var/home/core/zuul-output/0000755000175000017500000000000015113251636014530 5ustar corecorevar/home/core/zuul-output/logs/0000755000175000017500000000000015113256460015474 5ustar corecorevar/home/core/zuul-output/logs/kubelet.log0000644000000000000000004347232115113256451017706 0ustar rootrootDec 01 08:40:14 crc systemd[1]: Starting Kubernetes Kubelet... Dec 01 08:40:14 crc restorecon[4694]: Relabeled /var/lib/kubelet/config.json from system_u:object_r:unlabeled_t:s0 to system_u:object_r:container_var_lib_t:s0 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/device-plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/device-plugins/kubelet.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/volumes/kubernetes.io~configmap/nginx-conf/..2025_02_23_05_40_35.4114275528/nginx.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/22e96971 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/21c98286 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/0f1869e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/46889d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/5b6a5969 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/6c7921f5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4804f443 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/2a46b283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/a6b5573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4f88ee5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/5a4eee4b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/cd87c521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/38602af4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/1483b002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/0346718b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/d3ed4ada not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/3bb473a5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/8cd075a9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/00ab4760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/54a21c09 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/70478888 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/43802770 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/955a0edc not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/bca2d009 not reset as customized by admin to system_u:object_r:container_file_t:s0:c140,c1009 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/b295f9bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/bc46ea27 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5731fc1b not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5e1b2a3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/943f0936 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/3f764ee4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/8695e3f9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/aed7aa86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/c64d7448 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/0ba16bd2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/207a939f not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/54aa8cdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/1f5fa595 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/bf9c8153 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/47fba4ea not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/7ae55ce9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7906a268 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/ce43fa69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7fc7ea3a not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/d8c38b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/9ef015fb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/b9db6a41 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/b1733d79 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/afccd338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/9df0a185 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/18938cf8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/7ab4eb23 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/56930be6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_35.630010865 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/0d8e3722 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/d22b2e76 not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/e036759f not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/2734c483 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/57878fe7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/3f3c2e58 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/375bec3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/7bc41e08 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/48c7a72d not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/4b66701f not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/a5a1c202 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_40.1388695756 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/26f3df5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/6d8fb21d not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/50e94777 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208473b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/ec9e08ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3b787c39 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208eaed5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/93aa3a2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3c697968 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/ba950ec9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/cb5cdb37 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/f2df9827 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/fedaa673 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/9ca2df95 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/b2d7460e not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2207853c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/241c1c29 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2d910eaf not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/c6c0f2e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/399edc97 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8049f7cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/0cec5484 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/312446d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c406,c828 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8e56a35d not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/2d30ddb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/eca8053d not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/c3a25c9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c168,c522 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/b9609c22 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/e8b0eca9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/b36a9c3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/38af7b07 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/ae821620 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/baa23338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/2c534809 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/59b29eae not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/c91a8e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/4d87494a not reset as customized by admin to system_u:object_r:container_file_t:s0:c442,c857 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/1e33ca63 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/8dea7be2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d0b04a99 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d84f01e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/4109059b not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/a7258a3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/05bdf2b6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/f3261b51 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/315d045e not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/5fdcf278 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/d053f757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/c2850dc7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fcfb0b2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c7ac9b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fa0c0d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c609b6ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/2be6c296 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/89a32653 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/4eb9afeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/13af6efa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/b03f9724 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/e3d105cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/3aed4d83 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/0765fa6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/2cefc627 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/3dcc6345 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/365af391 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b1130c0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/236a5913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b9432e26 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/5ddb0e3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/986dc4fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/8a23ff9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/9728ae68 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/665f31d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/136c9b42 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/98a1575b not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/cac69136 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/5deb77a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/2ae53400 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/e46f2326 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/dc688d3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/3497c3cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/177eb008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/af5a2afa not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/d780cb1f not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/49b0f374 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/26fbb125 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/cf14125a not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/b7f86972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/e51d739c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/88ba6a69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/669a9acf not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/5cd51231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/75349ec7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/15c26839 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/45023dcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/2bb66a50 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/64d03bdd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/ab8e7ca0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/bb9be25f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/9a0b61d3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/d471b9d2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/8cb76b8e not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/11a00840 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/ec355a92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/992f735e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d59cdbbc not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/72133ff0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/c56c834c not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d13724c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/0a498258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa471982 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fc900d92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa7d68da not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/4bacf9b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/424021b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/fc2e31a3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/f51eefac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/c8997f2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/7481f599 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/fdafea19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/d0e1c571 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/ee398915 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/682bb6b8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a3e67855 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a989f289 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/915431bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/7796fdab not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/dcdb5f19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/a3aaa88c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/5508e3e6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/160585de not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/e99f8da3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/8bc85570 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/a5861c91 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/84db1135 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/9e1a6043 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/c1aba1c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/d55ccd6d not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/971cc9f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/8f2e3dcf not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/ceb35e9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/1c192745 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/5209e501 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/f83de4df not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/e7b978ac not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/c64304a1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/5384386b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/cce3e3ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/8fb75465 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/740f573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/32fd1134 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/0a861bd3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/80363026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/bfa952a8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..2025_02_23_05_33_31.333075221 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/793bf43d not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/7db1bb6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/4f6a0368 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/c12c7d86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/36c4a773 not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/4c1e98ae not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/a4c8115c not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/setup/7db1802e not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver/a008a7ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-syncer/2c836bac not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-regeneration-controller/0ce62299 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-insecure-readyz/945d2457 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-check-endpoints/7d5c1dd8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/index.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/bundle-v1.15.0.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/channel.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/package.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/bc8d0691 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/6b76097a not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/34d1af30 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/312ba61c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/645d5dd1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/16e825f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/4cf51fc9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/2a23d348 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/075dbd49 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/dd585ddd not reset as customized by admin to system_u:object_r:container_file_t:s0:c377,c642 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/17ebd0ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c343 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/005579f4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_23_11.1287037894 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/bf5f3b9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/af276eb7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/ea28e322 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/692e6683 not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/871746a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/4eb2e958 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/ca9b62da not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/0edd6fce not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/containers/controller-manager/89b4555f not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/655fcd71 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/0d43c002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/e68efd17 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/9acf9b65 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/5ae3ff11 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/1e59206a not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/27af16d1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c304,c1017 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/7918e729 not reset as customized by admin to system_u:object_r:container_file_t:s0:c853,c893 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/5d976d0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c585,c981 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/d7f55cbb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/f0812073 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/1a56cbeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/7fdd437e not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/cdfb5652 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/fix-audit-permissions/fb93119e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver/f1e8fc0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver-check-endpoints/218511f3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server/serving-certs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/ca8af7b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/72cc8a75 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/6e8a3760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4c3455c0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/2278acb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4b453e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/3ec09bda not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2/cacerts.bin not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java/cacerts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl/ca-bundle.trust.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/email-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/objsign-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2ae6433e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fde84897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75680d2e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/openshift-service-serving-signer_1740288168.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/facfc4fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f5a969c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CFCA_EV_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9ef4a08a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ingress-operator_1740288202.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2f332aed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/248c8271.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d10a21f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ACCVRAIZ1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a94d09e5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c9a4d3b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40193066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd8c0d63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b936d1c6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CA_Disig_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4fd49c6c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM_SERVIDORES_SEGUROS.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b81b93f0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f9a69fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b30d5fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ANF_Secure_Server_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b433981b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93851c9e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9282e51c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7dd1bc4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Actalis_Authentication_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/930ac5d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f47b495.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e113c810.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5931b5bc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Commercial.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2b349938.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e48193cf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/302904dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a716d4ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Networking.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93bc0acc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/86212b19.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b727005e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbc54cab.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f51bb24c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c28a8a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9c8dfbd4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ccc52f49.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cb1c3204.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ce5e74ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd08c599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6d41d539.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb5fa911.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e35234b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8cb5ee0f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a7c655d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f8fc53da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/de6d66f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d41b5e2a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/41a3f684.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1df5a75f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_2011.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e36a6752.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b872f2b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9576d26b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/228f89db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_ECC_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb717492.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d21b73c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b1b94ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/595e996b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_RSA_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b46e03d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/128f4b91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_3_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81f2d2b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Autoridad_de_Certificacion_Firmaprofesional_CIF_A62634068.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3bde41ac.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d16a5865.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_EC-384_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0179095f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ffa7f1eb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9482e63a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4dae3dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e359ba6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7e067d03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/95aff9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7746a63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Baltimore_CyberTrust_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/653b494a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3ad48a91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_2_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/54657681.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/82223c44.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8de2f56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d9dafe4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d96b65e2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee64a828.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40547a79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5a3f0ff8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a780d93.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/34d996fb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/eed8c118.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/89c02a45.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b1159c4c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d6325660.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4c339cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8312c4c1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_E1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8508e720.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5fdd185d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48bec511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/69105f4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b9bc432.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/32888f65.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b03dec0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/219d9499.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5acf816d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbf06781.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc99f41e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AAA_Certificate_Services.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/985c1f52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8794b4e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_BR_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7c037b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ef954a4e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_EV_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2add47b6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/90c5a3c8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0f3e76e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/53a1b57a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_EV_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5ad8a5d6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/68dd7389.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d04f354.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d6437c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/062cdee6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bd43e1dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7f3d5d1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c491639e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3513523f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/399e7759.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/feffd413.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d18e9066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/607986c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c90bc37d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1b0f7e5c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e08bfd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dd8e9d41.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed39abd0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a3418fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bc3f2570.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_High_Assurance_EV_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/244b5494.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81b9768f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4be590e0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_ECC_P384_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9846683b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/252252d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e8e7201.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_RSA4096_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d52c538d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c44cc0c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Trusted_Root_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75d1b2ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a2c66da8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ecccd8db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust.net_Certification_Authority__2048_.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/aee5f10d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e7271e8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0e59380.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4c3982f2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b99d060.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf64f35b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0a775a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/002c0b4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cc450945.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_EC1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/106f3e4d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b3fb433b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4042bcee.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/02265526.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/455f1b52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0d69c7e1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9f727ac7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5e98733a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0cd152c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc4d6a89.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6187b673.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/FIRMAPROFESIONAL_CA_ROOT-A_WEB.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ba8887ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/068570d1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f081611a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48a195d8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GDCA_TrustAUTH_R5_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f6fa695.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab59055e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b92fd57f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GLOBALTRUST_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fa5da96b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ec40989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7719f463.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1001acf7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f013ecaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/626dceaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c559d742.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1d3472b9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9479c8c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a81e292b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4bfab552.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e071171e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/57bcb2da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_ECC_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab5346f4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5046c355.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_RSA_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/865fbdf9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da0cfd1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/85cde254.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_ECC_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbb3f32b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureSign_RootCA11.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5860aaa6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/31188b5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HiPKI_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c7f1359b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f15c80c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hongkong_Post_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/09789157.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/18856ac4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e09d511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Commercial_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cf701eeb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d06393bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Public_Sector_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/10531352.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Izenpe.com.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureTrust_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0ed035a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsec_e-Szigno_Root_CA_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8160b96c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8651083.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2c63f966.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_ECC_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d89cda1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/01419da9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_RSA_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7a5b843.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_RSA_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf53fb88.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9591a472.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3afde786.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Gold_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NAVER_Global_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3fb36b73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d39b0a2c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a89d74c2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd58d51e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7db1890.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NetLock_Arany__Class_Gold__F__tan__s__tv__ny.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/988a38cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/60afe812.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f39fc864.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5443e9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GB_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e73d606e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dfc0fe80.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b66938e9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e1eab7c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GC_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/773e07ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c899c73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d59297b8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ddcda989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_1_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/749e9e03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/52b525c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7e8dc79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a819ef2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/08063a00.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b483515.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/064e0aa9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1f58a078.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6f7454b3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7fa05551.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76faf6c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9339512a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f387163d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee37c333.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e18bfb83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e442e424.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fe8a2cd8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/23f4c490.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5cd81ad7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0c70a8d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7892ad52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SZAFIR_ROOT_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4f316efb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_RSA_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/06dc52d5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/583d0756.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0bf05006.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/88950faa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9046744a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c860d51.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_RSA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6fa5da56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/33ee480d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Secure_Global_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/63a2c897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_ECC_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bdacca6f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ff34af3f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbff3a01.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_ECC_RootCA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_C1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/406c9bb1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_C3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Services_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Silver_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/99e1b953.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/14bc7599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TUBITAK_Kamu_SM_SSL_Kok_Sertifikasi_-_Surum_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a3adc42.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f459871d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_ECC_Root_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_RSA_Root_2023.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TeliaSonera_Root_CA_v1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telia_Root_CA_v2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f103249.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f058632f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-certificates.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9bf03295.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/98aaf404.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1cef98f5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/073bfcc5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2923b3f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f249de83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/edcbddb5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P256_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b5697b0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ae85e5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b74d2bd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P384_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d887a5bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9aef356c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TunTrust_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd64f3fc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e13665f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Extended_Validation_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f5dc4f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da7377f6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Global_G2_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c01eb047.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/304d27c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed858448.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f30dd6ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/04f60c28.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_ECC_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fc5a8f99.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/35105088.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee532fd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/XRamp_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/706f604c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76579174.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d86cdd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/882de061.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f618aec.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a9d40e02.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e-Szigno_Root_CA_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e868b802.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/83e9984f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ePKI_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca6e4ad9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d6523ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4b718d9b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/869fbf79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/containers/registry/f8d22bdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/6e8bbfac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/54dd7996 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/a4f1bb05 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/207129da not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/c1df39e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/15b8f1cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/77bd6913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/2382c1b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/704ce128 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/70d16fe0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/bfb95535 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/57a8e8e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/1b9d3e5e not reset as customized by admin to system_u:object_r:container_file_t:s0:c107,c917 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/fddb173c not reset as customized by admin to system_u:object_r:container_file_t:s0:c202,c983 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/95d3c6c4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/bfb5fff5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/2aef40aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/c0391cad not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/1119e69d not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/660608b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/8220bd53 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/85f99d5c not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/4b0225f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/9c2a3394 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/e820b243 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/1ca52ea0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/e6988e45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/6655f00b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/98bc3986 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/08e3458a not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/2a191cb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/6c4eeefb not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/f61a549c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/24891863 not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/fbdfd89c not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/9b63b3bc not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/8acde6d6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/node-driver-registrar/59ecbba3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/csi-provisioner/685d4be3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/containers/route-controller-manager/feaea55e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:14 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/63709497 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/d966b7fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/f5773757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/81c9edb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/57bf57ee not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/86f5e6aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/0aabe31d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/d2af85c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/09d157d9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c0fe7256 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c30319e4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/e6b1dd45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/2bb643f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/920de426 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/70fa1e87 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/a1c12a2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/9442e6c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/5b45ec72 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/3c9f3a59 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/1091c11b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/9a6821c6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/ec0c35e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/517f37e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/6214fe78 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/ba189c8b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/351e4f31 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/c0f219ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/8069f607 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/559c3d82 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/605ad488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/148df488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/3bf6dcb4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/022a2feb not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/938c3924 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/729fe23e not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/1fd5cbd4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/a96697e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/e155ddca not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/10dd0e0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/6f2c8392 not reset as customized by admin to system_u:object_r:container_file_t:s0:c267,c588 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/bd241ad9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/plugins/csi-hostpath not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/plugins/csi-hostpath/csi.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/plugins/kubernetes.io not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/plugins/kubernetes.io/csi not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983 not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/vol_data.json not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 01 08:40:15 crc restorecon[4694]: /var/lib/kubelet/plugins_registry not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 01 08:40:15 crc restorecon[4694]: Relabeled /var/usrlocal/bin/kubenswrapper from system_u:object_r:bin_t:s0 to system_u:object_r:kubelet_exec_t:s0 Dec 01 08:40:16 crc kubenswrapper[4813]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 01 08:40:16 crc kubenswrapper[4813]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Dec 01 08:40:16 crc kubenswrapper[4813]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 01 08:40:16 crc kubenswrapper[4813]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 01 08:40:16 crc kubenswrapper[4813]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Dec 01 08:40:16 crc kubenswrapper[4813]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.091438 4813 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.104171 4813 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.104216 4813 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.104229 4813 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.104240 4813 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.104251 4813 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.104262 4813 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.104278 4813 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.104295 4813 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.104308 4813 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.104320 4813 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.104330 4813 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.104341 4813 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.104352 4813 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.104362 4813 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.104372 4813 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.104383 4813 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.104393 4813 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.104403 4813 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.104414 4813 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.104425 4813 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.104439 4813 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.104450 4813 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.104462 4813 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.104472 4813 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.104484 4813 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.104494 4813 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.104504 4813 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.104515 4813 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.104564 4813 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.104577 4813 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.104591 4813 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.104604 4813 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.104618 4813 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.104627 4813 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.104637 4813 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.104647 4813 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.104657 4813 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.104668 4813 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.104678 4813 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.104688 4813 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.104698 4813 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.104708 4813 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.104719 4813 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.104729 4813 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.104739 4813 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.104751 4813 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.104760 4813 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.104770 4813 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.104780 4813 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.104789 4813 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.104801 4813 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.104810 4813 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.104820 4813 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.104830 4813 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.104840 4813 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.104851 4813 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.104865 4813 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.104876 4813 feature_gate.go:330] unrecognized feature gate: Example Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.104886 4813 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.104896 4813 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.104906 4813 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.104916 4813 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.104926 4813 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.104936 4813 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.104953 4813 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.104999 4813 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.105042 4813 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.105052 4813 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.105063 4813 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.105074 4813 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.105084 4813 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.105852 4813 flags.go:64] FLAG: --address="0.0.0.0" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.105890 4813 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.105911 4813 flags.go:64] FLAG: --anonymous-auth="true" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.105925 4813 flags.go:64] FLAG: --application-metrics-count-limit="100" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.105940 4813 flags.go:64] FLAG: --authentication-token-webhook="false" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.105951 4813 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.105999 4813 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.106012 4813 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.106023 4813 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.106033 4813 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.106045 4813 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.106055 4813 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.106066 4813 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.106076 4813 flags.go:64] FLAG: --cgroup-root="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.106086 4813 flags.go:64] FLAG: --cgroups-per-qos="true" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.106096 4813 flags.go:64] FLAG: --client-ca-file="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.106106 4813 flags.go:64] FLAG: --cloud-config="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.106116 4813 flags.go:64] FLAG: --cloud-provider="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.106125 4813 flags.go:64] FLAG: --cluster-dns="[]" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.106137 4813 flags.go:64] FLAG: --cluster-domain="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.106147 4813 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.106158 4813 flags.go:64] FLAG: --config-dir="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.106168 4813 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.106178 4813 flags.go:64] FLAG: --container-log-max-files="5" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.106191 4813 flags.go:64] FLAG: --container-log-max-size="10Mi" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.106201 4813 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.106212 4813 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.106223 4813 flags.go:64] FLAG: --containerd-namespace="k8s.io" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.106233 4813 flags.go:64] FLAG: --contention-profiling="false" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.106243 4813 flags.go:64] FLAG: --cpu-cfs-quota="true" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.106253 4813 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.106265 4813 flags.go:64] FLAG: --cpu-manager-policy="none" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.106275 4813 flags.go:64] FLAG: --cpu-manager-policy-options="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.106287 4813 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.106297 4813 flags.go:64] FLAG: --enable-controller-attach-detach="true" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.106307 4813 flags.go:64] FLAG: --enable-debugging-handlers="true" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.106319 4813 flags.go:64] FLAG: --enable-load-reader="false" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.106329 4813 flags.go:64] FLAG: --enable-server="true" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.106339 4813 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.106353 4813 flags.go:64] FLAG: --event-burst="100" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.106363 4813 flags.go:64] FLAG: --event-qps="50" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.106373 4813 flags.go:64] FLAG: --event-storage-age-limit="default=0" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.106383 4813 flags.go:64] FLAG: --event-storage-event-limit="default=0" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.106393 4813 flags.go:64] FLAG: --eviction-hard="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.106405 4813 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.106415 4813 flags.go:64] FLAG: --eviction-minimum-reclaim="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.106425 4813 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.106435 4813 flags.go:64] FLAG: --eviction-soft="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.106446 4813 flags.go:64] FLAG: --eviction-soft-grace-period="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.106455 4813 flags.go:64] FLAG: --exit-on-lock-contention="false" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.106465 4813 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.106475 4813 flags.go:64] FLAG: --experimental-mounter-path="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.106485 4813 flags.go:64] FLAG: --fail-cgroupv1="false" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.106495 4813 flags.go:64] FLAG: --fail-swap-on="true" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.106504 4813 flags.go:64] FLAG: --feature-gates="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.106516 4813 flags.go:64] FLAG: --file-check-frequency="20s" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.106526 4813 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.106537 4813 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.106547 4813 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.106557 4813 flags.go:64] FLAG: --healthz-port="10248" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.106567 4813 flags.go:64] FLAG: --help="false" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.106577 4813 flags.go:64] FLAG: --hostname-override="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.106588 4813 flags.go:64] FLAG: --housekeeping-interval="10s" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.106598 4813 flags.go:64] FLAG: --http-check-frequency="20s" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.106608 4813 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.106618 4813 flags.go:64] FLAG: --image-credential-provider-config="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.106627 4813 flags.go:64] FLAG: --image-gc-high-threshold="85" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.106637 4813 flags.go:64] FLAG: --image-gc-low-threshold="80" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.106647 4813 flags.go:64] FLAG: --image-service-endpoint="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.106657 4813 flags.go:64] FLAG: --kernel-memcg-notification="false" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.106666 4813 flags.go:64] FLAG: --kube-api-burst="100" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.106676 4813 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.106689 4813 flags.go:64] FLAG: --kube-api-qps="50" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.106699 4813 flags.go:64] FLAG: --kube-reserved="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.106709 4813 flags.go:64] FLAG: --kube-reserved-cgroup="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.106719 4813 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.106729 4813 flags.go:64] FLAG: --kubelet-cgroups="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.106739 4813 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.106749 4813 flags.go:64] FLAG: --lock-file="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.106759 4813 flags.go:64] FLAG: --log-cadvisor-usage="false" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.106769 4813 flags.go:64] FLAG: --log-flush-frequency="5s" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.106779 4813 flags.go:64] FLAG: --log-json-info-buffer-size="0" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.106828 4813 flags.go:64] FLAG: --log-json-split-stream="false" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.106839 4813 flags.go:64] FLAG: --log-text-info-buffer-size="0" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.106849 4813 flags.go:64] FLAG: --log-text-split-stream="false" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.106859 4813 flags.go:64] FLAG: --logging-format="text" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.106869 4813 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.106879 4813 flags.go:64] FLAG: --make-iptables-util-chains="true" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.106889 4813 flags.go:64] FLAG: --manifest-url="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.106899 4813 flags.go:64] FLAG: --manifest-url-header="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.106912 4813 flags.go:64] FLAG: --max-housekeeping-interval="15s" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.106923 4813 flags.go:64] FLAG: --max-open-files="1000000" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.106935 4813 flags.go:64] FLAG: --max-pods="110" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.106946 4813 flags.go:64] FLAG: --maximum-dead-containers="-1" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.106956 4813 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.106997 4813 flags.go:64] FLAG: --memory-manager-policy="None" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.107008 4813 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.107018 4813 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.107028 4813 flags.go:64] FLAG: --node-ip="192.168.126.11" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.107038 4813 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.107067 4813 flags.go:64] FLAG: --node-status-max-images="50" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.107078 4813 flags.go:64] FLAG: --node-status-update-frequency="10s" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.107087 4813 flags.go:64] FLAG: --oom-score-adj="-999" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.107097 4813 flags.go:64] FLAG: --pod-cidr="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.107107 4813 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.107123 4813 flags.go:64] FLAG: --pod-manifest-path="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.107133 4813 flags.go:64] FLAG: --pod-max-pids="-1" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.107143 4813 flags.go:64] FLAG: --pods-per-core="0" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.107154 4813 flags.go:64] FLAG: --port="10250" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.107165 4813 flags.go:64] FLAG: --protect-kernel-defaults="false" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.107175 4813 flags.go:64] FLAG: --provider-id="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.107185 4813 flags.go:64] FLAG: --qos-reserved="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.107194 4813 flags.go:64] FLAG: --read-only-port="10255" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.107204 4813 flags.go:64] FLAG: --register-node="true" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.107215 4813 flags.go:64] FLAG: --register-schedulable="true" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.107224 4813 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.107242 4813 flags.go:64] FLAG: --registry-burst="10" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.107251 4813 flags.go:64] FLAG: --registry-qps="5" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.107261 4813 flags.go:64] FLAG: --reserved-cpus="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.107270 4813 flags.go:64] FLAG: --reserved-memory="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.107283 4813 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.107293 4813 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.107304 4813 flags.go:64] FLAG: --rotate-certificates="false" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.107313 4813 flags.go:64] FLAG: --rotate-server-certificates="false" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.107323 4813 flags.go:64] FLAG: --runonce="false" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.107333 4813 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.107344 4813 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.107354 4813 flags.go:64] FLAG: --seccomp-default="false" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.107364 4813 flags.go:64] FLAG: --serialize-image-pulls="true" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.107374 4813 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.107384 4813 flags.go:64] FLAG: --storage-driver-db="cadvisor" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.107394 4813 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.107404 4813 flags.go:64] FLAG: --storage-driver-password="root" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.107414 4813 flags.go:64] FLAG: --storage-driver-secure="false" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.107423 4813 flags.go:64] FLAG: --storage-driver-table="stats" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.107433 4813 flags.go:64] FLAG: --storage-driver-user="root" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.107443 4813 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.107453 4813 flags.go:64] FLAG: --sync-frequency="1m0s" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.107464 4813 flags.go:64] FLAG: --system-cgroups="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.107473 4813 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.107490 4813 flags.go:64] FLAG: --system-reserved-cgroup="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.107499 4813 flags.go:64] FLAG: --tls-cert-file="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.107509 4813 flags.go:64] FLAG: --tls-cipher-suites="[]" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.107520 4813 flags.go:64] FLAG: --tls-min-version="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.107531 4813 flags.go:64] FLAG: --tls-private-key-file="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.107541 4813 flags.go:64] FLAG: --topology-manager-policy="none" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.107551 4813 flags.go:64] FLAG: --topology-manager-policy-options="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.107561 4813 flags.go:64] FLAG: --topology-manager-scope="container" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.107571 4813 flags.go:64] FLAG: --v="2" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.107584 4813 flags.go:64] FLAG: --version="false" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.107597 4813 flags.go:64] FLAG: --vmodule="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.107608 4813 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.107619 4813 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.107873 4813 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.107886 4813 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.107895 4813 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.107907 4813 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.107916 4813 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.107926 4813 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.107935 4813 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.107944 4813 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.107953 4813 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.107962 4813 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.107995 4813 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.108003 4813 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.108014 4813 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.108025 4813 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.108035 4813 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.108044 4813 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.108052 4813 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.108060 4813 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.108069 4813 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.108078 4813 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.108086 4813 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.108094 4813 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.108102 4813 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.108111 4813 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.108119 4813 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.108128 4813 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.108136 4813 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.108145 4813 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.108153 4813 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.108162 4813 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.108171 4813 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.108179 4813 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.108187 4813 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.108196 4813 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.108205 4813 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.108217 4813 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.108227 4813 feature_gate.go:330] unrecognized feature gate: Example Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.108238 4813 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.108249 4813 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.108261 4813 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.108272 4813 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.108281 4813 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.108290 4813 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.108299 4813 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.108309 4813 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.108318 4813 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.108326 4813 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.108334 4813 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.108343 4813 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.108351 4813 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.108360 4813 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.108368 4813 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.108376 4813 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.108386 4813 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.108394 4813 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.108403 4813 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.108411 4813 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.108419 4813 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.108427 4813 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.108436 4813 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.108444 4813 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.108453 4813 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.108461 4813 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.108472 4813 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.108480 4813 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.108489 4813 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.108497 4813 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.108506 4813 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.108515 4813 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.108524 4813 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.108532 4813 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.108555 4813 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.120527 4813 server.go:491] "Kubelet version" kubeletVersion="v1.31.5" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.120591 4813 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.120740 4813 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.120765 4813 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.120774 4813 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.120783 4813 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.120793 4813 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.120802 4813 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.120810 4813 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.120818 4813 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.120827 4813 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.120837 4813 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.120845 4813 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.120853 4813 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.120861 4813 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.120868 4813 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.120876 4813 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.120884 4813 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.120893 4813 feature_gate.go:330] unrecognized feature gate: Example Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.120900 4813 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.120908 4813 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.120916 4813 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.120924 4813 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.120933 4813 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.120941 4813 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.121003 4813 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.121015 4813 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.121028 4813 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.121038 4813 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.121048 4813 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.121058 4813 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.121067 4813 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.121076 4813 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.121087 4813 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.121095 4813 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.121103 4813 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.121111 4813 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.121119 4813 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.121127 4813 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.121135 4813 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.121143 4813 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.121151 4813 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.121158 4813 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.121169 4813 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.121179 4813 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.121188 4813 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.121196 4813 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.121206 4813 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.121214 4813 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.121223 4813 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.121231 4813 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.121239 4813 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.121247 4813 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.121256 4813 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.121265 4813 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.121275 4813 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.121286 4813 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.121295 4813 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.121304 4813 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.121313 4813 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.121321 4813 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.121329 4813 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.121337 4813 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.121345 4813 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.121352 4813 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.121360 4813 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.121368 4813 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.121376 4813 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.121384 4813 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.121392 4813 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.121400 4813 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.121407 4813 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.121415 4813 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.121438 4813 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.121668 4813 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.121681 4813 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.121690 4813 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.121698 4813 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.121707 4813 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.121715 4813 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.121724 4813 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.121731 4813 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.121739 4813 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.121748 4813 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.121756 4813 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.121764 4813 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.121772 4813 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.121780 4813 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.121787 4813 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.121795 4813 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.121803 4813 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.121813 4813 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.121822 4813 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.121830 4813 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.121837 4813 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.121846 4813 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.121854 4813 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.121861 4813 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.121869 4813 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.121877 4813 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.121885 4813 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.121895 4813 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.121907 4813 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.121917 4813 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.121927 4813 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.121937 4813 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.121951 4813 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.121988 4813 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.121997 4813 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.122008 4813 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.122017 4813 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.122027 4813 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.122035 4813 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.122043 4813 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.122050 4813 feature_gate.go:330] unrecognized feature gate: Example Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.122060 4813 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.122068 4813 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.122076 4813 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.122084 4813 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.122093 4813 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.122101 4813 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.122110 4813 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.122118 4813 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.122128 4813 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.122139 4813 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.122147 4813 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.122156 4813 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.122164 4813 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.122172 4813 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.122180 4813 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.122190 4813 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.122200 4813 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.122210 4813 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.122218 4813 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.122226 4813 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.122235 4813 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.122242 4813 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.122250 4813 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.122258 4813 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.122266 4813 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.122274 4813 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.122281 4813 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.122289 4813 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.122297 4813 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.122305 4813 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.122319 4813 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.122615 4813 server.go:940] "Client rotation is on, will bootstrap in background" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.129314 4813 bootstrap.go:85] "Current kubeconfig file contents are still valid, no bootstrap necessary" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.129530 4813 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.130758 4813 server.go:997] "Starting client certificate rotation" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.130815 4813 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.131444 4813 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-02-24 05:52:08 +0000 UTC, rotation deadline is 2025-12-18 06:06:30.56106964 +0000 UTC Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.131615 4813 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 405h26m14.42946846s for next certificate rotation Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.176025 4813 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.181508 4813 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.209391 4813 log.go:25] "Validated CRI v1 runtime API" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.245708 4813 log.go:25] "Validated CRI v1 image API" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.248187 4813 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.251832 4813 fs.go:133] Filesystem UUIDs: map[0b076daa-c26a-46d2-b3a6-72a8dbc6e257:/dev/vda4 2025-12-01-08-35-48-00:/dev/sr0 7B77-95E7:/dev/vda2 de0497b0-db1b-465a-b278-03db02455c71:/dev/vda3] Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.251882 4813 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:42 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:30 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:43 fsType:tmpfs blockSize:0}] Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.291218 4813 manager.go:217] Machine: {Timestamp:2025-12-01 08:40:16.285531552 +0000 UTC m=+0.908453218 CPUVendorID:AuthenticAMD NumCores:12 NumPhysicalCores:1 NumSockets:12 CpuFrequency:2800000 MemoryCapacity:33654124544 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:21801e6708c44f15b81395eb736a7cec SystemUUID:906e3783-eb6d-4f21-889c-7989aeaa25ba BootID:cd271ba1-a5c5-44e0-b03e-cfd1f5e8cb76 Filesystems:[{Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:16827060224 Type:vfs Inodes:4108169 HasInodes:true} {Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:6730825728 Type:vfs Inodes:819200 HasInodes:true} {Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true} {Device:/tmp DeviceMajor:0 DeviceMinor:30 Capacity:16827064320 Type:vfs Inodes:1048576 HasInodes:true} {Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true} {Device:/run/user/1000 DeviceMajor:0 DeviceMinor:42 Capacity:3365408768 Type:vfs Inodes:821633 HasInodes:true} {Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:43 Capacity:1073741824 Type:vfs Inodes:4108169 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:214748364800 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:fb:cd:5a Speed:0 Mtu:1500} {Name:br-int MacAddress:d6:39:55:2e:22:71 Speed:0 Mtu:1400} {Name:ens3 MacAddress:fa:16:3e:fb:cd:5a Speed:-1 Mtu:1500} {Name:ens7 MacAddress:fa:16:3e:d0:de:b1 Speed:-1 Mtu:1500} {Name:ens7.20 MacAddress:52:54:00:5b:d6:46 Speed:-1 Mtu:1496} {Name:ens7.21 MacAddress:52:54:00:79:aa:b8 Speed:-1 Mtu:1496} {Name:ens7.22 MacAddress:52:54:00:1e:94:17 Speed:-1 Mtu:1496} {Name:eth10 MacAddress:f2:d9:9d:ee:1e:60 Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:ae:85:8a:26:a3:c3 Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:33654124544 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[10] Caches:[{Id:10 Size:32768 Type:Data Level:1} {Id:10 Size:32768 Type:Instruction Level:1} {Id:10 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:10 Size:16777216 Type:Unified Level:3}] SocketID:10 BookID: DrawerID:} {Id:0 Threads:[11] Caches:[{Id:11 Size:32768 Type:Data Level:1} {Id:11 Size:32768 Type:Instruction Level:1} {Id:11 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:11 Size:16777216 Type:Unified Level:3}] SocketID:11 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:32768 Type:Data Level:1} {Id:4 Size:32768 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:32768 Type:Data Level:1} {Id:5 Size:32768 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:32768 Type:Data Level:1} {Id:6 Size:32768 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:32768 Type:Data Level:1} {Id:7 Size:32768 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:} {Id:0 Threads:[8] Caches:[{Id:8 Size:32768 Type:Data Level:1} {Id:8 Size:32768 Type:Instruction Level:1} {Id:8 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:8 Size:16777216 Type:Unified Level:3}] SocketID:8 BookID: DrawerID:} {Id:0 Threads:[9] Caches:[{Id:9 Size:32768 Type:Data Level:1} {Id:9 Size:32768 Type:Instruction Level:1} {Id:9 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:9 Size:16777216 Type:Unified Level:3}] SocketID:9 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.291574 4813 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.291751 4813 manager.go:233] Version: {KernelVersion:5.14.0-427.50.2.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202502100215-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.292154 4813 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.292417 4813 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.292460 4813 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.292766 4813 topology_manager.go:138] "Creating topology manager with none policy" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.292784 4813 container_manager_linux.go:303] "Creating device plugin manager" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.293107 4813 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.293175 4813 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.293448 4813 state_mem.go:36] "Initialized new in-memory state store" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.293798 4813 server.go:1245] "Using root directory" path="/var/lib/kubelet" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.295288 4813 kubelet.go:418] "Attempting to sync node with API server" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.295318 4813 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.295379 4813 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.295402 4813 kubelet.go:324] "Adding apiserver pod source" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.295432 4813 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.297947 4813 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.5-4.rhaos4.18.gitdad78d5.el9" apiVersion="v1" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.298473 4813 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.299802 4813 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.301360 4813 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.301401 4813 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.301417 4813 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.301449 4813 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.301474 4813 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.301487 4813 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.301500 4813 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.301518 4813 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.301533 4813 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.301547 4813 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.301565 4813 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.301577 4813 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.302034 4813 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.304062 4813 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.150:6443: connect: connection refused Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.304328 4813 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.150:6443: connect: connection refused Dec 01 08:40:16 crc kubenswrapper[4813]: E1201 08:40:16.304428 4813 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.150:6443: connect: connection refused" logger="UnhandledError" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.308361 4813 server.go:1280] "Started kubelet" Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.309074 4813 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.150:6443: connect: connection refused Dec 01 08:40:16 crc kubenswrapper[4813]: E1201 08:40:16.309215 4813 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.150:6443: connect: connection refused" logger="UnhandledError" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.309087 4813 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.309130 4813 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.310079 4813 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Dec 01 08:40:16 crc systemd[1]: Started Kubernetes Kubelet. Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.319071 4813 server.go:460] "Adding debug handlers to kubelet server" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.320691 4813 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.320770 4813 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.321175 4813 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-23 03:40:34.6966243 +0000 UTC Dec 01 08:40:16 crc kubenswrapper[4813]: E1201 08:40:16.316493 4813 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.150:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.187d0ab8720b4bba default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-01 08:40:16.302705594 +0000 UTC m=+0.925627210,LastTimestamp:2025-12-01 08:40:16.302705594 +0000 UTC m=+0.925627210,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.325009 4813 volume_manager.go:287] "The desired_state_of_world populator starts" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.325048 4813 volume_manager.go:289] "Starting Kubelet Volume Manager" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.325321 4813 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Dec 01 08:40:16 crc kubenswrapper[4813]: E1201 08:40:16.325555 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Dec 01 08:40:16 crc kubenswrapper[4813]: E1201 08:40:16.326826 4813 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.150:6443: connect: connection refused" interval="200ms" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.327365 4813 factory.go:55] Registering systemd factory Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.327440 4813 factory.go:221] Registration of the systemd container factory successfully Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.328172 4813 factory.go:153] Registering CRI-O factory Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.328313 4813 factory.go:221] Registration of the crio container factory successfully Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.328493 4813 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.328511 4813 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.150:6443: connect: connection refused Dec 01 08:40:16 crc kubenswrapper[4813]: E1201 08:40:16.328698 4813 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.150:6443: connect: connection refused" logger="UnhandledError" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.328870 4813 factory.go:103] Registering Raw factory Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.328921 4813 manager.go:1196] Started watching for new ooms in manager Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.330871 4813 manager.go:319] Starting recovery of all containers Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.347846 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" seLinuxMountContext="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.348259 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="44663579-783b-4372-86d6-acf235a62d72" volumeName="kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" seLinuxMountContext="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.348386 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" seLinuxMountContext="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.348516 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" seLinuxMountContext="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.348643 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" seLinuxMountContext="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.348766 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" seLinuxMountContext="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.348903 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" seLinuxMountContext="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.349061 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" seLinuxMountContext="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.349200 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" seLinuxMountContext="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.349321 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" seLinuxMountContext="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.349454 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" seLinuxMountContext="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.349577 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" seLinuxMountContext="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.349729 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" seLinuxMountContext="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.349859 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49ef4625-1d3a-4a9f-b595-c2433d32326d" volumeName="kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" seLinuxMountContext="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.350004 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" seLinuxMountContext="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.350137 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" seLinuxMountContext="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.350247 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" seLinuxMountContext="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.350355 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" seLinuxMountContext="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.350634 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" seLinuxMountContext="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.350781 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" seLinuxMountContext="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.350909 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" seLinuxMountContext="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.351084 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" seLinuxMountContext="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.351210 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" seLinuxMountContext="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.351348 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" seLinuxMountContext="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.351467 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" seLinuxMountContext="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.351552 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" seLinuxMountContext="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.351646 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d751cbb-f2e2-430d-9754-c882a5e924a5" volumeName="kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl" seLinuxMountContext="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.351732 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" seLinuxMountContext="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.351810 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" seLinuxMountContext="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.351896 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" seLinuxMountContext="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.352006 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" seLinuxMountContext="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.352129 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" seLinuxMountContext="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.352228 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" seLinuxMountContext="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.352381 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" seLinuxMountContext="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.352483 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" seLinuxMountContext="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.352601 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" seLinuxMountContext="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.352708 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" seLinuxMountContext="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.352814 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" seLinuxMountContext="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.352904 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" seLinuxMountContext="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.353035 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" seLinuxMountContext="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.353228 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" seLinuxMountContext="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.353312 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" seLinuxMountContext="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.353395 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" seLinuxMountContext="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.353478 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" seLinuxMountContext="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.353558 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" seLinuxMountContext="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.353643 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" seLinuxMountContext="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.353730 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" seLinuxMountContext="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.353810 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" seLinuxMountContext="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.353903 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" seLinuxMountContext="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.354063 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" seLinuxMountContext="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.354174 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" seLinuxMountContext="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.354272 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" seLinuxMountContext="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.354409 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" seLinuxMountContext="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.354492 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3b6479f0-333b-4a96-9adf-2099afdc2447" volumeName="kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr" seLinuxMountContext="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.354593 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" seLinuxMountContext="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.354680 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" seLinuxMountContext="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.354835 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" seLinuxMountContext="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.354934 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides" seLinuxMountContext="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.355055 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" seLinuxMountContext="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.355163 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" seLinuxMountContext="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.355256 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" seLinuxMountContext="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.355353 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" seLinuxMountContext="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.355441 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" seLinuxMountContext="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.355521 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" seLinuxMountContext="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.355610 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" seLinuxMountContext="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.355696 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert" seLinuxMountContext="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.355806 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" seLinuxMountContext="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.355896 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" seLinuxMountContext="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.356006 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert" seLinuxMountContext="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.356114 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" seLinuxMountContext="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.356202 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" seLinuxMountContext="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.356281 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" volumeName="kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" seLinuxMountContext="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.356388 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" seLinuxMountContext="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.356479 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" seLinuxMountContext="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.356564 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" seLinuxMountContext="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.356439 4813 manager.go:324] Recovery completed Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.356650 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" seLinuxMountContext="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.356919 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" seLinuxMountContext="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.357037 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" seLinuxMountContext="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.357143 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5" seLinuxMountContext="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.357226 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" seLinuxMountContext="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.357319 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" seLinuxMountContext="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.357407 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" seLinuxMountContext="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.357501 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" seLinuxMountContext="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.357594 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" volumeName="kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" seLinuxMountContext="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.357682 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb" seLinuxMountContext="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.357765 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" seLinuxMountContext="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.357844 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" seLinuxMountContext="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.357933 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" seLinuxMountContext="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.358045 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" seLinuxMountContext="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.358131 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" seLinuxMountContext="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.358223 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" seLinuxMountContext="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.358305 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" seLinuxMountContext="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.358390 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" seLinuxMountContext="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.358468 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" seLinuxMountContext="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.358548 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm" seLinuxMountContext="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.358627 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" seLinuxMountContext="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.358724 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" seLinuxMountContext="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.358817 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" seLinuxMountContext="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.358900 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" seLinuxMountContext="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.359022 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" seLinuxMountContext="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.359114 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" seLinuxMountContext="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.359192 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" seLinuxMountContext="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.359284 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" seLinuxMountContext="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.359373 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" seLinuxMountContext="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.359463 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.359546 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" seLinuxMountContext="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.359627 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" seLinuxMountContext="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.359709 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" seLinuxMountContext="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.359800 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" seLinuxMountContext="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.359888 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" seLinuxMountContext="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.359994 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" seLinuxMountContext="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.360121 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" seLinuxMountContext="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.360212 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" seLinuxMountContext="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.360306 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" seLinuxMountContext="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.360406 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" seLinuxMountContext="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.360550 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" seLinuxMountContext="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.360644 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls" seLinuxMountContext="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.365649 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" seLinuxMountContext="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.365777 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" seLinuxMountContext="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.365832 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" seLinuxMountContext="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.365863 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" seLinuxMountContext="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.365897 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" seLinuxMountContext="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.365931 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" seLinuxMountContext="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.371134 4813 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.371317 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" seLinuxMountContext="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.371356 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" seLinuxMountContext="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.371388 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" seLinuxMountContext="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.371407 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" seLinuxMountContext="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.371445 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" seLinuxMountContext="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.371475 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" seLinuxMountContext="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.371490 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" seLinuxMountContext="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.371509 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" seLinuxMountContext="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.371525 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" seLinuxMountContext="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.371592 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" seLinuxMountContext="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.371617 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" seLinuxMountContext="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.371646 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" seLinuxMountContext="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.371687 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" seLinuxMountContext="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.371736 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" seLinuxMountContext="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.371773 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf" seLinuxMountContext="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.371792 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" seLinuxMountContext="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.371805 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" seLinuxMountContext="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.371824 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" seLinuxMountContext="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.371837 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" seLinuxMountContext="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.371878 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" seLinuxMountContext="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.371919 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" seLinuxMountContext="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.371934 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" seLinuxMountContext="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.371951 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" seLinuxMountContext="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.372134 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" seLinuxMountContext="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.372149 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" seLinuxMountContext="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.372169 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" seLinuxMountContext="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.372183 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" seLinuxMountContext="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.372261 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" seLinuxMountContext="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.372307 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" seLinuxMountContext="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.372321 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" seLinuxMountContext="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.372338 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" seLinuxMountContext="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.372351 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" seLinuxMountContext="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.372419 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" seLinuxMountContext="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.372440 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" seLinuxMountContext="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.372454 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" seLinuxMountContext="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.372473 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" seLinuxMountContext="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.372505 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" seLinuxMountContext="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.372549 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" seLinuxMountContext="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.372568 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" seLinuxMountContext="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.372584 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" seLinuxMountContext="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.372601 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" seLinuxMountContext="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.372615 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" seLinuxMountContext="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.372628 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" seLinuxMountContext="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.372658 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" seLinuxMountContext="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.372684 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" seLinuxMountContext="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.372702 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" seLinuxMountContext="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.372737 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" seLinuxMountContext="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.372750 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" seLinuxMountContext="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.372767 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" seLinuxMountContext="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.372782 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" seLinuxMountContext="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.372798 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" seLinuxMountContext="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.372812 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" seLinuxMountContext="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.372841 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" seLinuxMountContext="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.372859 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" seLinuxMountContext="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.372872 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script" seLinuxMountContext="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.372885 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" seLinuxMountContext="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.372903 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" seLinuxMountContext="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.372943 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" seLinuxMountContext="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.372980 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" seLinuxMountContext="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.373028 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" seLinuxMountContext="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.373057 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" seLinuxMountContext="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.373075 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" seLinuxMountContext="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.373087 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" seLinuxMountContext="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.373125 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" seLinuxMountContext="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.373139 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf" seLinuxMountContext="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.373214 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" seLinuxMountContext="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.373233 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" seLinuxMountContext="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.373246 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" seLinuxMountContext="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.373293 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" seLinuxMountContext="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.373309 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" seLinuxMountContext="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.373323 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" seLinuxMountContext="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.373371 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" seLinuxMountContext="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.373387 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" seLinuxMountContext="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.373405 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" seLinuxMountContext="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.373419 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" seLinuxMountContext="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.373450 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" seLinuxMountContext="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.373501 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" seLinuxMountContext="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.373516 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" seLinuxMountContext="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.373550 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" seLinuxMountContext="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.373563 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" seLinuxMountContext="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.373576 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" seLinuxMountContext="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.373595 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" seLinuxMountContext="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.373609 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" seLinuxMountContext="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.373646 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" seLinuxMountContext="" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.373680 4813 reconstruct.go:97] "Volume reconstruction finished" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.373691 4813 reconciler.go:26] "Reconciler: start to sync state" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.385332 4813 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.387628 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.387693 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.387712 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.388871 4813 cpu_manager.go:225] "Starting CPU manager" policy="none" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.388900 4813 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.388936 4813 state_mem.go:36] "Initialized new in-memory state store" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.389567 4813 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.391811 4813 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.392077 4813 status_manager.go:217] "Starting to sync pod status with apiserver" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.392267 4813 kubelet.go:2335] "Starting kubelet main sync loop" Dec 01 08:40:16 crc kubenswrapper[4813]: E1201 08:40:16.392488 4813 kubelet.go:2359] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.394144 4813 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.150:6443: connect: connection refused Dec 01 08:40:16 crc kubenswrapper[4813]: E1201 08:40:16.394226 4813 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.150:6443: connect: connection refused" logger="UnhandledError" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.405572 4813 policy_none.go:49] "None policy: Start" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.406783 4813 memory_manager.go:170] "Starting memorymanager" policy="None" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.406845 4813 state_mem.go:35] "Initializing new in-memory state store" Dec 01 08:40:16 crc kubenswrapper[4813]: E1201 08:40:16.426646 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.464499 4813 manager.go:334] "Starting Device Plugin manager" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.464865 4813 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.464905 4813 server.go:79] "Starting device plugin registration server" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.465482 4813 eviction_manager.go:189] "Eviction manager: starting control loop" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.465526 4813 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.465694 4813 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.465838 4813 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.465857 4813 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Dec 01 08:40:16 crc kubenswrapper[4813]: E1201 08:40:16.475115 4813 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.492810 4813 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc","openshift-etcd/etcd-crc","openshift-kube-apiserver/kube-apiserver-crc","openshift-kube-controller-manager/kube-controller-manager-crc","openshift-kube-scheduler/openshift-kube-scheduler-crc"] Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.493078 4813 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.494716 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.494773 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.494795 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.495106 4813 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.495491 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.495569 4813 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.496360 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.496397 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.496411 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.496602 4813 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.496893 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.497005 4813 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.497525 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.497571 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.497591 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.497752 4813 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.497848 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.497892 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.497907 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.497952 4813 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.497913 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.498212 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.498247 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.498260 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.498915 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.498955 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.498995 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.499147 4813 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.499374 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.499456 4813 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.499376 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.499600 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.499623 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.500026 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.500072 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.500090 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.500306 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.500346 4813 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.500656 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.500701 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.500719 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.501252 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.501288 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.501302 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:16 crc kubenswrapper[4813]: E1201 08:40:16.527855 4813 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.150:6443: connect: connection refused" interval="400ms" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.566451 4813 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.567761 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.567805 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.567820 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.567854 4813 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 01 08:40:16 crc kubenswrapper[4813]: E1201 08:40:16.568493 4813 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.150:6443: connect: connection refused" node="crc" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.576483 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.576521 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.576549 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.576571 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.576670 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.576839 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.576883 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.576923 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.576952 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.577032 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.577056 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.577103 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.577128 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.577160 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.577191 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.678750 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.678823 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.678852 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.678876 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.678899 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.678953 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.679001 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.679025 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.679076 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.679103 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.679131 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.679139 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.679188 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.679144 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.679149 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.679152 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.679290 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.679302 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.679354 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.679403 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.679335 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.679207 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.679330 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.679474 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.679361 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.679359 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.679372 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.679537 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.679575 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.679635 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.768740 4813 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.771339 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.771463 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.771516 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.771607 4813 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 01 08:40:16 crc kubenswrapper[4813]: E1201 08:40:16.772472 4813 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.150:6443: connect: connection refused" node="crc" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.845393 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.877775 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.888228 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 08:40:16 crc kubenswrapper[4813]: E1201 08:40:16.928882 4813 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.150:6443: connect: connection refused" interval="800ms" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.930480 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 01 08:40:16 crc kubenswrapper[4813]: I1201 08:40:16.942551 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.947014 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2139d3e2895fc6797b9c76a1b4c9886d.slice/crio-51da8feac20ca696b7256e84e13cec16cca5bc35c4d4e90f30d4ddb3c887549b WatchSource:0}: Error finding container 51da8feac20ca696b7256e84e13cec16cca5bc35c4d4e90f30d4ddb3c887549b: Status 404 returned error can't find the container with id 51da8feac20ca696b7256e84e13cec16cca5bc35c4d4e90f30d4ddb3c887549b Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.947527 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf4b27818a5e8e43d0dc095d08835c792.slice/crio-080c19f9486e7361234e309f73a45deb05f90077ffba6005c2a313e4b94e74d2 WatchSource:0}: Error finding container 080c19f9486e7361234e309f73a45deb05f90077ffba6005c2a313e4b94e74d2: Status 404 returned error can't find the container with id 080c19f9486e7361234e309f73a45deb05f90077ffba6005c2a313e4b94e74d2 Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.949801 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1b160f5dda77d281dd8e69ec8d817f9.slice/crio-fc5f15c74aa041533197752e823ccf63b8a3f5523132a06765f6c90caa9dad01 WatchSource:0}: Error finding container fc5f15c74aa041533197752e823ccf63b8a3f5523132a06765f6c90caa9dad01: Status 404 returned error can't find the container with id fc5f15c74aa041533197752e823ccf63b8a3f5523132a06765f6c90caa9dad01 Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.951934 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf614b9022728cf315e60c057852e563e.slice/crio-bcb34f184655fc5fc46588bf205b5a644d927f742549e3fe87459633efb28d39 WatchSource:0}: Error finding container bcb34f184655fc5fc46588bf205b5a644d927f742549e3fe87459633efb28d39: Status 404 returned error can't find the container with id bcb34f184655fc5fc46588bf205b5a644d927f742549e3fe87459633efb28d39 Dec 01 08:40:16 crc kubenswrapper[4813]: W1201 08:40:16.962399 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3dcd261975c3d6b9a6ad6367fd4facd3.slice/crio-c07e3393a8b6b1794f824fd15654ee83716cd39689b3d4bf3e8017c67d2f345c WatchSource:0}: Error finding container c07e3393a8b6b1794f824fd15654ee83716cd39689b3d4bf3e8017c67d2f345c: Status 404 returned error can't find the container with id c07e3393a8b6b1794f824fd15654ee83716cd39689b3d4bf3e8017c67d2f345c Dec 01 08:40:17 crc kubenswrapper[4813]: I1201 08:40:17.172796 4813 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 08:40:17 crc kubenswrapper[4813]: I1201 08:40:17.174288 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:17 crc kubenswrapper[4813]: I1201 08:40:17.174325 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:17 crc kubenswrapper[4813]: I1201 08:40:17.174335 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:17 crc kubenswrapper[4813]: I1201 08:40:17.174360 4813 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 01 08:40:17 crc kubenswrapper[4813]: E1201 08:40:17.174761 4813 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.150:6443: connect: connection refused" node="crc" Dec 01 08:40:17 crc kubenswrapper[4813]: W1201 08:40:17.256468 4813 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.150:6443: connect: connection refused Dec 01 08:40:17 crc kubenswrapper[4813]: E1201 08:40:17.256553 4813 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.150:6443: connect: connection refused" logger="UnhandledError" Dec 01 08:40:17 crc kubenswrapper[4813]: I1201 08:40:17.305656 4813 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.150:6443: connect: connection refused Dec 01 08:40:17 crc kubenswrapper[4813]: I1201 08:40:17.321858 4813 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-18 06:33:13.019721631 +0000 UTC Dec 01 08:40:17 crc kubenswrapper[4813]: I1201 08:40:17.321935 4813 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 405h52m55.697789471s for next certificate rotation Dec 01 08:40:17 crc kubenswrapper[4813]: W1201 08:40:17.322432 4813 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.150:6443: connect: connection refused Dec 01 08:40:17 crc kubenswrapper[4813]: E1201 08:40:17.322498 4813 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.150:6443: connect: connection refused" logger="UnhandledError" Dec 01 08:40:17 crc kubenswrapper[4813]: I1201 08:40:17.395637 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"c07e3393a8b6b1794f824fd15654ee83716cd39689b3d4bf3e8017c67d2f345c"} Dec 01 08:40:17 crc kubenswrapper[4813]: I1201 08:40:17.397551 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"bcb34f184655fc5fc46588bf205b5a644d927f742549e3fe87459633efb28d39"} Dec 01 08:40:17 crc kubenswrapper[4813]: I1201 08:40:17.401656 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"080c19f9486e7361234e309f73a45deb05f90077ffba6005c2a313e4b94e74d2"} Dec 01 08:40:17 crc kubenswrapper[4813]: I1201 08:40:17.403105 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"51da8feac20ca696b7256e84e13cec16cca5bc35c4d4e90f30d4ddb3c887549b"} Dec 01 08:40:17 crc kubenswrapper[4813]: I1201 08:40:17.403897 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"fc5f15c74aa041533197752e823ccf63b8a3f5523132a06765f6c90caa9dad01"} Dec 01 08:40:17 crc kubenswrapper[4813]: W1201 08:40:17.406716 4813 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.150:6443: connect: connection refused Dec 01 08:40:17 crc kubenswrapper[4813]: E1201 08:40:17.407097 4813 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.150:6443: connect: connection refused" logger="UnhandledError" Dec 01 08:40:17 crc kubenswrapper[4813]: W1201 08:40:17.640419 4813 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.150:6443: connect: connection refused Dec 01 08:40:17 crc kubenswrapper[4813]: E1201 08:40:17.640512 4813 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.150:6443: connect: connection refused" logger="UnhandledError" Dec 01 08:40:17 crc kubenswrapper[4813]: E1201 08:40:17.730194 4813 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.150:6443: connect: connection refused" interval="1.6s" Dec 01 08:40:17 crc kubenswrapper[4813]: E1201 08:40:17.843441 4813 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.150:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.187d0ab8720b4bba default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-01 08:40:16.302705594 +0000 UTC m=+0.925627210,LastTimestamp:2025-12-01 08:40:16.302705594 +0000 UTC m=+0.925627210,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 01 08:40:17 crc kubenswrapper[4813]: I1201 08:40:17.975534 4813 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 08:40:17 crc kubenswrapper[4813]: I1201 08:40:17.980307 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:17 crc kubenswrapper[4813]: I1201 08:40:17.980354 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:17 crc kubenswrapper[4813]: I1201 08:40:17.980370 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:17 crc kubenswrapper[4813]: I1201 08:40:17.980400 4813 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 01 08:40:17 crc kubenswrapper[4813]: E1201 08:40:17.981082 4813 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.150:6443: connect: connection refused" node="crc" Dec 01 08:40:18 crc kubenswrapper[4813]: I1201 08:40:18.305869 4813 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.150:6443: connect: connection refused Dec 01 08:40:18 crc kubenswrapper[4813]: I1201 08:40:18.409182 4813 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="2e1a47da448b5c46933f8c322a953219afd20cd88855dba131cb27e1625545d0" exitCode=0 Dec 01 08:40:18 crc kubenswrapper[4813]: I1201 08:40:18.409336 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"2e1a47da448b5c46933f8c322a953219afd20cd88855dba131cb27e1625545d0"} Dec 01 08:40:18 crc kubenswrapper[4813]: I1201 08:40:18.409369 4813 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 08:40:18 crc kubenswrapper[4813]: I1201 08:40:18.411127 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:18 crc kubenswrapper[4813]: I1201 08:40:18.411179 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:18 crc kubenswrapper[4813]: I1201 08:40:18.411191 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:18 crc kubenswrapper[4813]: I1201 08:40:18.411742 4813 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 08:40:18 crc kubenswrapper[4813]: I1201 08:40:18.411774 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerDied","Data":"62135a7a7fa690060354d059965c832020df757ae8b6b0f9cefdb21f154b820e"} Dec 01 08:40:18 crc kubenswrapper[4813]: I1201 08:40:18.411607 4813 generic.go:334] "Generic (PLEG): container finished" podID="d1b160f5dda77d281dd8e69ec8d817f9" containerID="62135a7a7fa690060354d059965c832020df757ae8b6b0f9cefdb21f154b820e" exitCode=0 Dec 01 08:40:18 crc kubenswrapper[4813]: I1201 08:40:18.413191 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:18 crc kubenswrapper[4813]: I1201 08:40:18.413233 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:18 crc kubenswrapper[4813]: I1201 08:40:18.413255 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:18 crc kubenswrapper[4813]: I1201 08:40:18.414378 4813 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="53de1f0682f36aa83ad8e5c226b95daf9f97df4ae154ebfc2fc287f9f238a4a8" exitCode=0 Dec 01 08:40:18 crc kubenswrapper[4813]: I1201 08:40:18.414445 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"53de1f0682f36aa83ad8e5c226b95daf9f97df4ae154ebfc2fc287f9f238a4a8"} Dec 01 08:40:18 crc kubenswrapper[4813]: I1201 08:40:18.414499 4813 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 08:40:18 crc kubenswrapper[4813]: I1201 08:40:18.415550 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:18 crc kubenswrapper[4813]: I1201 08:40:18.415577 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:18 crc kubenswrapper[4813]: I1201 08:40:18.415592 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:18 crc kubenswrapper[4813]: I1201 08:40:18.418540 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"9902c00bf7738686c93f62e1772a484d05715d74b75ee98c84c8f23ed383824b"} Dec 01 08:40:18 crc kubenswrapper[4813]: I1201 08:40:18.418611 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"5e24ebfc52dcb28368a2d436e7d9ab78f07f2aec2a4c71405a6c7d3301eee96a"} Dec 01 08:40:18 crc kubenswrapper[4813]: I1201 08:40:18.418641 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"33898118bbf639c8e6eae5aafa10386b95795512444ed5f535db3c663f169df2"} Dec 01 08:40:18 crc kubenswrapper[4813]: I1201 08:40:18.421414 4813 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="05492df5703392cdf0932d6bf66025b18e53bd05fa0857a7cb2ad70979cbdc3e" exitCode=0 Dec 01 08:40:18 crc kubenswrapper[4813]: I1201 08:40:18.421460 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"05492df5703392cdf0932d6bf66025b18e53bd05fa0857a7cb2ad70979cbdc3e"} Dec 01 08:40:18 crc kubenswrapper[4813]: I1201 08:40:18.421637 4813 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 08:40:18 crc kubenswrapper[4813]: I1201 08:40:18.425053 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:18 crc kubenswrapper[4813]: I1201 08:40:18.425103 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:18 crc kubenswrapper[4813]: I1201 08:40:18.425118 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:18 crc kubenswrapper[4813]: I1201 08:40:18.431211 4813 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 08:40:18 crc kubenswrapper[4813]: I1201 08:40:18.432167 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:18 crc kubenswrapper[4813]: I1201 08:40:18.432202 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:18 crc kubenswrapper[4813]: I1201 08:40:18.432214 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:19 crc kubenswrapper[4813]: I1201 08:40:19.305132 4813 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.150:6443: connect: connection refused Dec 01 08:40:19 crc kubenswrapper[4813]: E1201 08:40:19.331390 4813 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.150:6443: connect: connection refused" interval="3.2s" Dec 01 08:40:19 crc kubenswrapper[4813]: I1201 08:40:19.446285 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"3657b47ec9f1563c83bfbe3d728c8fd123f7078e219d251b1874269e42c31eea"} Dec 01 08:40:19 crc kubenswrapper[4813]: I1201 08:40:19.446770 4813 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 08:40:19 crc kubenswrapper[4813]: I1201 08:40:19.448870 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:19 crc kubenswrapper[4813]: I1201 08:40:19.448913 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:19 crc kubenswrapper[4813]: I1201 08:40:19.448928 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:19 crc kubenswrapper[4813]: I1201 08:40:19.450558 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"6e97356ee274ea78d9f237ffbc756abb6ac8df0dbac244bf0a86ed215b8cbf43"} Dec 01 08:40:19 crc kubenswrapper[4813]: I1201 08:40:19.450605 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"d441a348087ae3cadda1ad8a7d1c4dff02cf3f6e1d346c294b278fdd9131edcc"} Dec 01 08:40:19 crc kubenswrapper[4813]: I1201 08:40:19.505843 4813 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="e4be2790fe443bc1f9774107455987501b3d623468e7026e6ad7f7980076dd77" exitCode=0 Dec 01 08:40:19 crc kubenswrapper[4813]: I1201 08:40:19.505934 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"e4be2790fe443bc1f9774107455987501b3d623468e7026e6ad7f7980076dd77"} Dec 01 08:40:19 crc kubenswrapper[4813]: I1201 08:40:19.506273 4813 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 08:40:19 crc kubenswrapper[4813]: I1201 08:40:19.509550 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:19 crc kubenswrapper[4813]: I1201 08:40:19.509589 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:19 crc kubenswrapper[4813]: I1201 08:40:19.509598 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:19 crc kubenswrapper[4813]: I1201 08:40:19.510189 4813 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 08:40:19 crc kubenswrapper[4813]: I1201 08:40:19.510306 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"86abafcbe7fed4bfbaa6c3b2f64d6a314afa7ed0fe2709bc8a01ba9fb17c1a75"} Dec 01 08:40:19 crc kubenswrapper[4813]: I1201 08:40:19.511188 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:19 crc kubenswrapper[4813]: I1201 08:40:19.511214 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:19 crc kubenswrapper[4813]: I1201 08:40:19.511224 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:19 crc kubenswrapper[4813]: I1201 08:40:19.513190 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"74f8cf73f07b4cde260af429d26413e4e6226e7b177b493dd9c0e5c043ab6e1a"} Dec 01 08:40:19 crc kubenswrapper[4813]: I1201 08:40:19.513753 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"fb48bd28fd3d29781695a5ae630f4913ef2745e97fbe7d463cca39a31e43ba0a"} Dec 01 08:40:19 crc kubenswrapper[4813]: W1201 08:40:19.546400 4813 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.150:6443: connect: connection refused Dec 01 08:40:19 crc kubenswrapper[4813]: E1201 08:40:19.546470 4813 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.150:6443: connect: connection refused" logger="UnhandledError" Dec 01 08:40:19 crc kubenswrapper[4813]: I1201 08:40:19.582007 4813 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 08:40:19 crc kubenswrapper[4813]: I1201 08:40:19.583711 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:19 crc kubenswrapper[4813]: I1201 08:40:19.583750 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:19 crc kubenswrapper[4813]: I1201 08:40:19.583759 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:19 crc kubenswrapper[4813]: I1201 08:40:19.583787 4813 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 01 08:40:19 crc kubenswrapper[4813]: E1201 08:40:19.584303 4813 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.150:6443: connect: connection refused" node="crc" Dec 01 08:40:19 crc kubenswrapper[4813]: I1201 08:40:19.857221 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 01 08:40:20 crc kubenswrapper[4813]: W1201 08:40:20.051078 4813 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.150:6443: connect: connection refused Dec 01 08:40:20 crc kubenswrapper[4813]: E1201 08:40:20.051180 4813 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.150:6443: connect: connection refused" logger="UnhandledError" Dec 01 08:40:20 crc kubenswrapper[4813]: W1201 08:40:20.130560 4813 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.150:6443: connect: connection refused Dec 01 08:40:20 crc kubenswrapper[4813]: E1201 08:40:20.130646 4813 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.150:6443: connect: connection refused" logger="UnhandledError" Dec 01 08:40:20 crc kubenswrapper[4813]: W1201 08:40:20.158050 4813 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.150:6443: connect: connection refused Dec 01 08:40:20 crc kubenswrapper[4813]: E1201 08:40:20.158153 4813 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.150:6443: connect: connection refused" logger="UnhandledError" Dec 01 08:40:20 crc kubenswrapper[4813]: I1201 08:40:20.305703 4813 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.150:6443: connect: connection refused Dec 01 08:40:20 crc kubenswrapper[4813]: I1201 08:40:20.517779 4813 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 08:40:20 crc kubenswrapper[4813]: I1201 08:40:20.517778 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"3aaf696cefbe398bb6b26acfc5be43bda7e2be18218c8060840350c4106e66a4"} Dec 01 08:40:20 crc kubenswrapper[4813]: I1201 08:40:20.518578 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:20 crc kubenswrapper[4813]: I1201 08:40:20.518628 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:20 crc kubenswrapper[4813]: I1201 08:40:20.518644 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:20 crc kubenswrapper[4813]: I1201 08:40:20.522056 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"32238885283b14552bdd563de01057b496e11a9e931d80b07f02315a53e9d29b"} Dec 01 08:40:20 crc kubenswrapper[4813]: I1201 08:40:20.522113 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"7354f9337d42115bb1d974bf3772f2d03365acba57f2155135c7a0bbab39879e"} Dec 01 08:40:20 crc kubenswrapper[4813]: I1201 08:40:20.522133 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"da67635daf6d9e32937f4d5d4ceeee3f8f09ae4fd2dd26c3e85be1286c5949de"} Dec 01 08:40:20 crc kubenswrapper[4813]: I1201 08:40:20.522133 4813 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 08:40:20 crc kubenswrapper[4813]: I1201 08:40:20.525006 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:20 crc kubenswrapper[4813]: I1201 08:40:20.525089 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:20 crc kubenswrapper[4813]: I1201 08:40:20.525124 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:20 crc kubenswrapper[4813]: I1201 08:40:20.527318 4813 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="ed78ffc127c55bf3726b69025c45af62a5f14f29ae9b1e3c3c5b2f00d8fd65cc" exitCode=0 Dec 01 08:40:20 crc kubenswrapper[4813]: I1201 08:40:20.527416 4813 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 08:40:20 crc kubenswrapper[4813]: I1201 08:40:20.527427 4813 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 08:40:20 crc kubenswrapper[4813]: I1201 08:40:20.527423 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"ed78ffc127c55bf3726b69025c45af62a5f14f29ae9b1e3c3c5b2f00d8fd65cc"} Dec 01 08:40:20 crc kubenswrapper[4813]: I1201 08:40:20.527594 4813 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 08:40:20 crc kubenswrapper[4813]: I1201 08:40:20.528369 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:20 crc kubenswrapper[4813]: I1201 08:40:20.528429 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:20 crc kubenswrapper[4813]: I1201 08:40:20.528448 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:20 crc kubenswrapper[4813]: I1201 08:40:20.528455 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:20 crc kubenswrapper[4813]: I1201 08:40:20.528485 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:20 crc kubenswrapper[4813]: I1201 08:40:20.528498 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:20 crc kubenswrapper[4813]: I1201 08:40:20.528987 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:20 crc kubenswrapper[4813]: I1201 08:40:20.529019 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:20 crc kubenswrapper[4813]: I1201 08:40:20.529032 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:21 crc kubenswrapper[4813]: I1201 08:40:21.171709 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 08:40:21 crc kubenswrapper[4813]: I1201 08:40:21.551794 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"094131af57d9160da70c4800886d937e51007738ab34dd63dc89f756c3cfc1d6"} Dec 01 08:40:21 crc kubenswrapper[4813]: I1201 08:40:21.551909 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"74c67c43ba4e723a94f6e4518e0cd0ae2f43a4920a1c9b4d107a43fd908887a8"} Dec 01 08:40:21 crc kubenswrapper[4813]: I1201 08:40:21.551923 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"1758729c056f33ff81cd649b56a2cb74a2ad23433cebd38366c72788f9eddd86"} Dec 01 08:40:21 crc kubenswrapper[4813]: I1201 08:40:21.554143 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Dec 01 08:40:21 crc kubenswrapper[4813]: I1201 08:40:21.556536 4813 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="32238885283b14552bdd563de01057b496e11a9e931d80b07f02315a53e9d29b" exitCode=255 Dec 01 08:40:21 crc kubenswrapper[4813]: I1201 08:40:21.556719 4813 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 01 08:40:21 crc kubenswrapper[4813]: I1201 08:40:21.556791 4813 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 08:40:21 crc kubenswrapper[4813]: I1201 08:40:21.556940 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"32238885283b14552bdd563de01057b496e11a9e931d80b07f02315a53e9d29b"} Dec 01 08:40:21 crc kubenswrapper[4813]: I1201 08:40:21.557099 4813 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 08:40:21 crc kubenswrapper[4813]: I1201 08:40:21.557177 4813 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 08:40:21 crc kubenswrapper[4813]: I1201 08:40:21.557992 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:21 crc kubenswrapper[4813]: I1201 08:40:21.558026 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:21 crc kubenswrapper[4813]: I1201 08:40:21.558039 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:21 crc kubenswrapper[4813]: I1201 08:40:21.558337 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:21 crc kubenswrapper[4813]: I1201 08:40:21.558443 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:21 crc kubenswrapper[4813]: I1201 08:40:21.558474 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:21 crc kubenswrapper[4813]: I1201 08:40:21.558837 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:21 crc kubenswrapper[4813]: I1201 08:40:21.558890 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:21 crc kubenswrapper[4813]: I1201 08:40:21.558909 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:21 crc kubenswrapper[4813]: I1201 08:40:21.560101 4813 scope.go:117] "RemoveContainer" containerID="32238885283b14552bdd563de01057b496e11a9e931d80b07f02315a53e9d29b" Dec 01 08:40:21 crc kubenswrapper[4813]: I1201 08:40:21.869146 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 01 08:40:22 crc kubenswrapper[4813]: I1201 08:40:22.563208 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Dec 01 08:40:22 crc kubenswrapper[4813]: I1201 08:40:22.566331 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"a0404f5a5ed5a9750d4c479b608e68e8960b448f455d5af9d219aa5edaeb789b"} Dec 01 08:40:22 crc kubenswrapper[4813]: I1201 08:40:22.566366 4813 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 01 08:40:22 crc kubenswrapper[4813]: I1201 08:40:22.566490 4813 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 08:40:22 crc kubenswrapper[4813]: I1201 08:40:22.568068 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:22 crc kubenswrapper[4813]: I1201 08:40:22.568162 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:22 crc kubenswrapper[4813]: I1201 08:40:22.568187 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:22 crc kubenswrapper[4813]: I1201 08:40:22.573046 4813 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 08:40:22 crc kubenswrapper[4813]: I1201 08:40:22.573083 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"f6dbff3de3c8acd4bf2da1db905cfe0c066112d80eec73f698225bc7cd9a97ca"} Dec 01 08:40:22 crc kubenswrapper[4813]: I1201 08:40:22.573146 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"57dda64219336064481bbf8c3eba6acf706f700072f2b411cdae8bf9e689dc2f"} Dec 01 08:40:22 crc kubenswrapper[4813]: I1201 08:40:22.573225 4813 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 08:40:22 crc kubenswrapper[4813]: I1201 08:40:22.574317 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:22 crc kubenswrapper[4813]: I1201 08:40:22.574379 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:22 crc kubenswrapper[4813]: I1201 08:40:22.574404 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:22 crc kubenswrapper[4813]: I1201 08:40:22.574426 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:22 crc kubenswrapper[4813]: I1201 08:40:22.574452 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:22 crc kubenswrapper[4813]: I1201 08:40:22.574462 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:22 crc kubenswrapper[4813]: I1201 08:40:22.785049 4813 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 08:40:22 crc kubenswrapper[4813]: I1201 08:40:22.787130 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:22 crc kubenswrapper[4813]: I1201 08:40:22.787171 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:22 crc kubenswrapper[4813]: I1201 08:40:22.787181 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:22 crc kubenswrapper[4813]: I1201 08:40:22.787218 4813 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 01 08:40:22 crc kubenswrapper[4813]: I1201 08:40:22.940494 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-crc" Dec 01 08:40:23 crc kubenswrapper[4813]: I1201 08:40:23.184028 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 01 08:40:23 crc kubenswrapper[4813]: I1201 08:40:23.575807 4813 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 08:40:23 crc kubenswrapper[4813]: I1201 08:40:23.575895 4813 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 08:40:23 crc kubenswrapper[4813]: I1201 08:40:23.576006 4813 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 01 08:40:23 crc kubenswrapper[4813]: I1201 08:40:23.576650 4813 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 08:40:23 crc kubenswrapper[4813]: I1201 08:40:23.577900 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:23 crc kubenswrapper[4813]: I1201 08:40:23.577942 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:23 crc kubenswrapper[4813]: I1201 08:40:23.577953 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:23 crc kubenswrapper[4813]: I1201 08:40:23.578110 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:23 crc kubenswrapper[4813]: I1201 08:40:23.578208 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:23 crc kubenswrapper[4813]: I1201 08:40:23.578330 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:23 crc kubenswrapper[4813]: I1201 08:40:23.578572 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:23 crc kubenswrapper[4813]: I1201 08:40:23.578616 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:23 crc kubenswrapper[4813]: I1201 08:40:23.578641 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:23 crc kubenswrapper[4813]: I1201 08:40:23.693421 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 01 08:40:23 crc kubenswrapper[4813]: I1201 08:40:23.701930 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 01 08:40:24 crc kubenswrapper[4813]: I1201 08:40:24.578658 4813 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 08:40:24 crc kubenswrapper[4813]: I1201 08:40:24.578665 4813 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 08:40:24 crc kubenswrapper[4813]: I1201 08:40:24.580099 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:24 crc kubenswrapper[4813]: I1201 08:40:24.580146 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:24 crc kubenswrapper[4813]: I1201 08:40:24.580164 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:24 crc kubenswrapper[4813]: I1201 08:40:24.581147 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:24 crc kubenswrapper[4813]: I1201 08:40:24.581175 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:24 crc kubenswrapper[4813]: I1201 08:40:24.581189 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:24 crc kubenswrapper[4813]: I1201 08:40:24.597116 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 01 08:40:24 crc kubenswrapper[4813]: I1201 08:40:24.597261 4813 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 08:40:24 crc kubenswrapper[4813]: I1201 08:40:24.598342 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:24 crc kubenswrapper[4813]: I1201 08:40:24.598413 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:24 crc kubenswrapper[4813]: I1201 08:40:24.598441 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:24 crc kubenswrapper[4813]: I1201 08:40:24.645157 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 08:40:24 crc kubenswrapper[4813]: I1201 08:40:24.645454 4813 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 01 08:40:24 crc kubenswrapper[4813]: I1201 08:40:24.645538 4813 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 08:40:24 crc kubenswrapper[4813]: I1201 08:40:24.647663 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:24 crc kubenswrapper[4813]: I1201 08:40:24.647722 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:24 crc kubenswrapper[4813]: I1201 08:40:24.647740 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:24 crc kubenswrapper[4813]: I1201 08:40:24.870193 4813 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 01 08:40:24 crc kubenswrapper[4813]: I1201 08:40:24.870310 4813 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 01 08:40:25 crc kubenswrapper[4813]: I1201 08:40:25.253594 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc" Dec 01 08:40:25 crc kubenswrapper[4813]: I1201 08:40:25.825414 4813 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 08:40:25 crc kubenswrapper[4813]: I1201 08:40:25.825430 4813 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 08:40:25 crc kubenswrapper[4813]: I1201 08:40:25.828982 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:25 crc kubenswrapper[4813]: I1201 08:40:25.828984 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:25 crc kubenswrapper[4813]: I1201 08:40:25.829137 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:25 crc kubenswrapper[4813]: I1201 08:40:25.829167 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:25 crc kubenswrapper[4813]: I1201 08:40:25.829304 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:25 crc kubenswrapper[4813]: I1201 08:40:25.829337 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:26 crc kubenswrapper[4813]: I1201 08:40:26.438686 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 08:40:26 crc kubenswrapper[4813]: I1201 08:40:26.439713 4813 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 08:40:26 crc kubenswrapper[4813]: I1201 08:40:26.443426 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:26 crc kubenswrapper[4813]: I1201 08:40:26.443510 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:26 crc kubenswrapper[4813]: I1201 08:40:26.443530 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:26 crc kubenswrapper[4813]: E1201 08:40:26.475738 4813 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Dec 01 08:40:31 crc kubenswrapper[4813]: I1201 08:40:31.172664 4813 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="Get \"https://192.168.126.11:6443/livez\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 01 08:40:31 crc kubenswrapper[4813]: I1201 08:40:31.172926 4813 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="Get \"https://192.168.126.11:6443/livez\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 01 08:40:31 crc kubenswrapper[4813]: I1201 08:40:31.307015 4813 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": net/http: TLS handshake timeout Dec 01 08:40:32 crc kubenswrapper[4813]: E1201 08:40:32.581612 4813 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" interval="6.4s" Dec 01 08:40:32 crc kubenswrapper[4813]: E1201 08:40:32.789500 4813 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": net/http: TLS handshake timeout" node="crc" Dec 01 08:40:33 crc kubenswrapper[4813]: I1201 08:40:33.011190 4813 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Dec 01 08:40:33 crc kubenswrapper[4813]: I1201 08:40:33.011596 4813 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Dec 01 08:40:33 crc kubenswrapper[4813]: I1201 08:40:33.197887 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 01 08:40:33 crc kubenswrapper[4813]: I1201 08:40:33.198703 4813 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 08:40:33 crc kubenswrapper[4813]: I1201 08:40:33.200907 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:33 crc kubenswrapper[4813]: I1201 08:40:33.200991 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:33 crc kubenswrapper[4813]: I1201 08:40:33.201028 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:34 crc kubenswrapper[4813]: I1201 08:40:34.870522 4813 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 01 08:40:34 crc kubenswrapper[4813]: I1201 08:40:34.870696 4813 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 01 08:40:35 crc kubenswrapper[4813]: I1201 08:40:35.284883 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Dec 01 08:40:35 crc kubenswrapper[4813]: I1201 08:40:35.285431 4813 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 08:40:35 crc kubenswrapper[4813]: I1201 08:40:35.287410 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:35 crc kubenswrapper[4813]: I1201 08:40:35.287545 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:35 crc kubenswrapper[4813]: I1201 08:40:35.287572 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:35 crc kubenswrapper[4813]: I1201 08:40:35.311194 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Dec 01 08:40:35 crc kubenswrapper[4813]: I1201 08:40:35.908863 4813 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 08:40:35 crc kubenswrapper[4813]: I1201 08:40:35.910663 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:35 crc kubenswrapper[4813]: I1201 08:40:35.910717 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:35 crc kubenswrapper[4813]: I1201 08:40:35.910730 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:36 crc kubenswrapper[4813]: I1201 08:40:36.182050 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 08:40:36 crc kubenswrapper[4813]: I1201 08:40:36.182406 4813 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 08:40:36 crc kubenswrapper[4813]: I1201 08:40:36.184460 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:36 crc kubenswrapper[4813]: I1201 08:40:36.184513 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:36 crc kubenswrapper[4813]: I1201 08:40:36.184530 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:36 crc kubenswrapper[4813]: I1201 08:40:36.189994 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 08:40:36 crc kubenswrapper[4813]: E1201 08:40:36.476100 4813 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Dec 01 08:40:36 crc kubenswrapper[4813]: I1201 08:40:36.911376 4813 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 08:40:36 crc kubenswrapper[4813]: I1201 08:40:36.912586 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:36 crc kubenswrapper[4813]: I1201 08:40:36.912717 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:36 crc kubenswrapper[4813]: I1201 08:40:36.912814 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.012813 4813 trace.go:236] Trace[1251003110]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (01-Dec-2025 08:40:25.177) (total time: 12835ms): Dec 01 08:40:38 crc kubenswrapper[4813]: Trace[1251003110]: ---"Objects listed" error: 12835ms (08:40:38.012) Dec 01 08:40:38 crc kubenswrapper[4813]: Trace[1251003110]: [12.835416047s] [12.835416047s] END Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.012870 4813 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.012815 4813 trace.go:236] Trace[1378667898]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (01-Dec-2025 08:40:24.250) (total time: 13762ms): Dec 01 08:40:38 crc kubenswrapper[4813]: Trace[1378667898]: ---"Objects listed" error: 13761ms (08:40:38.012) Dec 01 08:40:38 crc kubenswrapper[4813]: Trace[1378667898]: [13.762101582s] [13.762101582s] END Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.013024 4813 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.013160 4813 trace.go:236] Trace[339160333]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (01-Dec-2025 08:40:23.763) (total time: 14249ms): Dec 01 08:40:38 crc kubenswrapper[4813]: Trace[339160333]: ---"Objects listed" error: 14249ms (08:40:38.012) Dec 01 08:40:38 crc kubenswrapper[4813]: Trace[339160333]: [14.249318242s] [14.249318242s] END Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.013197 4813 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.016795 4813 trace.go:236] Trace[853153459]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (01-Dec-2025 08:40:24.497) (total time: 13519ms): Dec 01 08:40:38 crc kubenswrapper[4813]: Trace[853153459]: ---"Objects listed" error: 13519ms (08:40:38.016) Dec 01 08:40:38 crc kubenswrapper[4813]: Trace[853153459]: [13.519270112s] [13.519270112s] END Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.016831 4813 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.017398 4813 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.076870 4813 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": EOF" start-of-body= Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.076950 4813 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": EOF" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.080251 4813 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:56270->192.168.126.11:17697: read: connection reset by peer" start-of-body= Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.080312 4813 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:56270->192.168.126.11:17697: read: connection reset by peer" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.080603 4813 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.080642 4813 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.311142 4813 apiserver.go:52] "Watching apiserver" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.313641 4813 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.314209 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-network-console/networking-console-plugin-85b44fc459-gdk6g","openshift-network-diagnostics/network-check-source-55646444c4-trplf","openshift-network-diagnostics/network-check-target-xd92c","openshift-network-node-identity/network-node-identity-vrzqb","openshift-network-operator/iptables-alerter-4ln5h","openshift-network-operator/network-operator-58b4c7f79c-55gtf"] Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.315071 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.315096 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.315177 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 08:40:38 crc kubenswrapper[4813]: E1201 08:40:38.315549 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.315623 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.315677 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 01 08:40:38 crc kubenswrapper[4813]: E1201 08:40:38.315722 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.315767 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 08:40:38 crc kubenswrapper[4813]: E1201 08:40:38.315812 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.316853 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.316882 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.317112 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.317396 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.317701 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.317835 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.318296 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.318312 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.318753 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.326282 4813 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.342807 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.353640 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.361150 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.369595 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.377680 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.386158 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.394638 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.419100 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.419155 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.419179 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.419199 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.419228 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.419250 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.419269 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.419289 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.419308 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.419329 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.419350 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.419386 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.419406 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.419426 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.419446 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.419467 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.419488 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.419511 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.419533 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.419554 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.419572 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.419594 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.419614 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.419633 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.419654 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.419676 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.419701 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.419722 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.419744 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.419768 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.419791 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.419814 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.419845 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.419887 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.419914 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.419948 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.419998 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.420022 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.420052 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.420073 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.420092 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.420141 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.420166 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.420188 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.420209 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.420233 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.420258 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.420282 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.420304 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.420328 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.420358 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.420383 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.420406 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.420429 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.420454 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.420479 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.420499 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.420685 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.420712 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.420733 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.420754 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.420774 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.420795 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.420823 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.420847 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.420873 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.420897 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.420928 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.420953 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.421001 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.421026 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.421049 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.421070 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.421090 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.421109 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.421129 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.421149 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.421169 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.421189 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.421212 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.421236 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.421263 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.421288 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.421312 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.419735 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.421348 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.421373 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.419868 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.419874 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.420148 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.420296 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.420313 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.420628 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.420811 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.420628 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.420900 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.420951 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.421207 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.421267 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.421307 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.421640 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.421664 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.421923 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.421980 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.422202 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.422216 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.421395 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.422288 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.422317 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.422342 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.422294 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.422355 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.422361 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.422357 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.422382 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.422365 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.422422 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.422482 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.422516 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.422539 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.422562 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.422586 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.422608 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.422630 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.422650 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.422672 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.422692 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.422713 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.422733 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.422752 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.422772 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.422794 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.422816 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.422848 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.422873 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.422895 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.422917 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.422948 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.422988 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.423014 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.423037 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.423058 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.423079 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.423105 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.423129 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.423153 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.423178 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.423202 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.423239 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.423262 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.423282 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.423307 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.423328 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.423351 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.423374 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.423401 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.423425 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.423448 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.423472 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.423496 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.423520 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.423540 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.423562 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.423588 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.423614 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.423638 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.423661 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.423684 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.423705 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.423728 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.423753 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.423776 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.423800 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.423821 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.423845 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.423891 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.423916 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.423937 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.423975 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.424001 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.424024 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.424052 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.424080 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.424105 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.424129 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.424153 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.424203 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.424232 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.424265 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.424290 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.424315 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.424338 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.424359 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.424381 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.424407 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.424434 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.424467 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.424493 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.424516 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.424537 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.424560 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.424584 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.424607 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.424631 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.424656 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.424678 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.424699 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.424721 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.424746 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.424769 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.424796 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.424823 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.424848 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.424871 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.424895 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.424919 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.424949 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.424992 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.425018 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.425046 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.425068 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.425151 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.425174 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.425214 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.425231 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.425250 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.425269 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.425287 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.425311 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.425329 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.425346 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.425364 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.425382 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.425402 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.425422 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.422485 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.422690 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.425491 4813 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.422831 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.425502 4813 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.425513 4813 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.425523 4813 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.425534 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.425544 4813 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.425553 4813 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.425562 4813 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.425571 4813 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.425581 4813 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.425590 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.425601 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.425610 4813 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.425620 4813 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.425629 4813 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.425641 4813 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.425654 4813 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.425666 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.425679 4813 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.425694 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.425707 4813 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.425719 4813 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.425729 4813 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.425739 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.425748 4813 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.425758 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.423068 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.423083 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.423085 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.423108 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.423132 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.425805 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.423328 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.423453 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.423459 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.423586 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.423587 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.423654 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.423740 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.423957 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.424061 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.424072 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.424109 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.424174 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.424243 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.424453 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.424575 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.424691 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.424729 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.425731 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.425905 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.425798 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.425847 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.425858 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.426086 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.426096 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.426212 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.426265 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.426367 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.426407 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.426469 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.426500 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.426504 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.426581 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.426854 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.426935 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.426997 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.427019 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.427043 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.427075 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.427224 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.427249 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.427529 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.427575 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.427599 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.427615 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.427673 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.427681 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:40:38 crc kubenswrapper[4813]: E1201 08:40:38.427775 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 08:40:38.927729659 +0000 UTC m=+23.550651245 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.427912 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.427921 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.428138 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.428347 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.428508 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.428564 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.428647 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.428672 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.428844 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.428897 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.429033 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.429054 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.429241 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.429320 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.429392 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.429482 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.429520 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.429580 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.429601 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.429513 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.429743 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.430106 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.430293 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.430306 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.430322 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.430378 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.430449 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.430619 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.430691 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.430726 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.430740 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.430732 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.430753 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.431217 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.431371 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.431591 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.431701 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.432003 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.432118 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.432143 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.432160 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:40:38 crc kubenswrapper[4813]: E1201 08:40:38.432216 4813 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 01 08:40:38 crc kubenswrapper[4813]: E1201 08:40:38.432293 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-01 08:40:38.932269669 +0000 UTC m=+23.555191335 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.432430 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.432441 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.432469 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.432654 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.432706 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.432936 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.432714 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.432730 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.432930 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.432796 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.432894 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.433492 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.433666 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.433734 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.433850 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.434135 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.434280 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.434330 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.434341 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.434386 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.434452 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.434477 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.434737 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.434833 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.434847 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.434994 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.435701 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.436272 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:40:38 crc kubenswrapper[4813]: E1201 08:40:38.436278 4813 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 01 08:40:38 crc kubenswrapper[4813]: E1201 08:40:38.436534 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-01 08:40:38.936509239 +0000 UTC m=+23.559430895 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.436532 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.436366 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.436740 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.436780 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.437093 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.437110 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.437010 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.436628 4813 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.437201 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.437234 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.437391 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.437466 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.437585 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.437664 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.437881 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.437892 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.437949 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.452084 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.452132 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.452269 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.452266 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.452789 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.452864 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.452947 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.452979 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:40:38 crc kubenswrapper[4813]: E1201 08:40:38.453267 4813 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 01 08:40:38 crc kubenswrapper[4813]: E1201 08:40:38.453290 4813 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 01 08:40:38 crc kubenswrapper[4813]: E1201 08:40:38.453316 4813 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 08:40:38 crc kubenswrapper[4813]: E1201 08:40:38.453372 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-01 08:40:38.953353809 +0000 UTC m=+23.576275395 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.453453 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.453522 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.453534 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.453647 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:40:38 crc kubenswrapper[4813]: E1201 08:40:38.453736 4813 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 01 08:40:38 crc kubenswrapper[4813]: E1201 08:40:38.453765 4813 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.453762 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:40:38 crc kubenswrapper[4813]: E1201 08:40:38.453783 4813 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 08:40:38 crc kubenswrapper[4813]: E1201 08:40:38.453845 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-01 08:40:38.953823422 +0000 UTC m=+23.576745028 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.454555 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.452569 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.455131 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.455594 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.455620 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.455727 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.455761 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.456133 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.456275 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.456474 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.456746 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.457487 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.457827 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.457884 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.457908 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.458240 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.459004 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.459766 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.460039 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.471269 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.482196 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.497537 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.503111 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.526530 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.526812 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.526945 4813 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.527058 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.527141 4813 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.527233 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.527412 4813 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.527509 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.527610 4813 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.527697 4813 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.527729 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.527721 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.527942 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.528084 4813 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.528165 4813 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.528240 4813 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.528316 4813 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.528403 4813 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.528502 4813 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.528581 4813 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.528660 4813 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.528761 4813 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.528842 4813 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.528929 4813 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.529072 4813 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.529176 4813 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.529257 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.529333 4813 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.529449 4813 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.529563 4813 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.529676 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.529790 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.529900 4813 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.530030 4813 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.530122 4813 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.530205 4813 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.530283 4813 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.530364 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.530444 4813 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.530521 4813 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.530593 4813 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.530657 4813 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.530722 4813 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.530789 4813 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.530856 4813 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.530929 4813 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.531018 4813 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.531091 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.531159 4813 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.531224 4813 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.531288 4813 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.531448 4813 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.531519 4813 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.531584 4813 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.531649 4813 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.531715 4813 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.531792 4813 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.531868 4813 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.531942 4813 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.532028 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.532097 4813 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.532161 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.532234 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.532302 4813 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.532377 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.532456 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.532528 4813 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.532602 4813 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.532693 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.532774 4813 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.532853 4813 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.532932 4813 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.533041 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.533124 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.533210 4813 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.533289 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.533368 4813 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.533447 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.533530 4813 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.533612 4813 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.533687 4813 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.533761 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.533837 4813 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.533913 4813 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.534037 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.534132 4813 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.534215 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.534315 4813 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.534460 4813 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.534578 4813 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.534689 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.534777 4813 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.534934 4813 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.535032 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.535112 4813 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.535183 4813 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.535266 4813 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.535338 4813 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.535407 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.535493 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.535570 4813 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.535639 4813 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.535712 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.535791 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.535863 4813 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.535933 4813 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.536039 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.536150 4813 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.536226 4813 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.536294 4813 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.536555 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.536623 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.536694 4813 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.536772 4813 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.536854 4813 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.536935 4813 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.537031 4813 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.537111 4813 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.537183 4813 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.537261 4813 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.537338 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.537409 4813 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.537490 4813 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.537564 4813 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.537645 4813 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.537716 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.537785 4813 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.537853 4813 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.537925 4813 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.538047 4813 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.538134 4813 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.538205 4813 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.538280 4813 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.538357 4813 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.538428 4813 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.538498 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.538580 4813 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.538658 4813 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.538731 4813 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.538804 4813 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.538887 4813 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.539000 4813 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.539077 4813 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.539158 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.539229 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.539298 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.539372 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.539449 4813 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.539520 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.539589 4813 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.539661 4813 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.539730 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.539800 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.539873 4813 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.539946 4813 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.540035 4813 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.540117 4813 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.540198 4813 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.540271 4813 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.540346 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.540440 4813 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.540519 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.540599 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.540676 4813 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.540762 4813 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.540838 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.543303 4813 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.543363 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.543416 4813 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.543469 4813 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.543535 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.543592 4813 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.631497 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.638281 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.645889 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 01 08:40:38 crc kubenswrapper[4813]: W1201 08:40:38.663265 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod37a5e44f_9a88_4405_be8a_b645485e7312.slice/crio-7df2b2842f2f42b4e2d7390cea3731487ba0056f5185a5ae315a0bb715d94b34 WatchSource:0}: Error finding container 7df2b2842f2f42b4e2d7390cea3731487ba0056f5185a5ae315a0bb715d94b34: Status 404 returned error can't find the container with id 7df2b2842f2f42b4e2d7390cea3731487ba0056f5185a5ae315a0bb715d94b34 Dec 01 08:40:38 crc kubenswrapper[4813]: W1201 08:40:38.664384 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podef543e1b_8068_4ea3_b32a_61027b32e95d.slice/crio-3ab1a445a3b80b0e7c51f816f82ab8b9a184359865877881286ad96e2e8c8e09 WatchSource:0}: Error finding container 3ab1a445a3b80b0e7c51f816f82ab8b9a184359865877881286ad96e2e8c8e09: Status 404 returned error can't find the container with id 3ab1a445a3b80b0e7c51f816f82ab8b9a184359865877881286ad96e2e8c8e09 Dec 01 08:40:38 crc kubenswrapper[4813]: W1201 08:40:38.675178 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd75a4c96_2883_4a0b_bab2_0fab2b6c0b49.slice/crio-8ad822d04293c8fceffdab9990d32e3dfd900c77dd78ec544c49473958e86dd6 WatchSource:0}: Error finding container 8ad822d04293c8fceffdab9990d32e3dfd900c77dd78ec544c49473958e86dd6: Status 404 returned error can't find the container with id 8ad822d04293c8fceffdab9990d32e3dfd900c77dd78ec544c49473958e86dd6 Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.925239 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"8ad822d04293c8fceffdab9990d32e3dfd900c77dd78ec544c49473958e86dd6"} Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.927380 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"3264b3af996af452df2953c59bea6deeb5ea6af1b119ce89f7334595c27e70da"} Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.927443 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"b9e879b1f563bd66383d58fd7b435a8a81bd50cb387dc1fabb6fc47d5a3fa468"} Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.927454 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"3ab1a445a3b80b0e7c51f816f82ab8b9a184359865877881286ad96e2e8c8e09"} Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.929066 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"0023b0e358d75aacc889ce364846dda89ea513832c87fa4b066aea7dac57e8a2"} Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.929091 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"7df2b2842f2f42b4e2d7390cea3731487ba0056f5185a5ae315a0bb715d94b34"} Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.931683 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.932107 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.933605 4813 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="a0404f5a5ed5a9750d4c479b608e68e8960b448f455d5af9d219aa5edaeb789b" exitCode=255 Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.933661 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"a0404f5a5ed5a9750d4c479b608e68e8960b448f455d5af9d219aa5edaeb789b"} Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.933756 4813 scope.go:117] "RemoveContainer" containerID="32238885283b14552bdd563de01057b496e11a9e931d80b07f02315a53e9d29b" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.938581 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.944653 4813 scope.go:117] "RemoveContainer" containerID="a0404f5a5ed5a9750d4c479b608e68e8960b448f455d5af9d219aa5edaeb789b" Dec 01 08:40:38 crc kubenswrapper[4813]: E1201 08:40:38.944897 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.949951 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 08:40:38 crc kubenswrapper[4813]: E1201 08:40:38.950141 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 08:40:39.950104195 +0000 UTC m=+24.573025781 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.950207 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.950280 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 08:40:38 crc kubenswrapper[4813]: E1201 08:40:38.950536 4813 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 01 08:40:38 crc kubenswrapper[4813]: E1201 08:40:38.950631 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-01 08:40:39.950612569 +0000 UTC m=+24.573534235 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 01 08:40:38 crc kubenswrapper[4813]: E1201 08:40:38.950647 4813 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 01 08:40:38 crc kubenswrapper[4813]: E1201 08:40:38.950699 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-01 08:40:39.950682981 +0000 UTC m=+24.573604567 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.951240 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.951737 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.961530 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.970208 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 01 08:40:38 crc kubenswrapper[4813]: I1201 08:40:38.979802 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 01 08:40:39 crc kubenswrapper[4813]: I1201 08:40:39.001649 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3264b3af996af452df2953c59bea6deeb5ea6af1b119ce89f7334595c27e70da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9e879b1f563bd66383d58fd7b435a8a81bd50cb387dc1fabb6fc47d5a3fa468\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 01 08:40:39 crc kubenswrapper[4813]: I1201 08:40:39.020723 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0023b0e358d75aacc889ce364846dda89ea513832c87fa4b066aea7dac57e8a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 01 08:40:39 crc kubenswrapper[4813]: I1201 08:40:39.035801 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 01 08:40:39 crc kubenswrapper[4813]: I1201 08:40:39.044794 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 01 08:40:39 crc kubenswrapper[4813]: I1201 08:40:39.051182 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 08:40:39 crc kubenswrapper[4813]: I1201 08:40:39.051257 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 08:40:39 crc kubenswrapper[4813]: E1201 08:40:39.051369 4813 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 01 08:40:39 crc kubenswrapper[4813]: E1201 08:40:39.051409 4813 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 01 08:40:39 crc kubenswrapper[4813]: E1201 08:40:39.051425 4813 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 08:40:39 crc kubenswrapper[4813]: E1201 08:40:39.051433 4813 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 01 08:40:39 crc kubenswrapper[4813]: E1201 08:40:39.051470 4813 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 01 08:40:39 crc kubenswrapper[4813]: E1201 08:40:39.051484 4813 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 08:40:39 crc kubenswrapper[4813]: E1201 08:40:39.051497 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-01 08:40:40.051475402 +0000 UTC m=+24.674397038 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 08:40:39 crc kubenswrapper[4813]: E1201 08:40:39.051562 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-01 08:40:40.051522403 +0000 UTC m=+24.674444049 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 08:40:39 crc kubenswrapper[4813]: I1201 08:40:39.054864 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 01 08:40:39 crc kubenswrapper[4813]: I1201 08:40:39.066788 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"349332bd-6b4d-4306-84ae-07019ca3860f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d441a348087ae3cadda1ad8a7d1c4dff02cf3f6e1d346c294b278fdd9131edcc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://da67635daf6d9e32937f4d5d4ceeee3f8f09ae4fd2dd26c3e85be1286c5949de\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e97356ee274ea78d9f237ffbc756abb6ac8df0dbac244bf0a86ed215b8cbf43\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0404f5a5ed5a9750d4c479b608e68e8960b448f455d5af9d219aa5edaeb789b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://32238885283b14552bdd563de01057b496e11a9e931d80b07f02315a53e9d29b\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T08:40:20Z\\\",\\\"message\\\":\\\"W1201 08:40:20.327704 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1201 08:40:20.328141 1 crypto.go:601] Generating new CA for check-endpoints-signer@1764578420 cert, and key in /tmp/serving-cert-2947015701/serving-signer.crt, /tmp/serving-cert-2947015701/serving-signer.key\\\\nI1201 08:40:20.479057 1 observer_polling.go:159] Starting file observer\\\\nW1201 08:40:20.485329 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1201 08:40:20.485511 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1201 08:40:20.492359 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2947015701/tls.crt::/tmp/serving-cert-2947015701/tls.key\\\\\\\"\\\\nF1201 08:40:20.945612 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:20Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a0404f5a5ed5a9750d4c479b608e68e8960b448f455d5af9d219aa5edaeb789b\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"message\\\":\\\"\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764578421\\\\\\\\\\\\\\\" (2025-12-01 08:40:21 +0000 UTC to 2025-12-31 08:40:22 +0000 UTC (now=2025-12-01 08:40:38.066068451 +0000 UTC))\\\\\\\"\\\\nI1201 08:40:38.066144 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1201 08:40:38.066166 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1201 08:40:38.066220 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1201 08:40:38.066232 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1201 08:40:38.066338 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764578432\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764578432\\\\\\\\\\\\\\\" (2025-12-01 07:40:32 +0000 UTC to 2026-12-01 07:40:32 +0000 UTC (now=2025-12-01 08:40:38.066301217 +0000 UTC))\\\\\\\"\\\\nI1201 08:40:38.066357 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1201 08:40:38.066385 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1201 08:40:38.066402 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1201 08:40:38.066414 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-63890811/tls.crt::/tmp/serving-cert-63890811/tls.key\\\\\\\"\\\\nI1201 08:40:38.066419 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1201 08:40:38.066572 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1201 08:40:38.067159 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7354f9337d42115bb1d974bf3772f2d03365acba57f2155135c7a0bbab39879e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:19Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05492df5703392cdf0932d6bf66025b18e53bd05fa0857a7cb2ad70979cbdc3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05492df5703392cdf0932d6bf66025b18e53bd05fa0857a7cb2ad70979cbdc3e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:16Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 01 08:40:39 crc kubenswrapper[4813]: I1201 08:40:39.075649 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 01 08:40:39 crc kubenswrapper[4813]: I1201 08:40:39.088172 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3264b3af996af452df2953c59bea6deeb5ea6af1b119ce89f7334595c27e70da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9e879b1f563bd66383d58fd7b435a8a81bd50cb387dc1fabb6fc47d5a3fa468\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 01 08:40:39 crc kubenswrapper[4813]: I1201 08:40:39.190034 4813 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 08:40:39 crc kubenswrapper[4813]: I1201 08:40:39.192027 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:39 crc kubenswrapper[4813]: I1201 08:40:39.192071 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:39 crc kubenswrapper[4813]: I1201 08:40:39.192084 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:39 crc kubenswrapper[4813]: I1201 08:40:39.192162 4813 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 01 08:40:39 crc kubenswrapper[4813]: I1201 08:40:39.202469 4813 kubelet_node_status.go:115] "Node was previously registered" node="crc" Dec 01 08:40:39 crc kubenswrapper[4813]: I1201 08:40:39.202840 4813 kubelet_node_status.go:79] "Successfully registered node" node="crc" Dec 01 08:40:39 crc kubenswrapper[4813]: I1201 08:40:39.204141 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:39 crc kubenswrapper[4813]: I1201 08:40:39.204179 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:39 crc kubenswrapper[4813]: I1201 08:40:39.204229 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:39 crc kubenswrapper[4813]: I1201 08:40:39.204246 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:39 crc kubenswrapper[4813]: I1201 08:40:39.204269 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:39Z","lastTransitionTime":"2025-12-01T08:40:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:39 crc kubenswrapper[4813]: E1201 08:40:39.220310 4813 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:40:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:40:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:39Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:40:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:40:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:39Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cd271ba1-a5c5-44e0-b03e-cfd1f5e8cb76\\\",\\\"systemUUID\\\":\\\"906e3783-eb6d-4f21-889c-7989aeaa25ba\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 01 08:40:39 crc kubenswrapper[4813]: I1201 08:40:39.224828 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:39 crc kubenswrapper[4813]: I1201 08:40:39.224894 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:39 crc kubenswrapper[4813]: I1201 08:40:39.224911 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:39 crc kubenswrapper[4813]: I1201 08:40:39.224933 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:39 crc kubenswrapper[4813]: I1201 08:40:39.224948 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:39Z","lastTransitionTime":"2025-12-01T08:40:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:39 crc kubenswrapper[4813]: E1201 08:40:39.240135 4813 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:40:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:40:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:39Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:40:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:40:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:39Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cd271ba1-a5c5-44e0-b03e-cfd1f5e8cb76\\\",\\\"systemUUID\\\":\\\"906e3783-eb6d-4f21-889c-7989aeaa25ba\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 01 08:40:39 crc kubenswrapper[4813]: I1201 08:40:39.248766 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:39 crc kubenswrapper[4813]: I1201 08:40:39.248817 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:39 crc kubenswrapper[4813]: I1201 08:40:39.248826 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:39 crc kubenswrapper[4813]: I1201 08:40:39.248845 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:39 crc kubenswrapper[4813]: I1201 08:40:39.248855 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:39Z","lastTransitionTime":"2025-12-01T08:40:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:39 crc kubenswrapper[4813]: E1201 08:40:39.262052 4813 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:40:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:40:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:39Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:40:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:40:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:39Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cd271ba1-a5c5-44e0-b03e-cfd1f5e8cb76\\\",\\\"systemUUID\\\":\\\"906e3783-eb6d-4f21-889c-7989aeaa25ba\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 01 08:40:39 crc kubenswrapper[4813]: I1201 08:40:39.265522 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:39 crc kubenswrapper[4813]: I1201 08:40:39.265578 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:39 crc kubenswrapper[4813]: I1201 08:40:39.265590 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:39 crc kubenswrapper[4813]: I1201 08:40:39.265640 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:39 crc kubenswrapper[4813]: I1201 08:40:39.265654 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:39Z","lastTransitionTime":"2025-12-01T08:40:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:39 crc kubenswrapper[4813]: E1201 08:40:39.275934 4813 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:40:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:40:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:39Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:40:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:40:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:39Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cd271ba1-a5c5-44e0-b03e-cfd1f5e8cb76\\\",\\\"systemUUID\\\":\\\"906e3783-eb6d-4f21-889c-7989aeaa25ba\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 01 08:40:39 crc kubenswrapper[4813]: I1201 08:40:39.279940 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:39 crc kubenswrapper[4813]: I1201 08:40:39.279991 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:39 crc kubenswrapper[4813]: I1201 08:40:39.279999 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:39 crc kubenswrapper[4813]: I1201 08:40:39.280014 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:39 crc kubenswrapper[4813]: I1201 08:40:39.280023 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:39Z","lastTransitionTime":"2025-12-01T08:40:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:39 crc kubenswrapper[4813]: E1201 08:40:39.289068 4813 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:40:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:40:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:39Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:40:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:40:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:39Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cd271ba1-a5c5-44e0-b03e-cfd1f5e8cb76\\\",\\\"systemUUID\\\":\\\"906e3783-eb6d-4f21-889c-7989aeaa25ba\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 01 08:40:39 crc kubenswrapper[4813]: E1201 08:40:39.289340 4813 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 01 08:40:39 crc kubenswrapper[4813]: I1201 08:40:39.290924 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:39 crc kubenswrapper[4813]: I1201 08:40:39.291000 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:39 crc kubenswrapper[4813]: I1201 08:40:39.291046 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:39 crc kubenswrapper[4813]: I1201 08:40:39.291073 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:39 crc kubenswrapper[4813]: I1201 08:40:39.291092 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:39Z","lastTransitionTime":"2025-12-01T08:40:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:39 crc kubenswrapper[4813]: I1201 08:40:39.393994 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:39 crc kubenswrapper[4813]: I1201 08:40:39.394030 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:39 crc kubenswrapper[4813]: I1201 08:40:39.394042 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:39 crc kubenswrapper[4813]: I1201 08:40:39.394058 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:39 crc kubenswrapper[4813]: I1201 08:40:39.394070 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:39Z","lastTransitionTime":"2025-12-01T08:40:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:39 crc kubenswrapper[4813]: I1201 08:40:39.497374 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:39 crc kubenswrapper[4813]: I1201 08:40:39.497419 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:39 crc kubenswrapper[4813]: I1201 08:40:39.497431 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:39 crc kubenswrapper[4813]: I1201 08:40:39.497450 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:39 crc kubenswrapper[4813]: I1201 08:40:39.497461 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:39Z","lastTransitionTime":"2025-12-01T08:40:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:39 crc kubenswrapper[4813]: I1201 08:40:39.600421 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:39 crc kubenswrapper[4813]: I1201 08:40:39.600458 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:39 crc kubenswrapper[4813]: I1201 08:40:39.600469 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:39 crc kubenswrapper[4813]: I1201 08:40:39.600485 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:39 crc kubenswrapper[4813]: I1201 08:40:39.600498 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:39Z","lastTransitionTime":"2025-12-01T08:40:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:39 crc kubenswrapper[4813]: I1201 08:40:39.703185 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:39 crc kubenswrapper[4813]: I1201 08:40:39.703227 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:39 crc kubenswrapper[4813]: I1201 08:40:39.703237 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:39 crc kubenswrapper[4813]: I1201 08:40:39.703253 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:39 crc kubenswrapper[4813]: I1201 08:40:39.703262 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:39Z","lastTransitionTime":"2025-12-01T08:40:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:39 crc kubenswrapper[4813]: I1201 08:40:39.805230 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:39 crc kubenswrapper[4813]: I1201 08:40:39.805271 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:39 crc kubenswrapper[4813]: I1201 08:40:39.805284 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:39 crc kubenswrapper[4813]: I1201 08:40:39.805301 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:39 crc kubenswrapper[4813]: I1201 08:40:39.805314 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:39Z","lastTransitionTime":"2025-12-01T08:40:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:39 crc kubenswrapper[4813]: I1201 08:40:39.907751 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:39 crc kubenswrapper[4813]: I1201 08:40:39.907805 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:39 crc kubenswrapper[4813]: I1201 08:40:39.907821 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:39 crc kubenswrapper[4813]: I1201 08:40:39.907845 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:39 crc kubenswrapper[4813]: I1201 08:40:39.907859 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:39Z","lastTransitionTime":"2025-12-01T08:40:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:39 crc kubenswrapper[4813]: I1201 08:40:39.937284 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Dec 01 08:40:39 crc kubenswrapper[4813]: I1201 08:40:39.939615 4813 scope.go:117] "RemoveContainer" containerID="a0404f5a5ed5a9750d4c479b608e68e8960b448f455d5af9d219aa5edaeb789b" Dec 01 08:40:39 crc kubenswrapper[4813]: E1201 08:40:39.939749 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Dec 01 08:40:39 crc kubenswrapper[4813]: I1201 08:40:39.958358 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:39Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:39 crc kubenswrapper[4813]: I1201 08:40:39.959499 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 08:40:39 crc kubenswrapper[4813]: I1201 08:40:39.959585 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 08:40:39 crc kubenswrapper[4813]: E1201 08:40:39.959655 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 08:40:41.959633143 +0000 UTC m=+26.582554729 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:40:39 crc kubenswrapper[4813]: E1201 08:40:39.959709 4813 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 01 08:40:39 crc kubenswrapper[4813]: E1201 08:40:39.959756 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-01 08:40:41.959745176 +0000 UTC m=+26.582666762 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 01 08:40:39 crc kubenswrapper[4813]: E1201 08:40:39.959755 4813 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 01 08:40:39 crc kubenswrapper[4813]: I1201 08:40:39.959710 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 08:40:39 crc kubenswrapper[4813]: E1201 08:40:39.959801 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-01 08:40:41.959791097 +0000 UTC m=+26.582712683 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 01 08:40:39 crc kubenswrapper[4813]: I1201 08:40:39.979432 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3264b3af996af452df2953c59bea6deeb5ea6af1b119ce89f7334595c27e70da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9e879b1f563bd66383d58fd7b435a8a81bd50cb387dc1fabb6fc47d5a3fa468\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:39Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:39 crc kubenswrapper[4813]: I1201 08:40:39.996448 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"349332bd-6b4d-4306-84ae-07019ca3860f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d441a348087ae3cadda1ad8a7d1c4dff02cf3f6e1d346c294b278fdd9131edcc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://da67635daf6d9e32937f4d5d4ceeee3f8f09ae4fd2dd26c3e85be1286c5949de\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e97356ee274ea78d9f237ffbc756abb6ac8df0dbac244bf0a86ed215b8cbf43\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0404f5a5ed5a9750d4c479b608e68e8960b448f455d5af9d219aa5edaeb789b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a0404f5a5ed5a9750d4c479b608e68e8960b448f455d5af9d219aa5edaeb789b\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"message\\\":\\\"\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764578421\\\\\\\\\\\\\\\" (2025-12-01 08:40:21 +0000 UTC to 2025-12-31 08:40:22 +0000 UTC (now=2025-12-01 08:40:38.066068451 +0000 UTC))\\\\\\\"\\\\nI1201 08:40:38.066144 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1201 08:40:38.066166 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1201 08:40:38.066220 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1201 08:40:38.066232 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1201 08:40:38.066338 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764578432\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764578432\\\\\\\\\\\\\\\" (2025-12-01 07:40:32 +0000 UTC to 2026-12-01 07:40:32 +0000 UTC (now=2025-12-01 08:40:38.066301217 +0000 UTC))\\\\\\\"\\\\nI1201 08:40:38.066357 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1201 08:40:38.066385 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1201 08:40:38.066402 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1201 08:40:38.066414 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-63890811/tls.crt::/tmp/serving-cert-63890811/tls.key\\\\\\\"\\\\nI1201 08:40:38.066419 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1201 08:40:38.066572 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1201 08:40:38.067159 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7354f9337d42115bb1d974bf3772f2d03365acba57f2155135c7a0bbab39879e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:19Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05492df5703392cdf0932d6bf66025b18e53bd05fa0857a7cb2ad70979cbdc3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05492df5703392cdf0932d6bf66025b18e53bd05fa0857a7cb2ad70979cbdc3e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:16Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:39Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:40 crc kubenswrapper[4813]: I1201 08:40:40.010386 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:40 crc kubenswrapper[4813]: I1201 08:40:40.010418 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:40 crc kubenswrapper[4813]: I1201 08:40:40.010426 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:40 crc kubenswrapper[4813]: I1201 08:40:40.010444 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:40 crc kubenswrapper[4813]: I1201 08:40:40.010453 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:40Z","lastTransitionTime":"2025-12-01T08:40:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:40 crc kubenswrapper[4813]: I1201 08:40:40.012671 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:40Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:40 crc kubenswrapper[4813]: I1201 08:40:40.045463 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:40Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:40 crc kubenswrapper[4813]: I1201 08:40:40.065359 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 08:40:40 crc kubenswrapper[4813]: I1201 08:40:40.065413 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 08:40:40 crc kubenswrapper[4813]: E1201 08:40:40.065541 4813 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 01 08:40:40 crc kubenswrapper[4813]: E1201 08:40:40.065566 4813 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 01 08:40:40 crc kubenswrapper[4813]: E1201 08:40:40.065568 4813 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 01 08:40:40 crc kubenswrapper[4813]: E1201 08:40:40.065580 4813 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 08:40:40 crc kubenswrapper[4813]: E1201 08:40:40.065591 4813 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 01 08:40:40 crc kubenswrapper[4813]: E1201 08:40:40.065603 4813 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 08:40:40 crc kubenswrapper[4813]: E1201 08:40:40.065646 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-01 08:40:42.06562788 +0000 UTC m=+26.688549466 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 08:40:40 crc kubenswrapper[4813]: E1201 08:40:40.065665 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-01 08:40:42.065657611 +0000 UTC m=+26.688579197 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 08:40:40 crc kubenswrapper[4813]: I1201 08:40:40.081200 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:40Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:40 crc kubenswrapper[4813]: I1201 08:40:40.098841 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0023b0e358d75aacc889ce364846dda89ea513832c87fa4b066aea7dac57e8a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:40Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:40 crc kubenswrapper[4813]: I1201 08:40:40.112567 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:40 crc kubenswrapper[4813]: I1201 08:40:40.112616 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:40 crc kubenswrapper[4813]: I1201 08:40:40.112629 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:40 crc kubenswrapper[4813]: I1201 08:40:40.112649 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:40 crc kubenswrapper[4813]: I1201 08:40:40.112661 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:40Z","lastTransitionTime":"2025-12-01T08:40:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:40 crc kubenswrapper[4813]: I1201 08:40:40.216754 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:40 crc kubenswrapper[4813]: I1201 08:40:40.216793 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:40 crc kubenswrapper[4813]: I1201 08:40:40.216802 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:40 crc kubenswrapper[4813]: I1201 08:40:40.216819 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:40 crc kubenswrapper[4813]: I1201 08:40:40.216829 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:40Z","lastTransitionTime":"2025-12-01T08:40:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:40 crc kubenswrapper[4813]: I1201 08:40:40.328618 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:40 crc kubenswrapper[4813]: I1201 08:40:40.328671 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:40 crc kubenswrapper[4813]: I1201 08:40:40.328684 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:40 crc kubenswrapper[4813]: I1201 08:40:40.328703 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:40 crc kubenswrapper[4813]: I1201 08:40:40.328916 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:40Z","lastTransitionTime":"2025-12-01T08:40:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:40 crc kubenswrapper[4813]: I1201 08:40:40.469204 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 08:40:40 crc kubenswrapper[4813]: E1201 08:40:40.469468 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 08:40:40 crc kubenswrapper[4813]: I1201 08:40:40.469237 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 08:40:40 crc kubenswrapper[4813]: E1201 08:40:40.469647 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 08:40:40 crc kubenswrapper[4813]: I1201 08:40:40.469222 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 08:40:40 crc kubenswrapper[4813]: E1201 08:40:40.469832 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 08:40:40 crc kubenswrapper[4813]: I1201 08:40:40.488933 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:40 crc kubenswrapper[4813]: I1201 08:40:40.489068 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:40 crc kubenswrapper[4813]: I1201 08:40:40.489148 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:40 crc kubenswrapper[4813]: I1201 08:40:40.489175 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:40 crc kubenswrapper[4813]: I1201 08:40:40.489277 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:40Z","lastTransitionTime":"2025-12-01T08:40:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:40 crc kubenswrapper[4813]: I1201 08:40:40.489074 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Dec 01 08:40:40 crc kubenswrapper[4813]: I1201 08:40:40.490420 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Dec 01 08:40:40 crc kubenswrapper[4813]: I1201 08:40:40.491577 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Dec 01 08:40:40 crc kubenswrapper[4813]: I1201 08:40:40.492592 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Dec 01 08:40:40 crc kubenswrapper[4813]: I1201 08:40:40.493404 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Dec 01 08:40:40 crc kubenswrapper[4813]: I1201 08:40:40.495506 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Dec 01 08:40:40 crc kubenswrapper[4813]: I1201 08:40:40.496483 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Dec 01 08:40:40 crc kubenswrapper[4813]: I1201 08:40:40.497725 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Dec 01 08:40:40 crc kubenswrapper[4813]: I1201 08:40:40.499142 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Dec 01 08:40:40 crc kubenswrapper[4813]: I1201 08:40:40.499896 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Dec 01 08:40:40 crc kubenswrapper[4813]: I1201 08:40:40.501156 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Dec 01 08:40:40 crc kubenswrapper[4813]: I1201 08:40:40.502075 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Dec 01 08:40:40 crc kubenswrapper[4813]: I1201 08:40:40.503268 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Dec 01 08:40:40 crc kubenswrapper[4813]: I1201 08:40:40.504160 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Dec 01 08:40:40 crc kubenswrapper[4813]: I1201 08:40:40.505370 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Dec 01 08:40:40 crc kubenswrapper[4813]: I1201 08:40:40.507550 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Dec 01 08:40:40 crc kubenswrapper[4813]: I1201 08:40:40.509330 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Dec 01 08:40:40 crc kubenswrapper[4813]: I1201 08:40:40.510117 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Dec 01 08:40:40 crc kubenswrapper[4813]: I1201 08:40:40.512077 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Dec 01 08:40:40 crc kubenswrapper[4813]: I1201 08:40:40.514041 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Dec 01 08:40:40 crc kubenswrapper[4813]: I1201 08:40:40.514932 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Dec 01 08:40:40 crc kubenswrapper[4813]: I1201 08:40:40.516960 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Dec 01 08:40:40 crc kubenswrapper[4813]: I1201 08:40:40.517868 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Dec 01 08:40:40 crc kubenswrapper[4813]: I1201 08:40:40.520926 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Dec 01 08:40:40 crc kubenswrapper[4813]: I1201 08:40:40.527404 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Dec 01 08:40:40 crc kubenswrapper[4813]: I1201 08:40:40.530047 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Dec 01 08:40:40 crc kubenswrapper[4813]: I1201 08:40:40.531386 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Dec 01 08:40:40 crc kubenswrapper[4813]: I1201 08:40:40.535159 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Dec 01 08:40:40 crc kubenswrapper[4813]: I1201 08:40:40.537214 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Dec 01 08:40:40 crc kubenswrapper[4813]: I1201 08:40:40.539432 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Dec 01 08:40:40 crc kubenswrapper[4813]: I1201 08:40:40.540072 4813 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Dec 01 08:40:40 crc kubenswrapper[4813]: I1201 08:40:40.540181 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Dec 01 08:40:40 crc kubenswrapper[4813]: I1201 08:40:40.541791 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Dec 01 08:40:40 crc kubenswrapper[4813]: I1201 08:40:40.542353 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Dec 01 08:40:40 crc kubenswrapper[4813]: I1201 08:40:40.542743 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Dec 01 08:40:40 crc kubenswrapper[4813]: I1201 08:40:40.545422 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Dec 01 08:40:40 crc kubenswrapper[4813]: I1201 08:40:40.546150 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Dec 01 08:40:40 crc kubenswrapper[4813]: I1201 08:40:40.547060 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Dec 01 08:40:40 crc kubenswrapper[4813]: I1201 08:40:40.547680 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Dec 01 08:40:40 crc kubenswrapper[4813]: I1201 08:40:40.548740 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Dec 01 08:40:40 crc kubenswrapper[4813]: I1201 08:40:40.549262 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Dec 01 08:40:40 crc kubenswrapper[4813]: I1201 08:40:40.549831 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Dec 01 08:40:40 crc kubenswrapper[4813]: I1201 08:40:40.550916 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Dec 01 08:40:40 crc kubenswrapper[4813]: I1201 08:40:40.552141 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Dec 01 08:40:40 crc kubenswrapper[4813]: I1201 08:40:40.552742 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Dec 01 08:40:40 crc kubenswrapper[4813]: I1201 08:40:40.553887 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Dec 01 08:40:40 crc kubenswrapper[4813]: I1201 08:40:40.554511 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Dec 01 08:40:40 crc kubenswrapper[4813]: I1201 08:40:40.555705 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Dec 01 08:40:40 crc kubenswrapper[4813]: I1201 08:40:40.556241 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Dec 01 08:40:40 crc kubenswrapper[4813]: I1201 08:40:40.557351 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Dec 01 08:40:40 crc kubenswrapper[4813]: I1201 08:40:40.557856 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Dec 01 08:40:40 crc kubenswrapper[4813]: I1201 08:40:40.558387 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Dec 01 08:40:40 crc kubenswrapper[4813]: I1201 08:40:40.560523 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Dec 01 08:40:40 crc kubenswrapper[4813]: I1201 08:40:40.561086 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Dec 01 08:40:40 crc kubenswrapper[4813]: I1201 08:40:40.561553 4813 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 08:40:40 crc kubenswrapper[4813]: I1201 08:40:40.591374 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:40 crc kubenswrapper[4813]: I1201 08:40:40.591415 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:40 crc kubenswrapper[4813]: I1201 08:40:40.591428 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:40 crc kubenswrapper[4813]: I1201 08:40:40.591446 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:40 crc kubenswrapper[4813]: I1201 08:40:40.591458 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:40Z","lastTransitionTime":"2025-12-01T08:40:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:40 crc kubenswrapper[4813]: I1201 08:40:40.694037 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:40 crc kubenswrapper[4813]: I1201 08:40:40.694077 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:40 crc kubenswrapper[4813]: I1201 08:40:40.694086 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:40 crc kubenswrapper[4813]: I1201 08:40:40.694101 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:40 crc kubenswrapper[4813]: I1201 08:40:40.694110 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:40Z","lastTransitionTime":"2025-12-01T08:40:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:40 crc kubenswrapper[4813]: I1201 08:40:40.796234 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:40 crc kubenswrapper[4813]: I1201 08:40:40.796270 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:40 crc kubenswrapper[4813]: I1201 08:40:40.796281 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:40 crc kubenswrapper[4813]: I1201 08:40:40.796296 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:40 crc kubenswrapper[4813]: I1201 08:40:40.796307 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:40Z","lastTransitionTime":"2025-12-01T08:40:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:40 crc kubenswrapper[4813]: I1201 08:40:40.898695 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:40 crc kubenswrapper[4813]: I1201 08:40:40.898735 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:40 crc kubenswrapper[4813]: I1201 08:40:40.898743 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:40 crc kubenswrapper[4813]: I1201 08:40:40.898759 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:40 crc kubenswrapper[4813]: I1201 08:40:40.898769 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:40Z","lastTransitionTime":"2025-12-01T08:40:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:40 crc kubenswrapper[4813]: I1201 08:40:40.942227 4813 scope.go:117] "RemoveContainer" containerID="a0404f5a5ed5a9750d4c479b608e68e8960b448f455d5af9d219aa5edaeb789b" Dec 01 08:40:40 crc kubenswrapper[4813]: E1201 08:40:40.942367 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.001038 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.001100 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.001116 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.001140 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.001155 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:41Z","lastTransitionTime":"2025-12-01T08:40:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.104032 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.104076 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.104089 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.104108 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.104119 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:41Z","lastTransitionTime":"2025-12-01T08:40:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.164891 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/node-resolver-c576v"] Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.165290 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-c576v" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.167456 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.169519 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.170616 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.180356 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/17292cd0-1eac-40b5-a60e-67167c57409c-hosts-file\") pod \"node-resolver-c576v\" (UID: \"17292cd0-1eac-40b5-a60e-67167c57409c\") " pod="openshift-dns/node-resolver-c576v" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.180536 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7zwfx\" (UniqueName: \"kubernetes.io/projected/17292cd0-1eac-40b5-a60e-67167c57409c-kube-api-access-7zwfx\") pod \"node-resolver-c576v\" (UID: \"17292cd0-1eac-40b5-a60e-67167c57409c\") " pod="openshift-dns/node-resolver-c576v" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.183730 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0023b0e358d75aacc889ce364846dda89ea513832c87fa4b066aea7dac57e8a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:41Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.197404 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:41Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.206124 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.206307 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.206394 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.206471 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.206557 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:41Z","lastTransitionTime":"2025-12-01T08:40:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.211486 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:41Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.222599 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:41Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.235260 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"349332bd-6b4d-4306-84ae-07019ca3860f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d441a348087ae3cadda1ad8a7d1c4dff02cf3f6e1d346c294b278fdd9131edcc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://da67635daf6d9e32937f4d5d4ceeee3f8f09ae4fd2dd26c3e85be1286c5949de\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e97356ee274ea78d9f237ffbc756abb6ac8df0dbac244bf0a86ed215b8cbf43\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0404f5a5ed5a9750d4c479b608e68e8960b448f455d5af9d219aa5edaeb789b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a0404f5a5ed5a9750d4c479b608e68e8960b448f455d5af9d219aa5edaeb789b\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"message\\\":\\\"\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764578421\\\\\\\\\\\\\\\" (2025-12-01 08:40:21 +0000 UTC to 2025-12-31 08:40:22 +0000 UTC (now=2025-12-01 08:40:38.066068451 +0000 UTC))\\\\\\\"\\\\nI1201 08:40:38.066144 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1201 08:40:38.066166 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1201 08:40:38.066220 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1201 08:40:38.066232 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1201 08:40:38.066338 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764578432\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764578432\\\\\\\\\\\\\\\" (2025-12-01 07:40:32 +0000 UTC to 2026-12-01 07:40:32 +0000 UTC (now=2025-12-01 08:40:38.066301217 +0000 UTC))\\\\\\\"\\\\nI1201 08:40:38.066357 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1201 08:40:38.066385 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1201 08:40:38.066402 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1201 08:40:38.066414 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-63890811/tls.crt::/tmp/serving-cert-63890811/tls.key\\\\\\\"\\\\nI1201 08:40:38.066419 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1201 08:40:38.066572 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1201 08:40:38.067159 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7354f9337d42115bb1d974bf3772f2d03365acba57f2155135c7a0bbab39879e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:19Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05492df5703392cdf0932d6bf66025b18e53bd05fa0857a7cb2ad70979cbdc3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05492df5703392cdf0932d6bf66025b18e53bd05fa0857a7cb2ad70979cbdc3e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:16Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:41Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.248576 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:41Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.260375 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3264b3af996af452df2953c59bea6deeb5ea6af1b119ce89f7334595c27e70da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9e879b1f563bd66383d58fd7b435a8a81bd50cb387dc1fabb6fc47d5a3fa468\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:41Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.271267 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-c576v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"17292cd0-1eac-40b5-a60e-67167c57409c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7zwfx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:41Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-c576v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:41Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.281700 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/17292cd0-1eac-40b5-a60e-67167c57409c-hosts-file\") pod \"node-resolver-c576v\" (UID: \"17292cd0-1eac-40b5-a60e-67167c57409c\") " pod="openshift-dns/node-resolver-c576v" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.281884 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/17292cd0-1eac-40b5-a60e-67167c57409c-hosts-file\") pod \"node-resolver-c576v\" (UID: \"17292cd0-1eac-40b5-a60e-67167c57409c\") " pod="openshift-dns/node-resolver-c576v" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.281928 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7zwfx\" (UniqueName: \"kubernetes.io/projected/17292cd0-1eac-40b5-a60e-67167c57409c-kube-api-access-7zwfx\") pod \"node-resolver-c576v\" (UID: \"17292cd0-1eac-40b5-a60e-67167c57409c\") " pod="openshift-dns/node-resolver-c576v" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.300777 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7zwfx\" (UniqueName: \"kubernetes.io/projected/17292cd0-1eac-40b5-a60e-67167c57409c-kube-api-access-7zwfx\") pod \"node-resolver-c576v\" (UID: \"17292cd0-1eac-40b5-a60e-67167c57409c\") " pod="openshift-dns/node-resolver-c576v" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.309123 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.309188 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.309200 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.309217 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.309228 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:41Z","lastTransitionTime":"2025-12-01T08:40:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.412245 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.412294 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.412308 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.412325 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.412337 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:41Z","lastTransitionTime":"2025-12-01T08:40:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.477424 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-c576v" Dec 01 08:40:41 crc kubenswrapper[4813]: W1201 08:40:41.490125 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod17292cd0_1eac_40b5_a60e_67167c57409c.slice/crio-c55556334ef2df449eec30be9ed290f08cf5de74b42db031098da01aa6572437 WatchSource:0}: Error finding container c55556334ef2df449eec30be9ed290f08cf5de74b42db031098da01aa6572437: Status 404 returned error can't find the container with id c55556334ef2df449eec30be9ed290f08cf5de74b42db031098da01aa6572437 Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.515310 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.515368 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.515379 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.515399 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.515409 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:41Z","lastTransitionTime":"2025-12-01T08:40:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.529593 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-daemon-9wn4q"] Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.530234 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-b6qqv"] Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.530486 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-9wn4q" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.530879 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-b6qqv" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.533013 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-j4pcc"] Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.533742 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.533863 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-j4pcc" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.535173 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.536029 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.536072 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.536162 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.536619 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.537237 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.538323 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.539010 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.539148 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.539235 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.539375 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.539548 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.539548 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.539653 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-additional-cni-plugins-h7lvd"] Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.539730 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.540314 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-h7lvd" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.541358 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.541744 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.549285 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.549330 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.554093 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"349332bd-6b4d-4306-84ae-07019ca3860f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d441a348087ae3cadda1ad8a7d1c4dff02cf3f6e1d346c294b278fdd9131edcc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://da67635daf6d9e32937f4d5d4ceeee3f8f09ae4fd2dd26c3e85be1286c5949de\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e97356ee274ea78d9f237ffbc756abb6ac8df0dbac244bf0a86ed215b8cbf43\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0404f5a5ed5a9750d4c479b608e68e8960b448f455d5af9d219aa5edaeb789b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a0404f5a5ed5a9750d4c479b608e68e8960b448f455d5af9d219aa5edaeb789b\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"message\\\":\\\"\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764578421\\\\\\\\\\\\\\\" (2025-12-01 08:40:21 +0000 UTC to 2025-12-31 08:40:22 +0000 UTC (now=2025-12-01 08:40:38.066068451 +0000 UTC))\\\\\\\"\\\\nI1201 08:40:38.066144 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1201 08:40:38.066166 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1201 08:40:38.066220 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1201 08:40:38.066232 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1201 08:40:38.066338 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764578432\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764578432\\\\\\\\\\\\\\\" (2025-12-01 07:40:32 +0000 UTC to 2026-12-01 07:40:32 +0000 UTC (now=2025-12-01 08:40:38.066301217 +0000 UTC))\\\\\\\"\\\\nI1201 08:40:38.066357 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1201 08:40:38.066385 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1201 08:40:38.066402 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1201 08:40:38.066414 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-63890811/tls.crt::/tmp/serving-cert-63890811/tls.key\\\\\\\"\\\\nI1201 08:40:38.066419 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1201 08:40:38.066572 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1201 08:40:38.067159 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7354f9337d42115bb1d974bf3772f2d03365acba57f2155135c7a0bbab39879e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:19Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05492df5703392cdf0932d6bf66025b18e53bd05fa0857a7cb2ad70979cbdc3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05492df5703392cdf0932d6bf66025b18e53bd05fa0857a7cb2ad70979cbdc3e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:16Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:41Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.569809 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:41Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.582719 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3264b3af996af452df2953c59bea6deeb5ea6af1b119ce89f7334595c27e70da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9e879b1f563bd66383d58fd7b435a8a81bd50cb387dc1fabb6fc47d5a3fa468\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:41Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.584013 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/b9656a03-0096-4c89-b0da-e37e4103611b-host-run-multus-certs\") pod \"multus-b6qqv\" (UID: \"b9656a03-0096-4c89-b0da-e37e4103611b\") " pod="openshift-multus/multus-b6qqv" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.584048 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/d431b613-a91b-4fff-88b9-447804766289-cnibin\") pod \"multus-additional-cni-plugins-h7lvd\" (UID: \"d431b613-a91b-4fff-88b9-447804766289\") " pod="openshift-multus/multus-additional-cni-plugins-h7lvd" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.584072 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/b7505e14-3560-4d0d-be2c-d6a7a5953865-host-cni-netd\") pod \"ovnkube-node-j4pcc\" (UID: \"b7505e14-3560-4d0d-be2c-d6a7a5953865\") " pod="openshift-ovn-kubernetes/ovnkube-node-j4pcc" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.584095 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4rzxd\" (UniqueName: \"kubernetes.io/projected/3d423402-e08f-42e9-b60c-6eb9bf2f6d48-kube-api-access-4rzxd\") pod \"machine-config-daemon-9wn4q\" (UID: \"3d423402-e08f-42e9-b60c-6eb9bf2f6d48\") " pod="openshift-machine-config-operator/machine-config-daemon-9wn4q" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.584118 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/d431b613-a91b-4fff-88b9-447804766289-system-cni-dir\") pod \"multus-additional-cni-plugins-h7lvd\" (UID: \"d431b613-a91b-4fff-88b9-447804766289\") " pod="openshift-multus/multus-additional-cni-plugins-h7lvd" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.584140 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/b7505e14-3560-4d0d-be2c-d6a7a5953865-host-run-netns\") pod \"ovnkube-node-j4pcc\" (UID: \"b7505e14-3560-4d0d-be2c-d6a7a5953865\") " pod="openshift-ovn-kubernetes/ovnkube-node-j4pcc" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.584160 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/b7505e14-3560-4d0d-be2c-d6a7a5953865-host-run-ovn-kubernetes\") pod \"ovnkube-node-j4pcc\" (UID: \"b7505e14-3560-4d0d-be2c-d6a7a5953865\") " pod="openshift-ovn-kubernetes/ovnkube-node-j4pcc" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.584182 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/b7505e14-3560-4d0d-be2c-d6a7a5953865-env-overrides\") pod \"ovnkube-node-j4pcc\" (UID: \"b7505e14-3560-4d0d-be2c-d6a7a5953865\") " pod="openshift-ovn-kubernetes/ovnkube-node-j4pcc" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.584202 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/3d423402-e08f-42e9-b60c-6eb9bf2f6d48-proxy-tls\") pod \"machine-config-daemon-9wn4q\" (UID: \"3d423402-e08f-42e9-b60c-6eb9bf2f6d48\") " pod="openshift-machine-config-operator/machine-config-daemon-9wn4q" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.584222 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/3d423402-e08f-42e9-b60c-6eb9bf2f6d48-mcd-auth-proxy-config\") pod \"machine-config-daemon-9wn4q\" (UID: \"3d423402-e08f-42e9-b60c-6eb9bf2f6d48\") " pod="openshift-machine-config-operator/machine-config-daemon-9wn4q" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.584254 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/b9656a03-0096-4c89-b0da-e37e4103611b-host-var-lib-kubelet\") pod \"multus-b6qqv\" (UID: \"b9656a03-0096-4c89-b0da-e37e4103611b\") " pod="openshift-multus/multus-b6qqv" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.584352 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/b9656a03-0096-4c89-b0da-e37e4103611b-hostroot\") pod \"multus-b6qqv\" (UID: \"b9656a03-0096-4c89-b0da-e37e4103611b\") " pod="openshift-multus/multus-b6qqv" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.584414 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/d431b613-a91b-4fff-88b9-447804766289-os-release\") pod \"multus-additional-cni-plugins-h7lvd\" (UID: \"d431b613-a91b-4fff-88b9-447804766289\") " pod="openshift-multus/multus-additional-cni-plugins-h7lvd" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.584444 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/b7505e14-3560-4d0d-be2c-d6a7a5953865-log-socket\") pod \"ovnkube-node-j4pcc\" (UID: \"b7505e14-3560-4d0d-be2c-d6a7a5953865\") " pod="openshift-ovn-kubernetes/ovnkube-node-j4pcc" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.584780 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/b7505e14-3560-4d0d-be2c-d6a7a5953865-systemd-units\") pod \"ovnkube-node-j4pcc\" (UID: \"b7505e14-3560-4d0d-be2c-d6a7a5953865\") " pod="openshift-ovn-kubernetes/ovnkube-node-j4pcc" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.584833 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/b7505e14-3560-4d0d-be2c-d6a7a5953865-run-systemd\") pod \"ovnkube-node-j4pcc\" (UID: \"b7505e14-3560-4d0d-be2c-d6a7a5953865\") " pod="openshift-ovn-kubernetes/ovnkube-node-j4pcc" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.584895 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/b7505e14-3560-4d0d-be2c-d6a7a5953865-host-cni-bin\") pod \"ovnkube-node-j4pcc\" (UID: \"b7505e14-3560-4d0d-be2c-d6a7a5953865\") " pod="openshift-ovn-kubernetes/ovnkube-node-j4pcc" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.585026 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/b7505e14-3560-4d0d-be2c-d6a7a5953865-ovn-node-metrics-cert\") pod \"ovnkube-node-j4pcc\" (UID: \"b7505e14-3560-4d0d-be2c-d6a7a5953865\") " pod="openshift-ovn-kubernetes/ovnkube-node-j4pcc" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.585052 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/b9656a03-0096-4c89-b0da-e37e4103611b-multus-cni-dir\") pod \"multus-b6qqv\" (UID: \"b9656a03-0096-4c89-b0da-e37e4103611b\") " pod="openshift-multus/multus-b6qqv" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.585073 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/b9656a03-0096-4c89-b0da-e37e4103611b-multus-daemon-config\") pod \"multus-b6qqv\" (UID: \"b9656a03-0096-4c89-b0da-e37e4103611b\") " pod="openshift-multus/multus-b6qqv" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.585108 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/b9656a03-0096-4c89-b0da-e37e4103611b-cnibin\") pod \"multus-b6qqv\" (UID: \"b9656a03-0096-4c89-b0da-e37e4103611b\") " pod="openshift-multus/multus-b6qqv" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.585196 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/b9656a03-0096-4c89-b0da-e37e4103611b-os-release\") pod \"multus-b6qqv\" (UID: \"b9656a03-0096-4c89-b0da-e37e4103611b\") " pod="openshift-multus/multus-b6qqv" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.585399 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/b9656a03-0096-4c89-b0da-e37e4103611b-host-run-k8s-cni-cncf-io\") pod \"multus-b6qqv\" (UID: \"b9656a03-0096-4c89-b0da-e37e4103611b\") " pod="openshift-multus/multus-b6qqv" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.585431 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/b7505e14-3560-4d0d-be2c-d6a7a5953865-host-kubelet\") pod \"ovnkube-node-j4pcc\" (UID: \"b7505e14-3560-4d0d-be2c-d6a7a5953865\") " pod="openshift-ovn-kubernetes/ovnkube-node-j4pcc" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.585453 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/b7505e14-3560-4d0d-be2c-d6a7a5953865-host-slash\") pod \"ovnkube-node-j4pcc\" (UID: \"b7505e14-3560-4d0d-be2c-d6a7a5953865\") " pod="openshift-ovn-kubernetes/ovnkube-node-j4pcc" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.585507 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/b9656a03-0096-4c89-b0da-e37e4103611b-cni-binary-copy\") pod \"multus-b6qqv\" (UID: \"b9656a03-0096-4c89-b0da-e37e4103611b\") " pod="openshift-multus/multus-b6qqv" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.585529 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/b7505e14-3560-4d0d-be2c-d6a7a5953865-var-lib-openvswitch\") pod \"ovnkube-node-j4pcc\" (UID: \"b7505e14-3560-4d0d-be2c-d6a7a5953865\") " pod="openshift-ovn-kubernetes/ovnkube-node-j4pcc" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.585581 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/b9656a03-0096-4c89-b0da-e37e4103611b-host-var-lib-cni-multus\") pod \"multus-b6qqv\" (UID: \"b9656a03-0096-4c89-b0da-e37e4103611b\") " pod="openshift-multus/multus-b6qqv" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.585613 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cjvnn\" (UniqueName: \"kubernetes.io/projected/b9656a03-0096-4c89-b0da-e37e4103611b-kube-api-access-cjvnn\") pod \"multus-b6qqv\" (UID: \"b9656a03-0096-4c89-b0da-e37e4103611b\") " pod="openshift-multus/multus-b6qqv" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.585665 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/b7505e14-3560-4d0d-be2c-d6a7a5953865-run-ovn\") pod \"ovnkube-node-j4pcc\" (UID: \"b7505e14-3560-4d0d-be2c-d6a7a5953865\") " pod="openshift-ovn-kubernetes/ovnkube-node-j4pcc" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.585687 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/b7505e14-3560-4d0d-be2c-d6a7a5953865-ovnkube-config\") pod \"ovnkube-node-j4pcc\" (UID: \"b7505e14-3560-4d0d-be2c-d6a7a5953865\") " pod="openshift-ovn-kubernetes/ovnkube-node-j4pcc" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.585735 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dgqwb\" (UniqueName: \"kubernetes.io/projected/b7505e14-3560-4d0d-be2c-d6a7a5953865-kube-api-access-dgqwb\") pod \"ovnkube-node-j4pcc\" (UID: \"b7505e14-3560-4d0d-be2c-d6a7a5953865\") " pod="openshift-ovn-kubernetes/ovnkube-node-j4pcc" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.585761 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/b9656a03-0096-4c89-b0da-e37e4103611b-etc-kubernetes\") pod \"multus-b6qqv\" (UID: \"b9656a03-0096-4c89-b0da-e37e4103611b\") " pod="openshift-multus/multus-b6qqv" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.585781 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/b9656a03-0096-4c89-b0da-e37e4103611b-host-run-netns\") pod \"multus-b6qqv\" (UID: \"b9656a03-0096-4c89-b0da-e37e4103611b\") " pod="openshift-multus/multus-b6qqv" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.585829 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/d431b613-a91b-4fff-88b9-447804766289-cni-binary-copy\") pod \"multus-additional-cni-plugins-h7lvd\" (UID: \"d431b613-a91b-4fff-88b9-447804766289\") " pod="openshift-multus/multus-additional-cni-plugins-h7lvd" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.585851 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bq6d5\" (UniqueName: \"kubernetes.io/projected/d431b613-a91b-4fff-88b9-447804766289-kube-api-access-bq6d5\") pod \"multus-additional-cni-plugins-h7lvd\" (UID: \"d431b613-a91b-4fff-88b9-447804766289\") " pod="openshift-multus/multus-additional-cni-plugins-h7lvd" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.585871 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/b7505e14-3560-4d0d-be2c-d6a7a5953865-etc-openvswitch\") pod \"ovnkube-node-j4pcc\" (UID: \"b7505e14-3560-4d0d-be2c-d6a7a5953865\") " pod="openshift-ovn-kubernetes/ovnkube-node-j4pcc" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.585902 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/b9656a03-0096-4c89-b0da-e37e4103611b-multus-conf-dir\") pod \"multus-b6qqv\" (UID: \"b9656a03-0096-4c89-b0da-e37e4103611b\") " pod="openshift-multus/multus-b6qqv" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.585927 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/b9656a03-0096-4c89-b0da-e37e4103611b-system-cni-dir\") pod \"multus-b6qqv\" (UID: \"b9656a03-0096-4c89-b0da-e37e4103611b\") " pod="openshift-multus/multus-b6qqv" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.585949 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/d431b613-a91b-4fff-88b9-447804766289-tuning-conf-dir\") pod \"multus-additional-cni-plugins-h7lvd\" (UID: \"d431b613-a91b-4fff-88b9-447804766289\") " pod="openshift-multus/multus-additional-cni-plugins-h7lvd" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.585984 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/b7505e14-3560-4d0d-be2c-d6a7a5953865-node-log\") pod \"ovnkube-node-j4pcc\" (UID: \"b7505e14-3560-4d0d-be2c-d6a7a5953865\") " pod="openshift-ovn-kubernetes/ovnkube-node-j4pcc" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.586008 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/b7505e14-3560-4d0d-be2c-d6a7a5953865-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-j4pcc\" (UID: \"b7505e14-3560-4d0d-be2c-d6a7a5953865\") " pod="openshift-ovn-kubernetes/ovnkube-node-j4pcc" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.586032 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/3d423402-e08f-42e9-b60c-6eb9bf2f6d48-rootfs\") pod \"machine-config-daemon-9wn4q\" (UID: \"3d423402-e08f-42e9-b60c-6eb9bf2f6d48\") " pod="openshift-machine-config-operator/machine-config-daemon-9wn4q" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.586057 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/b9656a03-0096-4c89-b0da-e37e4103611b-multus-socket-dir-parent\") pod \"multus-b6qqv\" (UID: \"b9656a03-0096-4c89-b0da-e37e4103611b\") " pod="openshift-multus/multus-b6qqv" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.586081 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/d431b613-a91b-4fff-88b9-447804766289-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-h7lvd\" (UID: \"d431b613-a91b-4fff-88b9-447804766289\") " pod="openshift-multus/multus-additional-cni-plugins-h7lvd" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.586101 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/b7505e14-3560-4d0d-be2c-d6a7a5953865-run-openvswitch\") pod \"ovnkube-node-j4pcc\" (UID: \"b7505e14-3560-4d0d-be2c-d6a7a5953865\") " pod="openshift-ovn-kubernetes/ovnkube-node-j4pcc" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.586124 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/b7505e14-3560-4d0d-be2c-d6a7a5953865-ovnkube-script-lib\") pod \"ovnkube-node-j4pcc\" (UID: \"b7505e14-3560-4d0d-be2c-d6a7a5953865\") " pod="openshift-ovn-kubernetes/ovnkube-node-j4pcc" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.586187 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/b9656a03-0096-4c89-b0da-e37e4103611b-host-var-lib-cni-bin\") pod \"multus-b6qqv\" (UID: \"b9656a03-0096-4c89-b0da-e37e4103611b\") " pod="openshift-multus/multus-b6qqv" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.596247 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-c576v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"17292cd0-1eac-40b5-a60e-67167c57409c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7zwfx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:41Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-c576v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:41Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.610250 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0023b0e358d75aacc889ce364846dda89ea513832c87fa4b066aea7dac57e8a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:41Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.618341 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.618409 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.618424 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.618446 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.618461 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:41Z","lastTransitionTime":"2025-12-01T08:40:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.625003 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:41Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.638081 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:41Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.651716 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:41Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.666658 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-9wn4q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3d423402-e08f-42e9-b60c-6eb9bf2f6d48\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rzxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rzxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:41Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-9wn4q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:41Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.682154 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"349332bd-6b4d-4306-84ae-07019ca3860f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d441a348087ae3cadda1ad8a7d1c4dff02cf3f6e1d346c294b278fdd9131edcc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://da67635daf6d9e32937f4d5d4ceeee3f8f09ae4fd2dd26c3e85be1286c5949de\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e97356ee274ea78d9f237ffbc756abb6ac8df0dbac244bf0a86ed215b8cbf43\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0404f5a5ed5a9750d4c479b608e68e8960b448f455d5af9d219aa5edaeb789b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a0404f5a5ed5a9750d4c479b608e68e8960b448f455d5af9d219aa5edaeb789b\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"message\\\":\\\"\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764578421\\\\\\\\\\\\\\\" (2025-12-01 08:40:21 +0000 UTC to 2025-12-31 08:40:22 +0000 UTC (now=2025-12-01 08:40:38.066068451 +0000 UTC))\\\\\\\"\\\\nI1201 08:40:38.066144 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1201 08:40:38.066166 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1201 08:40:38.066220 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1201 08:40:38.066232 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1201 08:40:38.066338 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764578432\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764578432\\\\\\\\\\\\\\\" (2025-12-01 07:40:32 +0000 UTC to 2026-12-01 07:40:32 +0000 UTC (now=2025-12-01 08:40:38.066301217 +0000 UTC))\\\\\\\"\\\\nI1201 08:40:38.066357 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1201 08:40:38.066385 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1201 08:40:38.066402 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1201 08:40:38.066414 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-63890811/tls.crt::/tmp/serving-cert-63890811/tls.key\\\\\\\"\\\\nI1201 08:40:38.066419 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1201 08:40:38.066572 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1201 08:40:38.067159 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7354f9337d42115bb1d974bf3772f2d03365acba57f2155135c7a0bbab39879e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:19Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05492df5703392cdf0932d6bf66025b18e53bd05fa0857a7cb2ad70979cbdc3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05492df5703392cdf0932d6bf66025b18e53bd05fa0857a7cb2ad70979cbdc3e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:16Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:41Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.686503 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/b7505e14-3560-4d0d-be2c-d6a7a5953865-host-run-netns\") pod \"ovnkube-node-j4pcc\" (UID: \"b7505e14-3560-4d0d-be2c-d6a7a5953865\") " pod="openshift-ovn-kubernetes/ovnkube-node-j4pcc" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.686535 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/b7505e14-3560-4d0d-be2c-d6a7a5953865-host-run-ovn-kubernetes\") pod \"ovnkube-node-j4pcc\" (UID: \"b7505e14-3560-4d0d-be2c-d6a7a5953865\") " pod="openshift-ovn-kubernetes/ovnkube-node-j4pcc" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.686553 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/b7505e14-3560-4d0d-be2c-d6a7a5953865-env-overrides\") pod \"ovnkube-node-j4pcc\" (UID: \"b7505e14-3560-4d0d-be2c-d6a7a5953865\") " pod="openshift-ovn-kubernetes/ovnkube-node-j4pcc" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.686573 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4rzxd\" (UniqueName: \"kubernetes.io/projected/3d423402-e08f-42e9-b60c-6eb9bf2f6d48-kube-api-access-4rzxd\") pod \"machine-config-daemon-9wn4q\" (UID: \"3d423402-e08f-42e9-b60c-6eb9bf2f6d48\") " pod="openshift-machine-config-operator/machine-config-daemon-9wn4q" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.686590 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/d431b613-a91b-4fff-88b9-447804766289-system-cni-dir\") pod \"multus-additional-cni-plugins-h7lvd\" (UID: \"d431b613-a91b-4fff-88b9-447804766289\") " pod="openshift-multus/multus-additional-cni-plugins-h7lvd" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.686609 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/b9656a03-0096-4c89-b0da-e37e4103611b-host-var-lib-kubelet\") pod \"multus-b6qqv\" (UID: \"b9656a03-0096-4c89-b0da-e37e4103611b\") " pod="openshift-multus/multus-b6qqv" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.686625 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/b9656a03-0096-4c89-b0da-e37e4103611b-hostroot\") pod \"multus-b6qqv\" (UID: \"b9656a03-0096-4c89-b0da-e37e4103611b\") " pod="openshift-multus/multus-b6qqv" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.686644 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/3d423402-e08f-42e9-b60c-6eb9bf2f6d48-proxy-tls\") pod \"machine-config-daemon-9wn4q\" (UID: \"3d423402-e08f-42e9-b60c-6eb9bf2f6d48\") " pod="openshift-machine-config-operator/machine-config-daemon-9wn4q" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.686661 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/3d423402-e08f-42e9-b60c-6eb9bf2f6d48-mcd-auth-proxy-config\") pod \"machine-config-daemon-9wn4q\" (UID: \"3d423402-e08f-42e9-b60c-6eb9bf2f6d48\") " pod="openshift-machine-config-operator/machine-config-daemon-9wn4q" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.686680 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/d431b613-a91b-4fff-88b9-447804766289-os-release\") pod \"multus-additional-cni-plugins-h7lvd\" (UID: \"d431b613-a91b-4fff-88b9-447804766289\") " pod="openshift-multus/multus-additional-cni-plugins-h7lvd" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.686696 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/b7505e14-3560-4d0d-be2c-d6a7a5953865-log-socket\") pod \"ovnkube-node-j4pcc\" (UID: \"b7505e14-3560-4d0d-be2c-d6a7a5953865\") " pod="openshift-ovn-kubernetes/ovnkube-node-j4pcc" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.686713 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/b7505e14-3560-4d0d-be2c-d6a7a5953865-systemd-units\") pod \"ovnkube-node-j4pcc\" (UID: \"b7505e14-3560-4d0d-be2c-d6a7a5953865\") " pod="openshift-ovn-kubernetes/ovnkube-node-j4pcc" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.686729 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/b7505e14-3560-4d0d-be2c-d6a7a5953865-run-systemd\") pod \"ovnkube-node-j4pcc\" (UID: \"b7505e14-3560-4d0d-be2c-d6a7a5953865\") " pod="openshift-ovn-kubernetes/ovnkube-node-j4pcc" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.686754 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/b7505e14-3560-4d0d-be2c-d6a7a5953865-host-cni-bin\") pod \"ovnkube-node-j4pcc\" (UID: \"b7505e14-3560-4d0d-be2c-d6a7a5953865\") " pod="openshift-ovn-kubernetes/ovnkube-node-j4pcc" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.686771 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/b7505e14-3560-4d0d-be2c-d6a7a5953865-ovn-node-metrics-cert\") pod \"ovnkube-node-j4pcc\" (UID: \"b7505e14-3560-4d0d-be2c-d6a7a5953865\") " pod="openshift-ovn-kubernetes/ovnkube-node-j4pcc" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.686796 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/b9656a03-0096-4c89-b0da-e37e4103611b-multus-cni-dir\") pod \"multus-b6qqv\" (UID: \"b9656a03-0096-4c89-b0da-e37e4103611b\") " pod="openshift-multus/multus-b6qqv" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.686814 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/b9656a03-0096-4c89-b0da-e37e4103611b-multus-daemon-config\") pod \"multus-b6qqv\" (UID: \"b9656a03-0096-4c89-b0da-e37e4103611b\") " pod="openshift-multus/multus-b6qqv" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.686830 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/b9656a03-0096-4c89-b0da-e37e4103611b-os-release\") pod \"multus-b6qqv\" (UID: \"b9656a03-0096-4c89-b0da-e37e4103611b\") " pod="openshift-multus/multus-b6qqv" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.686850 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/b9656a03-0096-4c89-b0da-e37e4103611b-host-run-k8s-cni-cncf-io\") pod \"multus-b6qqv\" (UID: \"b9656a03-0096-4c89-b0da-e37e4103611b\") " pod="openshift-multus/multus-b6qqv" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.686867 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/b7505e14-3560-4d0d-be2c-d6a7a5953865-host-kubelet\") pod \"ovnkube-node-j4pcc\" (UID: \"b7505e14-3560-4d0d-be2c-d6a7a5953865\") " pod="openshift-ovn-kubernetes/ovnkube-node-j4pcc" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.686883 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/b7505e14-3560-4d0d-be2c-d6a7a5953865-host-slash\") pod \"ovnkube-node-j4pcc\" (UID: \"b7505e14-3560-4d0d-be2c-d6a7a5953865\") " pod="openshift-ovn-kubernetes/ovnkube-node-j4pcc" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.686908 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/b9656a03-0096-4c89-b0da-e37e4103611b-cnibin\") pod \"multus-b6qqv\" (UID: \"b9656a03-0096-4c89-b0da-e37e4103611b\") " pod="openshift-multus/multus-b6qqv" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.686925 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/b9656a03-0096-4c89-b0da-e37e4103611b-cni-binary-copy\") pod \"multus-b6qqv\" (UID: \"b9656a03-0096-4c89-b0da-e37e4103611b\") " pod="openshift-multus/multus-b6qqv" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.686948 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/b7505e14-3560-4d0d-be2c-d6a7a5953865-var-lib-openvswitch\") pod \"ovnkube-node-j4pcc\" (UID: \"b7505e14-3560-4d0d-be2c-d6a7a5953865\") " pod="openshift-ovn-kubernetes/ovnkube-node-j4pcc" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.686981 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/b7505e14-3560-4d0d-be2c-d6a7a5953865-run-ovn\") pod \"ovnkube-node-j4pcc\" (UID: \"b7505e14-3560-4d0d-be2c-d6a7a5953865\") " pod="openshift-ovn-kubernetes/ovnkube-node-j4pcc" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.687000 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/b7505e14-3560-4d0d-be2c-d6a7a5953865-ovnkube-config\") pod \"ovnkube-node-j4pcc\" (UID: \"b7505e14-3560-4d0d-be2c-d6a7a5953865\") " pod="openshift-ovn-kubernetes/ovnkube-node-j4pcc" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.687021 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dgqwb\" (UniqueName: \"kubernetes.io/projected/b7505e14-3560-4d0d-be2c-d6a7a5953865-kube-api-access-dgqwb\") pod \"ovnkube-node-j4pcc\" (UID: \"b7505e14-3560-4d0d-be2c-d6a7a5953865\") " pod="openshift-ovn-kubernetes/ovnkube-node-j4pcc" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.687039 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/b9656a03-0096-4c89-b0da-e37e4103611b-host-var-lib-cni-multus\") pod \"multus-b6qqv\" (UID: \"b9656a03-0096-4c89-b0da-e37e4103611b\") " pod="openshift-multus/multus-b6qqv" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.687056 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cjvnn\" (UniqueName: \"kubernetes.io/projected/b9656a03-0096-4c89-b0da-e37e4103611b-kube-api-access-cjvnn\") pod \"multus-b6qqv\" (UID: \"b9656a03-0096-4c89-b0da-e37e4103611b\") " pod="openshift-multus/multus-b6qqv" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.687074 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/b9656a03-0096-4c89-b0da-e37e4103611b-etc-kubernetes\") pod \"multus-b6qqv\" (UID: \"b9656a03-0096-4c89-b0da-e37e4103611b\") " pod="openshift-multus/multus-b6qqv" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.687093 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/b9656a03-0096-4c89-b0da-e37e4103611b-host-run-netns\") pod \"multus-b6qqv\" (UID: \"b9656a03-0096-4c89-b0da-e37e4103611b\") " pod="openshift-multus/multus-b6qqv" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.687150 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bq6d5\" (UniqueName: \"kubernetes.io/projected/d431b613-a91b-4fff-88b9-447804766289-kube-api-access-bq6d5\") pod \"multus-additional-cni-plugins-h7lvd\" (UID: \"d431b613-a91b-4fff-88b9-447804766289\") " pod="openshift-multus/multus-additional-cni-plugins-h7lvd" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.687168 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/b7505e14-3560-4d0d-be2c-d6a7a5953865-etc-openvswitch\") pod \"ovnkube-node-j4pcc\" (UID: \"b7505e14-3560-4d0d-be2c-d6a7a5953865\") " pod="openshift-ovn-kubernetes/ovnkube-node-j4pcc" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.687186 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/d431b613-a91b-4fff-88b9-447804766289-cni-binary-copy\") pod \"multus-additional-cni-plugins-h7lvd\" (UID: \"d431b613-a91b-4fff-88b9-447804766289\") " pod="openshift-multus/multus-additional-cni-plugins-h7lvd" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.687209 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/b9656a03-0096-4c89-b0da-e37e4103611b-multus-conf-dir\") pod \"multus-b6qqv\" (UID: \"b9656a03-0096-4c89-b0da-e37e4103611b\") " pod="openshift-multus/multus-b6qqv" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.687237 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/b9656a03-0096-4c89-b0da-e37e4103611b-system-cni-dir\") pod \"multus-b6qqv\" (UID: \"b9656a03-0096-4c89-b0da-e37e4103611b\") " pod="openshift-multus/multus-b6qqv" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.687255 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/d431b613-a91b-4fff-88b9-447804766289-tuning-conf-dir\") pod \"multus-additional-cni-plugins-h7lvd\" (UID: \"d431b613-a91b-4fff-88b9-447804766289\") " pod="openshift-multus/multus-additional-cni-plugins-h7lvd" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.687273 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/b7505e14-3560-4d0d-be2c-d6a7a5953865-node-log\") pod \"ovnkube-node-j4pcc\" (UID: \"b7505e14-3560-4d0d-be2c-d6a7a5953865\") " pod="openshift-ovn-kubernetes/ovnkube-node-j4pcc" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.687290 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/b7505e14-3560-4d0d-be2c-d6a7a5953865-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-j4pcc\" (UID: \"b7505e14-3560-4d0d-be2c-d6a7a5953865\") " pod="openshift-ovn-kubernetes/ovnkube-node-j4pcc" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.687309 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/b7505e14-3560-4d0d-be2c-d6a7a5953865-run-openvswitch\") pod \"ovnkube-node-j4pcc\" (UID: \"b7505e14-3560-4d0d-be2c-d6a7a5953865\") " pod="openshift-ovn-kubernetes/ovnkube-node-j4pcc" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.687326 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/b7505e14-3560-4d0d-be2c-d6a7a5953865-ovnkube-script-lib\") pod \"ovnkube-node-j4pcc\" (UID: \"b7505e14-3560-4d0d-be2c-d6a7a5953865\") " pod="openshift-ovn-kubernetes/ovnkube-node-j4pcc" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.687342 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/3d423402-e08f-42e9-b60c-6eb9bf2f6d48-rootfs\") pod \"machine-config-daemon-9wn4q\" (UID: \"3d423402-e08f-42e9-b60c-6eb9bf2f6d48\") " pod="openshift-machine-config-operator/machine-config-daemon-9wn4q" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.687358 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/b9656a03-0096-4c89-b0da-e37e4103611b-multus-socket-dir-parent\") pod \"multus-b6qqv\" (UID: \"b9656a03-0096-4c89-b0da-e37e4103611b\") " pod="openshift-multus/multus-b6qqv" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.687376 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/d431b613-a91b-4fff-88b9-447804766289-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-h7lvd\" (UID: \"d431b613-a91b-4fff-88b9-447804766289\") " pod="openshift-multus/multus-additional-cni-plugins-h7lvd" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.687402 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/b9656a03-0096-4c89-b0da-e37e4103611b-host-var-lib-cni-bin\") pod \"multus-b6qqv\" (UID: \"b9656a03-0096-4c89-b0da-e37e4103611b\") " pod="openshift-multus/multus-b6qqv" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.687420 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/b9656a03-0096-4c89-b0da-e37e4103611b-host-run-multus-certs\") pod \"multus-b6qqv\" (UID: \"b9656a03-0096-4c89-b0da-e37e4103611b\") " pod="openshift-multus/multus-b6qqv" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.687438 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/d431b613-a91b-4fff-88b9-447804766289-cnibin\") pod \"multus-additional-cni-plugins-h7lvd\" (UID: \"d431b613-a91b-4fff-88b9-447804766289\") " pod="openshift-multus/multus-additional-cni-plugins-h7lvd" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.687455 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/b7505e14-3560-4d0d-be2c-d6a7a5953865-host-cni-netd\") pod \"ovnkube-node-j4pcc\" (UID: \"b7505e14-3560-4d0d-be2c-d6a7a5953865\") " pod="openshift-ovn-kubernetes/ovnkube-node-j4pcc" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.687541 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/b7505e14-3560-4d0d-be2c-d6a7a5953865-host-cni-netd\") pod \"ovnkube-node-j4pcc\" (UID: \"b7505e14-3560-4d0d-be2c-d6a7a5953865\") " pod="openshift-ovn-kubernetes/ovnkube-node-j4pcc" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.687577 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/b7505e14-3560-4d0d-be2c-d6a7a5953865-host-run-netns\") pod \"ovnkube-node-j4pcc\" (UID: \"b7505e14-3560-4d0d-be2c-d6a7a5953865\") " pod="openshift-ovn-kubernetes/ovnkube-node-j4pcc" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.687600 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/b7505e14-3560-4d0d-be2c-d6a7a5953865-host-run-ovn-kubernetes\") pod \"ovnkube-node-j4pcc\" (UID: \"b7505e14-3560-4d0d-be2c-d6a7a5953865\") " pod="openshift-ovn-kubernetes/ovnkube-node-j4pcc" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.688132 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/b7505e14-3560-4d0d-be2c-d6a7a5953865-env-overrides\") pod \"ovnkube-node-j4pcc\" (UID: \"b7505e14-3560-4d0d-be2c-d6a7a5953865\") " pod="openshift-ovn-kubernetes/ovnkube-node-j4pcc" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.688366 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/d431b613-a91b-4fff-88b9-447804766289-system-cni-dir\") pod \"multus-additional-cni-plugins-h7lvd\" (UID: \"d431b613-a91b-4fff-88b9-447804766289\") " pod="openshift-multus/multus-additional-cni-plugins-h7lvd" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.688398 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/b9656a03-0096-4c89-b0da-e37e4103611b-host-var-lib-kubelet\") pod \"multus-b6qqv\" (UID: \"b9656a03-0096-4c89-b0da-e37e4103611b\") " pod="openshift-multus/multus-b6qqv" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.688423 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/b9656a03-0096-4c89-b0da-e37e4103611b-hostroot\") pod \"multus-b6qqv\" (UID: \"b9656a03-0096-4c89-b0da-e37e4103611b\") " pod="openshift-multus/multus-b6qqv" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.688766 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/b9656a03-0096-4c89-b0da-e37e4103611b-host-var-lib-cni-multus\") pod \"multus-b6qqv\" (UID: \"b9656a03-0096-4c89-b0da-e37e4103611b\") " pod="openshift-multus/multus-b6qqv" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.689031 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/b7505e14-3560-4d0d-be2c-d6a7a5953865-node-log\") pod \"ovnkube-node-j4pcc\" (UID: \"b7505e14-3560-4d0d-be2c-d6a7a5953865\") " pod="openshift-ovn-kubernetes/ovnkube-node-j4pcc" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.689208 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/b9656a03-0096-4c89-b0da-e37e4103611b-etc-kubernetes\") pod \"multus-b6qqv\" (UID: \"b9656a03-0096-4c89-b0da-e37e4103611b\") " pod="openshift-multus/multus-b6qqv" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.689241 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/b9656a03-0096-4c89-b0da-e37e4103611b-host-run-netns\") pod \"multus-b6qqv\" (UID: \"b9656a03-0096-4c89-b0da-e37e4103611b\") " pod="openshift-multus/multus-b6qqv" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.689297 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/3d423402-e08f-42e9-b60c-6eb9bf2f6d48-mcd-auth-proxy-config\") pod \"machine-config-daemon-9wn4q\" (UID: \"3d423402-e08f-42e9-b60c-6eb9bf2f6d48\") " pod="openshift-machine-config-operator/machine-config-daemon-9wn4q" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.689358 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/d431b613-a91b-4fff-88b9-447804766289-os-release\") pod \"multus-additional-cni-plugins-h7lvd\" (UID: \"d431b613-a91b-4fff-88b9-447804766289\") " pod="openshift-multus/multus-additional-cni-plugins-h7lvd" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.689384 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/b7505e14-3560-4d0d-be2c-d6a7a5953865-etc-openvswitch\") pod \"ovnkube-node-j4pcc\" (UID: \"b7505e14-3560-4d0d-be2c-d6a7a5953865\") " pod="openshift-ovn-kubernetes/ovnkube-node-j4pcc" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.689389 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/b7505e14-3560-4d0d-be2c-d6a7a5953865-log-socket\") pod \"ovnkube-node-j4pcc\" (UID: \"b7505e14-3560-4d0d-be2c-d6a7a5953865\") " pod="openshift-ovn-kubernetes/ovnkube-node-j4pcc" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.689410 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/b7505e14-3560-4d0d-be2c-d6a7a5953865-systemd-units\") pod \"ovnkube-node-j4pcc\" (UID: \"b7505e14-3560-4d0d-be2c-d6a7a5953865\") " pod="openshift-ovn-kubernetes/ovnkube-node-j4pcc" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.689436 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/b7505e14-3560-4d0d-be2c-d6a7a5953865-run-systemd\") pod \"ovnkube-node-j4pcc\" (UID: \"b7505e14-3560-4d0d-be2c-d6a7a5953865\") " pod="openshift-ovn-kubernetes/ovnkube-node-j4pcc" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.689462 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/b7505e14-3560-4d0d-be2c-d6a7a5953865-host-cni-bin\") pod \"ovnkube-node-j4pcc\" (UID: \"b7505e14-3560-4d0d-be2c-d6a7a5953865\") " pod="openshift-ovn-kubernetes/ovnkube-node-j4pcc" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.689893 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/d431b613-a91b-4fff-88b9-447804766289-cni-binary-copy\") pod \"multus-additional-cni-plugins-h7lvd\" (UID: \"d431b613-a91b-4fff-88b9-447804766289\") " pod="openshift-multus/multus-additional-cni-plugins-h7lvd" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.689933 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/b9656a03-0096-4c89-b0da-e37e4103611b-multus-conf-dir\") pod \"multus-b6qqv\" (UID: \"b9656a03-0096-4c89-b0da-e37e4103611b\") " pod="openshift-multus/multus-b6qqv" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.690055 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/b9656a03-0096-4c89-b0da-e37e4103611b-system-cni-dir\") pod \"multus-b6qqv\" (UID: \"b9656a03-0096-4c89-b0da-e37e4103611b\") " pod="openshift-multus/multus-b6qqv" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.690791 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/b7505e14-3560-4d0d-be2c-d6a7a5953865-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-j4pcc\" (UID: \"b7505e14-3560-4d0d-be2c-d6a7a5953865\") " pod="openshift-ovn-kubernetes/ovnkube-node-j4pcc" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.690829 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/b7505e14-3560-4d0d-be2c-d6a7a5953865-run-openvswitch\") pod \"ovnkube-node-j4pcc\" (UID: \"b7505e14-3560-4d0d-be2c-d6a7a5953865\") " pod="openshift-ovn-kubernetes/ovnkube-node-j4pcc" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.690979 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/b9656a03-0096-4c89-b0da-e37e4103611b-multus-cni-dir\") pod \"multus-b6qqv\" (UID: \"b9656a03-0096-4c89-b0da-e37e4103611b\") " pod="openshift-multus/multus-b6qqv" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.691334 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/b7505e14-3560-4d0d-be2c-d6a7a5953865-ovnkube-script-lib\") pod \"ovnkube-node-j4pcc\" (UID: \"b7505e14-3560-4d0d-be2c-d6a7a5953865\") " pod="openshift-ovn-kubernetes/ovnkube-node-j4pcc" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.691374 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/3d423402-e08f-42e9-b60c-6eb9bf2f6d48-rootfs\") pod \"machine-config-daemon-9wn4q\" (UID: \"3d423402-e08f-42e9-b60c-6eb9bf2f6d48\") " pod="openshift-machine-config-operator/machine-config-daemon-9wn4q" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.691412 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/b9656a03-0096-4c89-b0da-e37e4103611b-multus-socket-dir-parent\") pod \"multus-b6qqv\" (UID: \"b9656a03-0096-4c89-b0da-e37e4103611b\") " pod="openshift-multus/multus-b6qqv" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.691653 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/b9656a03-0096-4c89-b0da-e37e4103611b-multus-daemon-config\") pod \"multus-b6qqv\" (UID: \"b9656a03-0096-4c89-b0da-e37e4103611b\") " pod="openshift-multus/multus-b6qqv" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.691733 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/b9656a03-0096-4c89-b0da-e37e4103611b-os-release\") pod \"multus-b6qqv\" (UID: \"b9656a03-0096-4c89-b0da-e37e4103611b\") " pod="openshift-multus/multus-b6qqv" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.691772 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/b9656a03-0096-4c89-b0da-e37e4103611b-host-run-k8s-cni-cncf-io\") pod \"multus-b6qqv\" (UID: \"b9656a03-0096-4c89-b0da-e37e4103611b\") " pod="openshift-multus/multus-b6qqv" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.691798 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/b7505e14-3560-4d0d-be2c-d6a7a5953865-host-kubelet\") pod \"ovnkube-node-j4pcc\" (UID: \"b7505e14-3560-4d0d-be2c-d6a7a5953865\") " pod="openshift-ovn-kubernetes/ovnkube-node-j4pcc" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.691826 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/b7505e14-3560-4d0d-be2c-d6a7a5953865-host-slash\") pod \"ovnkube-node-j4pcc\" (UID: \"b7505e14-3560-4d0d-be2c-d6a7a5953865\") " pod="openshift-ovn-kubernetes/ovnkube-node-j4pcc" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.691837 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/d431b613-a91b-4fff-88b9-447804766289-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-h7lvd\" (UID: \"d431b613-a91b-4fff-88b9-447804766289\") " pod="openshift-multus/multus-additional-cni-plugins-h7lvd" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.691870 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/b9656a03-0096-4c89-b0da-e37e4103611b-cnibin\") pod \"multus-b6qqv\" (UID: \"b9656a03-0096-4c89-b0da-e37e4103611b\") " pod="openshift-multus/multus-b6qqv" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.691878 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/b9656a03-0096-4c89-b0da-e37e4103611b-host-var-lib-cni-bin\") pod \"multus-b6qqv\" (UID: \"b9656a03-0096-4c89-b0da-e37e4103611b\") " pod="openshift-multus/multus-b6qqv" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.691908 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/b9656a03-0096-4c89-b0da-e37e4103611b-host-run-multus-certs\") pod \"multus-b6qqv\" (UID: \"b9656a03-0096-4c89-b0da-e37e4103611b\") " pod="openshift-multus/multus-b6qqv" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.691934 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/d431b613-a91b-4fff-88b9-447804766289-cnibin\") pod \"multus-additional-cni-plugins-h7lvd\" (UID: \"d431b613-a91b-4fff-88b9-447804766289\") " pod="openshift-multus/multus-additional-cni-plugins-h7lvd" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.691977 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/b7505e14-3560-4d0d-be2c-d6a7a5953865-run-ovn\") pod \"ovnkube-node-j4pcc\" (UID: \"b7505e14-3560-4d0d-be2c-d6a7a5953865\") " pod="openshift-ovn-kubernetes/ovnkube-node-j4pcc" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.692008 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/b7505e14-3560-4d0d-be2c-d6a7a5953865-var-lib-openvswitch\") pod \"ovnkube-node-j4pcc\" (UID: \"b7505e14-3560-4d0d-be2c-d6a7a5953865\") " pod="openshift-ovn-kubernetes/ovnkube-node-j4pcc" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.692331 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/b9656a03-0096-4c89-b0da-e37e4103611b-cni-binary-copy\") pod \"multus-b6qqv\" (UID: \"b9656a03-0096-4c89-b0da-e37e4103611b\") " pod="openshift-multus/multus-b6qqv" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.692485 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/b7505e14-3560-4d0d-be2c-d6a7a5953865-ovnkube-config\") pod \"ovnkube-node-j4pcc\" (UID: \"b7505e14-3560-4d0d-be2c-d6a7a5953865\") " pod="openshift-ovn-kubernetes/ovnkube-node-j4pcc" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.693525 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/3d423402-e08f-42e9-b60c-6eb9bf2f6d48-proxy-tls\") pod \"machine-config-daemon-9wn4q\" (UID: \"3d423402-e08f-42e9-b60c-6eb9bf2f6d48\") " pod="openshift-machine-config-operator/machine-config-daemon-9wn4q" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.693688 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/d431b613-a91b-4fff-88b9-447804766289-tuning-conf-dir\") pod \"multus-additional-cni-plugins-h7lvd\" (UID: \"d431b613-a91b-4fff-88b9-447804766289\") " pod="openshift-multus/multus-additional-cni-plugins-h7lvd" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.695099 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/b7505e14-3560-4d0d-be2c-d6a7a5953865-ovn-node-metrics-cert\") pod \"ovnkube-node-j4pcc\" (UID: \"b7505e14-3560-4d0d-be2c-d6a7a5953865\") " pod="openshift-ovn-kubernetes/ovnkube-node-j4pcc" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.723030 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.723081 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.723097 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.723119 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.723134 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:41Z","lastTransitionTime":"2025-12-01T08:40:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.725680 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3264b3af996af452df2953c59bea6deeb5ea6af1b119ce89f7334595c27e70da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9e879b1f563bd66383d58fd7b435a8a81bd50cb387dc1fabb6fc47d5a3fa468\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:41Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.734119 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cjvnn\" (UniqueName: \"kubernetes.io/projected/b9656a03-0096-4c89-b0da-e37e4103611b-kube-api-access-cjvnn\") pod \"multus-b6qqv\" (UID: \"b9656a03-0096-4c89-b0da-e37e4103611b\") " pod="openshift-multus/multus-b6qqv" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.734549 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4rzxd\" (UniqueName: \"kubernetes.io/projected/3d423402-e08f-42e9-b60c-6eb9bf2f6d48-kube-api-access-4rzxd\") pod \"machine-config-daemon-9wn4q\" (UID: \"3d423402-e08f-42e9-b60c-6eb9bf2f6d48\") " pod="openshift-machine-config-operator/machine-config-daemon-9wn4q" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.735237 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dgqwb\" (UniqueName: \"kubernetes.io/projected/b7505e14-3560-4d0d-be2c-d6a7a5953865-kube-api-access-dgqwb\") pod \"ovnkube-node-j4pcc\" (UID: \"b7505e14-3560-4d0d-be2c-d6a7a5953865\") " pod="openshift-ovn-kubernetes/ovnkube-node-j4pcc" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.737241 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bq6d5\" (UniqueName: \"kubernetes.io/projected/d431b613-a91b-4fff-88b9-447804766289-kube-api-access-bq6d5\") pod \"multus-additional-cni-plugins-h7lvd\" (UID: \"d431b613-a91b-4fff-88b9-447804766289\") " pod="openshift-multus/multus-additional-cni-plugins-h7lvd" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.741011 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-b6qqv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9656a03-0096-4c89-b0da-e37e4103611b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cjvnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:41Z\\\"}}\" for pod \"openshift-multus\"/\"multus-b6qqv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:41Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.777136 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-9wn4q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3d423402-e08f-42e9-b60c-6eb9bf2f6d48\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rzxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rzxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:41Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-9wn4q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:41Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.827620 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.827655 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.827666 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.827682 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.827693 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:41Z","lastTransitionTime":"2025-12-01T08:40:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.840798 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:41Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.852955 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-9wn4q" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.862847 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-b6qqv" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.870332 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-j4pcc" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.885342 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.906195 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.908659 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-h7lvd" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.922021 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-c576v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"17292cd0-1eac-40b5-a60e-67167c57409c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7zwfx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:41Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-c576v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:41Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.929395 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.929423 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.929435 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.929453 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.929462 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:41Z","lastTransitionTime":"2025-12-01T08:40:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:41 crc kubenswrapper[4813]: W1201 08:40:41.935146 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb9656a03_0096_4c89_b0da_e37e4103611b.slice/crio-8f81f892f09ab7808761cf18e77adde9315a9fa92e23b580e175715eb257b1e4 WatchSource:0}: Error finding container 8f81f892f09ab7808761cf18e77adde9315a9fa92e23b580e175715eb257b1e4: Status 404 returned error can't find the container with id 8f81f892f09ab7808761cf18e77adde9315a9fa92e23b580e175715eb257b1e4 Dec 01 08:40:41 crc kubenswrapper[4813]: W1201 08:40:41.937583 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd431b613_a91b_4fff_88b9_447804766289.slice/crio-28a18044ecafa1ecd70a61955c49b1045ab8b11be46dca9212305b475abf21a7 WatchSource:0}: Error finding container 28a18044ecafa1ecd70a61955c49b1045ab8b11be46dca9212305b475abf21a7: Status 404 returned error can't find the container with id 28a18044ecafa1ecd70a61955c49b1045ab8b11be46dca9212305b475abf21a7 Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.958209 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/kube-controller-manager-crc"] Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.960283 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"c7574f892d5e9142296b26a0b484eb6759ab53f571139fb6e23b93dc3da05780"} Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.961524 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-h7lvd" event={"ID":"d431b613-a91b-4fff-88b9-447804766289","Type":"ContainerStarted","Data":"28a18044ecafa1ecd70a61955c49b1045ab8b11be46dca9212305b475abf21a7"} Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.962415 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-b6qqv" event={"ID":"b9656a03-0096-4c89-b0da-e37e4103611b","Type":"ContainerStarted","Data":"8f81f892f09ab7808761cf18e77adde9315a9fa92e23b580e175715eb257b1e4"} Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.963590 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-j4pcc" event={"ID":"b7505e14-3560-4d0d-be2c-d6a7a5953865","Type":"ContainerStarted","Data":"4418f2beb2999b7488df1221f849f6a9c01c3a4439ea419a6462443d759f7309"} Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.964620 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-9wn4q" event={"ID":"3d423402-e08f-42e9-b60c-6eb9bf2f6d48","Type":"ContainerStarted","Data":"00326916424a8dc8389475b4189b139d312ff15104c579e21d0e78603db630e0"} Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.971369 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-c576v" event={"ID":"17292cd0-1eac-40b5-a60e-67167c57409c","Type":"ContainerStarted","Data":"a42206c914d0493ca74788ea1a941de80920ab7b810fb95227cd9b8711e8f4b5"} Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.971408 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-c576v" event={"ID":"17292cd0-1eac-40b5-a60e-67167c57409c","Type":"ContainerStarted","Data":"c55556334ef2df449eec30be9ed290f08cf5de74b42db031098da01aa6572437"} Dec 01 08:40:41 crc kubenswrapper[4813]: I1201 08:40:41.990075 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-j4pcc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b7505e14-3560-4d0d-be2c-d6a7a5953865\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:41Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-j4pcc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:41Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:42 crc kubenswrapper[4813]: I1201 08:40:42.009833 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-h7lvd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d431b613-a91b-4fff-88b9-447804766289\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:41Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-h7lvd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:42Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:42 crc kubenswrapper[4813]: I1201 08:40:42.010027 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 08:40:42 crc kubenswrapper[4813]: I1201 08:40:42.010128 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 08:40:42 crc kubenswrapper[4813]: E1201 08:40:42.010186 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 08:40:46.010168254 +0000 UTC m=+30.633089830 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:40:42 crc kubenswrapper[4813]: I1201 08:40:42.010240 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 08:40:42 crc kubenswrapper[4813]: E1201 08:40:42.010524 4813 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 01 08:40:42 crc kubenswrapper[4813]: E1201 08:40:42.010567 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-01 08:40:46.010554535 +0000 UTC m=+30.633476121 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 01 08:40:42 crc kubenswrapper[4813]: E1201 08:40:42.011068 4813 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 01 08:40:42 crc kubenswrapper[4813]: E1201 08:40:42.011110 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-01 08:40:46.011101071 +0000 UTC m=+30.634022657 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 01 08:40:42 crc kubenswrapper[4813]: I1201 08:40:42.033195 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:42 crc kubenswrapper[4813]: I1201 08:40:42.033236 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:42 crc kubenswrapper[4813]: I1201 08:40:42.033250 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:42 crc kubenswrapper[4813]: I1201 08:40:42.033272 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:42 crc kubenswrapper[4813]: I1201 08:40:42.033286 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:42Z","lastTransitionTime":"2025-12-01T08:40:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:42 crc kubenswrapper[4813]: I1201 08:40:42.038023 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0023b0e358d75aacc889ce364846dda89ea513832c87fa4b066aea7dac57e8a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:42Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:42 crc kubenswrapper[4813]: I1201 08:40:42.052477 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:42Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:42 crc kubenswrapper[4813]: I1201 08:40:42.074683 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:42Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:42 crc kubenswrapper[4813]: I1201 08:40:42.088011 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:42Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:42 crc kubenswrapper[4813]: I1201 08:40:42.101121 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"349332bd-6b4d-4306-84ae-07019ca3860f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d441a348087ae3cadda1ad8a7d1c4dff02cf3f6e1d346c294b278fdd9131edcc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://da67635daf6d9e32937f4d5d4ceeee3f8f09ae4fd2dd26c3e85be1286c5949de\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e97356ee274ea78d9f237ffbc756abb6ac8df0dbac244bf0a86ed215b8cbf43\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0404f5a5ed5a9750d4c479b608e68e8960b448f455d5af9d219aa5edaeb789b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a0404f5a5ed5a9750d4c479b608e68e8960b448f455d5af9d219aa5edaeb789b\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"message\\\":\\\"\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764578421\\\\\\\\\\\\\\\" (2025-12-01 08:40:21 +0000 UTC to 2025-12-31 08:40:22 +0000 UTC (now=2025-12-01 08:40:38.066068451 +0000 UTC))\\\\\\\"\\\\nI1201 08:40:38.066144 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1201 08:40:38.066166 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1201 08:40:38.066220 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1201 08:40:38.066232 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1201 08:40:38.066338 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764578432\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764578432\\\\\\\\\\\\\\\" (2025-12-01 07:40:32 +0000 UTC to 2026-12-01 07:40:32 +0000 UTC (now=2025-12-01 08:40:38.066301217 +0000 UTC))\\\\\\\"\\\\nI1201 08:40:38.066357 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1201 08:40:38.066385 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1201 08:40:38.066402 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1201 08:40:38.066414 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-63890811/tls.crt::/tmp/serving-cert-63890811/tls.key\\\\\\\"\\\\nI1201 08:40:38.066419 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1201 08:40:38.066572 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1201 08:40:38.067159 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7354f9337d42115bb1d974bf3772f2d03365acba57f2155135c7a0bbab39879e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:19Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05492df5703392cdf0932d6bf66025b18e53bd05fa0857a7cb2ad70979cbdc3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05492df5703392cdf0932d6bf66025b18e53bd05fa0857a7cb2ad70979cbdc3e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:16Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:42Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:42 crc kubenswrapper[4813]: I1201 08:40:42.111115 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 08:40:42 crc kubenswrapper[4813]: I1201 08:40:42.111178 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 08:40:42 crc kubenswrapper[4813]: E1201 08:40:42.111309 4813 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 01 08:40:42 crc kubenswrapper[4813]: E1201 08:40:42.111325 4813 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 01 08:40:42 crc kubenswrapper[4813]: E1201 08:40:42.111336 4813 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 08:40:42 crc kubenswrapper[4813]: E1201 08:40:42.111388 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-01 08:40:46.111374106 +0000 UTC m=+30.734295692 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 08:40:42 crc kubenswrapper[4813]: E1201 08:40:42.111432 4813 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 01 08:40:42 crc kubenswrapper[4813]: E1201 08:40:42.111465 4813 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 01 08:40:42 crc kubenswrapper[4813]: E1201 08:40:42.111509 4813 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 08:40:42 crc kubenswrapper[4813]: E1201 08:40:42.111601 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-01 08:40:46.111576112 +0000 UTC m=+30.734497698 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 08:40:42 crc kubenswrapper[4813]: I1201 08:40:42.117141 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3264b3af996af452df2953c59bea6deeb5ea6af1b119ce89f7334595c27e70da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9e879b1f563bd66383d58fd7b435a8a81bd50cb387dc1fabb6fc47d5a3fa468\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:42Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:42 crc kubenswrapper[4813]: I1201 08:40:42.130784 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-b6qqv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9656a03-0096-4c89-b0da-e37e4103611b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cjvnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:41Z\\\"}}\" for pod \"openshift-multus\"/\"multus-b6qqv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:42Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:42 crc kubenswrapper[4813]: I1201 08:40:42.137525 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:42 crc kubenswrapper[4813]: I1201 08:40:42.137774 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:42 crc kubenswrapper[4813]: I1201 08:40:42.137864 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:42 crc kubenswrapper[4813]: I1201 08:40:42.137987 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:42 crc kubenswrapper[4813]: I1201 08:40:42.138069 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:42Z","lastTransitionTime":"2025-12-01T08:40:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:42 crc kubenswrapper[4813]: I1201 08:40:42.141428 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-9wn4q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3d423402-e08f-42e9-b60c-6eb9bf2f6d48\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rzxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rzxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:41Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-9wn4q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:42Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:42 crc kubenswrapper[4813]: I1201 08:40:42.161868 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:42Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:42 crc kubenswrapper[4813]: I1201 08:40:42.173864 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-c576v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"17292cd0-1eac-40b5-a60e-67167c57409c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a42206c914d0493ca74788ea1a941de80920ab7b810fb95227cd9b8711e8f4b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7zwfx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:41Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-c576v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:42Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:42 crc kubenswrapper[4813]: I1201 08:40:42.191520 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-j4pcc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b7505e14-3560-4d0d-be2c-d6a7a5953865\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:41Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-j4pcc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:42Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:42 crc kubenswrapper[4813]: I1201 08:40:42.206939 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-h7lvd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d431b613-a91b-4fff-88b9-447804766289\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:41Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-h7lvd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:42Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:42 crc kubenswrapper[4813]: I1201 08:40:42.219295 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6fb5a648-ae4f-4e46-a530-007e9bf84abb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e24ebfc52dcb28368a2d436e7d9ab78f07f2aec2a4c71405a6c7d3301eee96a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33898118bbf639c8e6eae5aafa10386b95795512444ed5f535db3c663f169df2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9902c00bf7738686c93f62e1772a484d05715d74b75ee98c84c8f23ed383824b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3657b47ec9f1563c83bfbe3d728c8fd123f7078e219d251b1874269e42c31eea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:16Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:42Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:42 crc kubenswrapper[4813]: I1201 08:40:42.240583 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:42 crc kubenswrapper[4813]: I1201 08:40:42.240706 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:42 crc kubenswrapper[4813]: I1201 08:40:42.240772 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:42 crc kubenswrapper[4813]: I1201 08:40:42.240849 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:42 crc kubenswrapper[4813]: I1201 08:40:42.240919 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:42Z","lastTransitionTime":"2025-12-01T08:40:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:42 crc kubenswrapper[4813]: I1201 08:40:42.242081 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0023b0e358d75aacc889ce364846dda89ea513832c87fa4b066aea7dac57e8a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:42Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:42 crc kubenswrapper[4813]: I1201 08:40:42.264809 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:42Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:42 crc kubenswrapper[4813]: I1201 08:40:42.283915 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c7574f892d5e9142296b26a0b484eb6759ab53f571139fb6e23b93dc3da05780\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:42Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:42 crc kubenswrapper[4813]: I1201 08:40:42.294789 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:42Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:42 crc kubenswrapper[4813]: I1201 08:40:42.343358 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:42 crc kubenswrapper[4813]: I1201 08:40:42.343385 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:42 crc kubenswrapper[4813]: I1201 08:40:42.343393 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:42 crc kubenswrapper[4813]: I1201 08:40:42.343406 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:42 crc kubenswrapper[4813]: I1201 08:40:42.343420 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:42Z","lastTransitionTime":"2025-12-01T08:40:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:42 crc kubenswrapper[4813]: I1201 08:40:42.419177 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 08:40:42 crc kubenswrapper[4813]: E1201 08:40:42.419353 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 08:40:42 crc kubenswrapper[4813]: I1201 08:40:42.422801 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 08:40:42 crc kubenswrapper[4813]: E1201 08:40:42.423010 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 08:40:42 crc kubenswrapper[4813]: I1201 08:40:42.423145 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 08:40:42 crc kubenswrapper[4813]: E1201 08:40:42.423223 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 08:40:42 crc kubenswrapper[4813]: I1201 08:40:42.461654 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:42 crc kubenswrapper[4813]: I1201 08:40:42.461708 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:42 crc kubenswrapper[4813]: I1201 08:40:42.461726 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:42 crc kubenswrapper[4813]: I1201 08:40:42.506528 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:42 crc kubenswrapper[4813]: I1201 08:40:42.506584 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:42Z","lastTransitionTime":"2025-12-01T08:40:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:42 crc kubenswrapper[4813]: I1201 08:40:42.608639 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:42 crc kubenswrapper[4813]: I1201 08:40:42.608675 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:42 crc kubenswrapper[4813]: I1201 08:40:42.608685 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:42 crc kubenswrapper[4813]: I1201 08:40:42.608701 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:42 crc kubenswrapper[4813]: I1201 08:40:42.608713 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:42Z","lastTransitionTime":"2025-12-01T08:40:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:42 crc kubenswrapper[4813]: I1201 08:40:42.752226 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:42 crc kubenswrapper[4813]: I1201 08:40:42.752271 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:42 crc kubenswrapper[4813]: I1201 08:40:42.752284 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:42 crc kubenswrapper[4813]: I1201 08:40:42.752304 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:42 crc kubenswrapper[4813]: I1201 08:40:42.752317 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:42Z","lastTransitionTime":"2025-12-01T08:40:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:42 crc kubenswrapper[4813]: I1201 08:40:42.860097 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:42 crc kubenswrapper[4813]: I1201 08:40:42.860121 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:42 crc kubenswrapper[4813]: I1201 08:40:42.860128 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:42 crc kubenswrapper[4813]: I1201 08:40:42.860142 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:42 crc kubenswrapper[4813]: I1201 08:40:42.860150 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:42Z","lastTransitionTime":"2025-12-01T08:40:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:42 crc kubenswrapper[4813]: I1201 08:40:42.962472 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:42 crc kubenswrapper[4813]: I1201 08:40:42.962512 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:42 crc kubenswrapper[4813]: I1201 08:40:42.962520 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:42 crc kubenswrapper[4813]: I1201 08:40:42.962537 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:42 crc kubenswrapper[4813]: I1201 08:40:42.962546 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:42Z","lastTransitionTime":"2025-12-01T08:40:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:42 crc kubenswrapper[4813]: I1201 08:40:42.977921 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-9wn4q" event={"ID":"3d423402-e08f-42e9-b60c-6eb9bf2f6d48","Type":"ContainerStarted","Data":"b3049b842978c6580e1033a442a37766a2e76e278aface9a89fc0b2c57faf0df"} Dec 01 08:40:42 crc kubenswrapper[4813]: I1201 08:40:42.977983 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-9wn4q" event={"ID":"3d423402-e08f-42e9-b60c-6eb9bf2f6d48","Type":"ContainerStarted","Data":"1c915733072b32a63237e2401cbf84b059aca396c2c0db884f0f439d599a35d2"} Dec 01 08:40:42 crc kubenswrapper[4813]: I1201 08:40:42.979622 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-h7lvd" event={"ID":"d431b613-a91b-4fff-88b9-447804766289","Type":"ContainerStarted","Data":"495593f99e92a9c28100a95e27d06ff5847008c5d996366cb474bedd9b69d5ef"} Dec 01 08:40:42 crc kubenswrapper[4813]: I1201 08:40:42.980933 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-b6qqv" event={"ID":"b9656a03-0096-4c89-b0da-e37e4103611b","Type":"ContainerStarted","Data":"c4195aebc1b2e388d480920a7f72257f6660afc99ad8c1a09478ef445739b272"} Dec 01 08:40:42 crc kubenswrapper[4813]: I1201 08:40:42.982034 4813 generic.go:334] "Generic (PLEG): container finished" podID="b7505e14-3560-4d0d-be2c-d6a7a5953865" containerID="dd67634e0c806323334495401a4e20b48fea4bc5013acd973d2abd3d70870f99" exitCode=0 Dec 01 08:40:42 crc kubenswrapper[4813]: I1201 08:40:42.982063 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-j4pcc" event={"ID":"b7505e14-3560-4d0d-be2c-d6a7a5953865","Type":"ContainerDied","Data":"dd67634e0c806323334495401a4e20b48fea4bc5013acd973d2abd3d70870f99"} Dec 01 08:40:43 crc kubenswrapper[4813]: E1201 08:40:43.025879 4813 kubelet.go:1929] "Failed creating a mirror pod for" err="pods \"kube-controller-manager-crc\" already exists" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 01 08:40:43 crc kubenswrapper[4813]: I1201 08:40:43.036185 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-9wn4q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3d423402-e08f-42e9-b60c-6eb9bf2f6d48\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3049b842978c6580e1033a442a37766a2e76e278aface9a89fc0b2c57faf0df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rzxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c915733072b32a63237e2401cbf84b059aca396c2c0db884f0f439d599a35d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rzxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:41Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-9wn4q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:43Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:43 crc kubenswrapper[4813]: I1201 08:40:43.065437 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:43 crc kubenswrapper[4813]: I1201 08:40:43.065475 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:43 crc kubenswrapper[4813]: I1201 08:40:43.065483 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:43 crc kubenswrapper[4813]: I1201 08:40:43.065501 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:43 crc kubenswrapper[4813]: I1201 08:40:43.065511 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:43Z","lastTransitionTime":"2025-12-01T08:40:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:43 crc kubenswrapper[4813]: I1201 08:40:43.083772 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-j4pcc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b7505e14-3560-4d0d-be2c-d6a7a5953865\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:41Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-j4pcc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:43Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:43 crc kubenswrapper[4813]: I1201 08:40:43.105099 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-h7lvd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d431b613-a91b-4fff-88b9-447804766289\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:41Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-h7lvd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:43Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:43 crc kubenswrapper[4813]: I1201 08:40:43.168336 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:43 crc kubenswrapper[4813]: I1201 08:40:43.168370 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:43 crc kubenswrapper[4813]: I1201 08:40:43.168378 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:43 crc kubenswrapper[4813]: I1201 08:40:43.168394 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:43 crc kubenswrapper[4813]: I1201 08:40:43.168404 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:43Z","lastTransitionTime":"2025-12-01T08:40:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:43 crc kubenswrapper[4813]: I1201 08:40:43.235194 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:43Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:43 crc kubenswrapper[4813]: I1201 08:40:43.256987 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-c576v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"17292cd0-1eac-40b5-a60e-67167c57409c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a42206c914d0493ca74788ea1a941de80920ab7b810fb95227cd9b8711e8f4b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7zwfx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:41Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-c576v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:43Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:43 crc kubenswrapper[4813]: I1201 08:40:43.268526 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c7574f892d5e9142296b26a0b484eb6759ab53f571139fb6e23b93dc3da05780\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:43Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:43 crc kubenswrapper[4813]: I1201 08:40:43.269994 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:43 crc kubenswrapper[4813]: I1201 08:40:43.270027 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:43 crc kubenswrapper[4813]: I1201 08:40:43.270039 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:43 crc kubenswrapper[4813]: I1201 08:40:43.270055 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:43 crc kubenswrapper[4813]: I1201 08:40:43.270068 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:43Z","lastTransitionTime":"2025-12-01T08:40:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:43 crc kubenswrapper[4813]: I1201 08:40:43.291507 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:43Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:43 crc kubenswrapper[4813]: I1201 08:40:43.301920 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6fb5a648-ae4f-4e46-a530-007e9bf84abb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e24ebfc52dcb28368a2d436e7d9ab78f07f2aec2a4c71405a6c7d3301eee96a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33898118bbf639c8e6eae5aafa10386b95795512444ed5f535db3c663f169df2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9902c00bf7738686c93f62e1772a484d05715d74b75ee98c84c8f23ed383824b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3657b47ec9f1563c83bfbe3d728c8fd123f7078e219d251b1874269e42c31eea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:16Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:43Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:43 crc kubenswrapper[4813]: I1201 08:40:43.316762 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0023b0e358d75aacc889ce364846dda89ea513832c87fa4b066aea7dac57e8a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:43Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:43 crc kubenswrapper[4813]: I1201 08:40:43.334464 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:43Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:43 crc kubenswrapper[4813]: I1201 08:40:43.345783 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-b6qqv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9656a03-0096-4c89-b0da-e37e4103611b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cjvnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:41Z\\\"}}\" for pod \"openshift-multus\"/\"multus-b6qqv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:43Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:43 crc kubenswrapper[4813]: I1201 08:40:43.363075 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"349332bd-6b4d-4306-84ae-07019ca3860f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d441a348087ae3cadda1ad8a7d1c4dff02cf3f6e1d346c294b278fdd9131edcc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://da67635daf6d9e32937f4d5d4ceeee3f8f09ae4fd2dd26c3e85be1286c5949de\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e97356ee274ea78d9f237ffbc756abb6ac8df0dbac244bf0a86ed215b8cbf43\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0404f5a5ed5a9750d4c479b608e68e8960b448f455d5af9d219aa5edaeb789b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a0404f5a5ed5a9750d4c479b608e68e8960b448f455d5af9d219aa5edaeb789b\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"message\\\":\\\"\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764578421\\\\\\\\\\\\\\\" (2025-12-01 08:40:21 +0000 UTC to 2025-12-31 08:40:22 +0000 UTC (now=2025-12-01 08:40:38.066068451 +0000 UTC))\\\\\\\"\\\\nI1201 08:40:38.066144 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1201 08:40:38.066166 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1201 08:40:38.066220 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1201 08:40:38.066232 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1201 08:40:38.066338 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764578432\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764578432\\\\\\\\\\\\\\\" (2025-12-01 07:40:32 +0000 UTC to 2026-12-01 07:40:32 +0000 UTC (now=2025-12-01 08:40:38.066301217 +0000 UTC))\\\\\\\"\\\\nI1201 08:40:38.066357 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1201 08:40:38.066385 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1201 08:40:38.066402 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1201 08:40:38.066414 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-63890811/tls.crt::/tmp/serving-cert-63890811/tls.key\\\\\\\"\\\\nI1201 08:40:38.066419 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1201 08:40:38.066572 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1201 08:40:38.067159 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7354f9337d42115bb1d974bf3772f2d03365acba57f2155135c7a0bbab39879e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:19Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05492df5703392cdf0932d6bf66025b18e53bd05fa0857a7cb2ad70979cbdc3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05492df5703392cdf0932d6bf66025b18e53bd05fa0857a7cb2ad70979cbdc3e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:16Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:43Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:43 crc kubenswrapper[4813]: I1201 08:40:43.372158 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:43 crc kubenswrapper[4813]: I1201 08:40:43.372196 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:43 crc kubenswrapper[4813]: I1201 08:40:43.372205 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:43 crc kubenswrapper[4813]: I1201 08:40:43.372219 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:43 crc kubenswrapper[4813]: I1201 08:40:43.372227 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:43Z","lastTransitionTime":"2025-12-01T08:40:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:43 crc kubenswrapper[4813]: I1201 08:40:43.384167 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3264b3af996af452df2953c59bea6deeb5ea6af1b119ce89f7334595c27e70da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9e879b1f563bd66383d58fd7b435a8a81bd50cb387dc1fabb6fc47d5a3fa468\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:43Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:43 crc kubenswrapper[4813]: I1201 08:40:43.416335 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"349332bd-6b4d-4306-84ae-07019ca3860f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d441a348087ae3cadda1ad8a7d1c4dff02cf3f6e1d346c294b278fdd9131edcc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://da67635daf6d9e32937f4d5d4ceeee3f8f09ae4fd2dd26c3e85be1286c5949de\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e97356ee274ea78d9f237ffbc756abb6ac8df0dbac244bf0a86ed215b8cbf43\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0404f5a5ed5a9750d4c479b608e68e8960b448f455d5af9d219aa5edaeb789b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a0404f5a5ed5a9750d4c479b608e68e8960b448f455d5af9d219aa5edaeb789b\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"message\\\":\\\"\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764578421\\\\\\\\\\\\\\\" (2025-12-01 08:40:21 +0000 UTC to 2025-12-31 08:40:22 +0000 UTC (now=2025-12-01 08:40:38.066068451 +0000 UTC))\\\\\\\"\\\\nI1201 08:40:38.066144 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1201 08:40:38.066166 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1201 08:40:38.066220 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1201 08:40:38.066232 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1201 08:40:38.066338 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764578432\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764578432\\\\\\\\\\\\\\\" (2025-12-01 07:40:32 +0000 UTC to 2026-12-01 07:40:32 +0000 UTC (now=2025-12-01 08:40:38.066301217 +0000 UTC))\\\\\\\"\\\\nI1201 08:40:38.066357 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1201 08:40:38.066385 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1201 08:40:38.066402 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1201 08:40:38.066414 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-63890811/tls.crt::/tmp/serving-cert-63890811/tls.key\\\\\\\"\\\\nI1201 08:40:38.066419 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1201 08:40:38.066572 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1201 08:40:38.067159 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7354f9337d42115bb1d974bf3772f2d03365acba57f2155135c7a0bbab39879e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:19Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05492df5703392cdf0932d6bf66025b18e53bd05fa0857a7cb2ad70979cbdc3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05492df5703392cdf0932d6bf66025b18e53bd05fa0857a7cb2ad70979cbdc3e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:16Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:43Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:43 crc kubenswrapper[4813]: I1201 08:40:43.432588 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3264b3af996af452df2953c59bea6deeb5ea6af1b119ce89f7334595c27e70da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9e879b1f563bd66383d58fd7b435a8a81bd50cb387dc1fabb6fc47d5a3fa468\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:43Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:43 crc kubenswrapper[4813]: I1201 08:40:43.447248 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-b6qqv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9656a03-0096-4c89-b0da-e37e4103611b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4195aebc1b2e388d480920a7f72257f6660afc99ad8c1a09478ef445739b272\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cjvnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:41Z\\\"}}\" for pod \"openshift-multus\"/\"multus-b6qqv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:43Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:43 crc kubenswrapper[4813]: I1201 08:40:43.464064 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-9wn4q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3d423402-e08f-42e9-b60c-6eb9bf2f6d48\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3049b842978c6580e1033a442a37766a2e76e278aface9a89fc0b2c57faf0df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rzxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c915733072b32a63237e2401cbf84b059aca396c2c0db884f0f439d599a35d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rzxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:41Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-9wn4q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:43Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:43 crc kubenswrapper[4813]: I1201 08:40:43.474663 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:43 crc kubenswrapper[4813]: I1201 08:40:43.474721 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:43 crc kubenswrapper[4813]: I1201 08:40:43.474736 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:43 crc kubenswrapper[4813]: I1201 08:40:43.474754 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:43 crc kubenswrapper[4813]: I1201 08:40:43.474767 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:43Z","lastTransitionTime":"2025-12-01T08:40:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:43 crc kubenswrapper[4813]: I1201 08:40:43.478327 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:43Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:43 crc kubenswrapper[4813]: I1201 08:40:43.492713 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-c576v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"17292cd0-1eac-40b5-a60e-67167c57409c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a42206c914d0493ca74788ea1a941de80920ab7b810fb95227cd9b8711e8f4b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7zwfx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:41Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-c576v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:43Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:43 crc kubenswrapper[4813]: I1201 08:40:43.512990 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-j4pcc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b7505e14-3560-4d0d-be2c-d6a7a5953865\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd67634e0c806323334495401a4e20b48fea4bc5013acd973d2abd3d70870f99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd67634e0c806323334495401a4e20b48fea4bc5013acd973d2abd3d70870f99\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:41Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-j4pcc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:43Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:43 crc kubenswrapper[4813]: I1201 08:40:43.530571 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-h7lvd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d431b613-a91b-4fff-88b9-447804766289\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://495593f99e92a9c28100a95e27d06ff5847008c5d996366cb474bedd9b69d5ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:41Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-h7lvd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:43Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:43 crc kubenswrapper[4813]: I1201 08:40:43.545413 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6fb5a648-ae4f-4e46-a530-007e9bf84abb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e24ebfc52dcb28368a2d436e7d9ab78f07f2aec2a4c71405a6c7d3301eee96a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33898118bbf639c8e6eae5aafa10386b95795512444ed5f535db3c663f169df2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9902c00bf7738686c93f62e1772a484d05715d74b75ee98c84c8f23ed383824b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3657b47ec9f1563c83bfbe3d728c8fd123f7078e219d251b1874269e42c31eea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:16Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:43Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:43 crc kubenswrapper[4813]: I1201 08:40:43.561868 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0023b0e358d75aacc889ce364846dda89ea513832c87fa4b066aea7dac57e8a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:43Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:43 crc kubenswrapper[4813]: I1201 08:40:43.573556 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:43Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:43 crc kubenswrapper[4813]: I1201 08:40:43.579001 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:43 crc kubenswrapper[4813]: I1201 08:40:43.579040 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:43 crc kubenswrapper[4813]: I1201 08:40:43.579052 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:43 crc kubenswrapper[4813]: I1201 08:40:43.579068 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:43 crc kubenswrapper[4813]: I1201 08:40:43.579078 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:43Z","lastTransitionTime":"2025-12-01T08:40:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:43 crc kubenswrapper[4813]: I1201 08:40:43.588143 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c7574f892d5e9142296b26a0b484eb6759ab53f571139fb6e23b93dc3da05780\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:43Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:43 crc kubenswrapper[4813]: I1201 08:40:43.597380 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:43Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:43 crc kubenswrapper[4813]: I1201 08:40:43.626859 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/node-ca-xpnlc"] Dec 01 08:40:43 crc kubenswrapper[4813]: I1201 08:40:43.627416 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-xpnlc" Dec 01 08:40:43 crc kubenswrapper[4813]: I1201 08:40:43.629240 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Dec 01 08:40:43 crc kubenswrapper[4813]: I1201 08:40:43.629688 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Dec 01 08:40:43 crc kubenswrapper[4813]: I1201 08:40:43.629860 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Dec 01 08:40:43 crc kubenswrapper[4813]: I1201 08:40:43.632102 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Dec 01 08:40:43 crc kubenswrapper[4813]: I1201 08:40:43.644692 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"349332bd-6b4d-4306-84ae-07019ca3860f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d441a348087ae3cadda1ad8a7d1c4dff02cf3f6e1d346c294b278fdd9131edcc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://da67635daf6d9e32937f4d5d4ceeee3f8f09ae4fd2dd26c3e85be1286c5949de\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e97356ee274ea78d9f237ffbc756abb6ac8df0dbac244bf0a86ed215b8cbf43\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0404f5a5ed5a9750d4c479b608e68e8960b448f455d5af9d219aa5edaeb789b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a0404f5a5ed5a9750d4c479b608e68e8960b448f455d5af9d219aa5edaeb789b\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"message\\\":\\\"\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764578421\\\\\\\\\\\\\\\" (2025-12-01 08:40:21 +0000 UTC to 2025-12-31 08:40:22 +0000 UTC (now=2025-12-01 08:40:38.066068451 +0000 UTC))\\\\\\\"\\\\nI1201 08:40:38.066144 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1201 08:40:38.066166 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1201 08:40:38.066220 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1201 08:40:38.066232 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1201 08:40:38.066338 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764578432\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764578432\\\\\\\\\\\\\\\" (2025-12-01 07:40:32 +0000 UTC to 2026-12-01 07:40:32 +0000 UTC (now=2025-12-01 08:40:38.066301217 +0000 UTC))\\\\\\\"\\\\nI1201 08:40:38.066357 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1201 08:40:38.066385 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1201 08:40:38.066402 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1201 08:40:38.066414 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-63890811/tls.crt::/tmp/serving-cert-63890811/tls.key\\\\\\\"\\\\nI1201 08:40:38.066419 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1201 08:40:38.066572 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1201 08:40:38.067159 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7354f9337d42115bb1d974bf3772f2d03365acba57f2155135c7a0bbab39879e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:19Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05492df5703392cdf0932d6bf66025b18e53bd05fa0857a7cb2ad70979cbdc3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05492df5703392cdf0932d6bf66025b18e53bd05fa0857a7cb2ad70979cbdc3e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:16Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:43Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:43 crc kubenswrapper[4813]: I1201 08:40:43.664811 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3264b3af996af452df2953c59bea6deeb5ea6af1b119ce89f7334595c27e70da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9e879b1f563bd66383d58fd7b435a8a81bd50cb387dc1fabb6fc47d5a3fa468\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:43Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:43 crc kubenswrapper[4813]: I1201 08:40:43.678844 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-b6qqv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9656a03-0096-4c89-b0da-e37e4103611b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4195aebc1b2e388d480920a7f72257f6660afc99ad8c1a09478ef445739b272\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cjvnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:41Z\\\"}}\" for pod \"openshift-multus\"/\"multus-b6qqv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:43Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:43 crc kubenswrapper[4813]: I1201 08:40:43.687081 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-9wn4q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3d423402-e08f-42e9-b60c-6eb9bf2f6d48\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3049b842978c6580e1033a442a37766a2e76e278aface9a89fc0b2c57faf0df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rzxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c915733072b32a63237e2401cbf84b059aca396c2c0db884f0f439d599a35d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rzxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:41Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-9wn4q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:43Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:43 crc kubenswrapper[4813]: I1201 08:40:43.697137 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:43Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:43 crc kubenswrapper[4813]: I1201 08:40:43.714567 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-c576v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"17292cd0-1eac-40b5-a60e-67167c57409c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a42206c914d0493ca74788ea1a941de80920ab7b810fb95227cd9b8711e8f4b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7zwfx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:41Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-c576v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:43Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:43 crc kubenswrapper[4813]: I1201 08:40:43.715830 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:43 crc kubenswrapper[4813]: I1201 08:40:43.715879 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:43 crc kubenswrapper[4813]: I1201 08:40:43.715889 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:43 crc kubenswrapper[4813]: I1201 08:40:43.715908 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:43 crc kubenswrapper[4813]: I1201 08:40:43.715918 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:43Z","lastTransitionTime":"2025-12-01T08:40:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:43 crc kubenswrapper[4813]: I1201 08:40:43.731787 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/22da1fcf-0c57-4708-90d8-77692af17344-serviceca\") pod \"node-ca-xpnlc\" (UID: \"22da1fcf-0c57-4708-90d8-77692af17344\") " pod="openshift-image-registry/node-ca-xpnlc" Dec 01 08:40:43 crc kubenswrapper[4813]: I1201 08:40:43.731849 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xc44d\" (UniqueName: \"kubernetes.io/projected/22da1fcf-0c57-4708-90d8-77692af17344-kube-api-access-xc44d\") pod \"node-ca-xpnlc\" (UID: \"22da1fcf-0c57-4708-90d8-77692af17344\") " pod="openshift-image-registry/node-ca-xpnlc" Dec 01 08:40:43 crc kubenswrapper[4813]: I1201 08:40:43.731866 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/22da1fcf-0c57-4708-90d8-77692af17344-host\") pod \"node-ca-xpnlc\" (UID: \"22da1fcf-0c57-4708-90d8-77692af17344\") " pod="openshift-image-registry/node-ca-xpnlc" Dec 01 08:40:43 crc kubenswrapper[4813]: I1201 08:40:43.739090 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-j4pcc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b7505e14-3560-4d0d-be2c-d6a7a5953865\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd67634e0c806323334495401a4e20b48fea4bc5013acd973d2abd3d70870f99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd67634e0c806323334495401a4e20b48fea4bc5013acd973d2abd3d70870f99\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:41Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-j4pcc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:43Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:43 crc kubenswrapper[4813]: I1201 08:40:43.753674 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-h7lvd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d431b613-a91b-4fff-88b9-447804766289\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://495593f99e92a9c28100a95e27d06ff5847008c5d996366cb474bedd9b69d5ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:41Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-h7lvd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:43Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:43 crc kubenswrapper[4813]: I1201 08:40:43.774870 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0023b0e358d75aacc889ce364846dda89ea513832c87fa4b066aea7dac57e8a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:43Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:43 crc kubenswrapper[4813]: I1201 08:40:43.789037 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:43Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:43 crc kubenswrapper[4813]: I1201 08:40:43.800900 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c7574f892d5e9142296b26a0b484eb6759ab53f571139fb6e23b93dc3da05780\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:43Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:43 crc kubenswrapper[4813]: I1201 08:40:43.813980 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:43Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:43 crc kubenswrapper[4813]: I1201 08:40:43.818766 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:43 crc kubenswrapper[4813]: I1201 08:40:43.818814 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:43 crc kubenswrapper[4813]: I1201 08:40:43.818824 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:43 crc kubenswrapper[4813]: I1201 08:40:43.818841 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:43 crc kubenswrapper[4813]: I1201 08:40:43.818850 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:43Z","lastTransitionTime":"2025-12-01T08:40:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:43 crc kubenswrapper[4813]: I1201 08:40:43.825747 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-xpnlc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"22da1fcf-0c57-4708-90d8-77692af17344\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xc44d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-xpnlc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:43Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:43 crc kubenswrapper[4813]: I1201 08:40:43.832624 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xc44d\" (UniqueName: \"kubernetes.io/projected/22da1fcf-0c57-4708-90d8-77692af17344-kube-api-access-xc44d\") pod \"node-ca-xpnlc\" (UID: \"22da1fcf-0c57-4708-90d8-77692af17344\") " pod="openshift-image-registry/node-ca-xpnlc" Dec 01 08:40:43 crc kubenswrapper[4813]: I1201 08:40:43.832750 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/22da1fcf-0c57-4708-90d8-77692af17344-host\") pod \"node-ca-xpnlc\" (UID: \"22da1fcf-0c57-4708-90d8-77692af17344\") " pod="openshift-image-registry/node-ca-xpnlc" Dec 01 08:40:43 crc kubenswrapper[4813]: I1201 08:40:43.832867 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/22da1fcf-0c57-4708-90d8-77692af17344-host\") pod \"node-ca-xpnlc\" (UID: \"22da1fcf-0c57-4708-90d8-77692af17344\") " pod="openshift-image-registry/node-ca-xpnlc" Dec 01 08:40:43 crc kubenswrapper[4813]: I1201 08:40:43.832946 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/22da1fcf-0c57-4708-90d8-77692af17344-serviceca\") pod \"node-ca-xpnlc\" (UID: \"22da1fcf-0c57-4708-90d8-77692af17344\") " pod="openshift-image-registry/node-ca-xpnlc" Dec 01 08:40:43 crc kubenswrapper[4813]: I1201 08:40:43.833773 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/22da1fcf-0c57-4708-90d8-77692af17344-serviceca\") pod \"node-ca-xpnlc\" (UID: \"22da1fcf-0c57-4708-90d8-77692af17344\") " pod="openshift-image-registry/node-ca-xpnlc" Dec 01 08:40:43 crc kubenswrapper[4813]: I1201 08:40:43.844439 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6fb5a648-ae4f-4e46-a530-007e9bf84abb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e24ebfc52dcb28368a2d436e7d9ab78f07f2aec2a4c71405a6c7d3301eee96a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33898118bbf639c8e6eae5aafa10386b95795512444ed5f535db3c663f169df2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9902c00bf7738686c93f62e1772a484d05715d74b75ee98c84c8f23ed383824b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3657b47ec9f1563c83bfbe3d728c8fd123f7078e219d251b1874269e42c31eea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:16Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:43Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:43 crc kubenswrapper[4813]: I1201 08:40:43.858425 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xc44d\" (UniqueName: \"kubernetes.io/projected/22da1fcf-0c57-4708-90d8-77692af17344-kube-api-access-xc44d\") pod \"node-ca-xpnlc\" (UID: \"22da1fcf-0c57-4708-90d8-77692af17344\") " pod="openshift-image-registry/node-ca-xpnlc" Dec 01 08:40:43 crc kubenswrapper[4813]: I1201 08:40:43.924239 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:43 crc kubenswrapper[4813]: I1201 08:40:43.924281 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:43 crc kubenswrapper[4813]: I1201 08:40:43.924291 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:43 crc kubenswrapper[4813]: I1201 08:40:43.924311 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:43 crc kubenswrapper[4813]: I1201 08:40:43.924321 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:43Z","lastTransitionTime":"2025-12-01T08:40:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:43 crc kubenswrapper[4813]: I1201 08:40:43.940867 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-xpnlc" Dec 01 08:40:44 crc kubenswrapper[4813]: I1201 08:40:44.026632 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:44 crc kubenswrapper[4813]: I1201 08:40:44.026675 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:44 crc kubenswrapper[4813]: I1201 08:40:44.026687 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:44 crc kubenswrapper[4813]: I1201 08:40:44.026702 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:44 crc kubenswrapper[4813]: I1201 08:40:44.026710 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:44Z","lastTransitionTime":"2025-12-01T08:40:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:44 crc kubenswrapper[4813]: W1201 08:40:44.032847 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod22da1fcf_0c57_4708_90d8_77692af17344.slice/crio-93d12e4efa331a594dd87abfb249200439214260bcfd0c029e6a24e3d9929560 WatchSource:0}: Error finding container 93d12e4efa331a594dd87abfb249200439214260bcfd0c029e6a24e3d9929560: Status 404 returned error can't find the container with id 93d12e4efa331a594dd87abfb249200439214260bcfd0c029e6a24e3d9929560 Dec 01 08:40:44 crc kubenswrapper[4813]: I1201 08:40:44.045315 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-j4pcc" event={"ID":"b7505e14-3560-4d0d-be2c-d6a7a5953865","Type":"ContainerStarted","Data":"7f19c67b1c6c80c7f3223512102810d46ede1445464d21ae3bbea83424539b2b"} Dec 01 08:40:44 crc kubenswrapper[4813]: I1201 08:40:44.045378 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-j4pcc" event={"ID":"b7505e14-3560-4d0d-be2c-d6a7a5953865","Type":"ContainerStarted","Data":"944977e9722025088720107530d0c91fc693593f380e4b49678f28a56da180cc"} Dec 01 08:40:44 crc kubenswrapper[4813]: I1201 08:40:44.045392 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-j4pcc" event={"ID":"b7505e14-3560-4d0d-be2c-d6a7a5953865","Type":"ContainerStarted","Data":"ccc30a63dc1c23ff9880fb04af1f31f0656b2696d359988e23deed23989f480a"} Dec 01 08:40:44 crc kubenswrapper[4813]: I1201 08:40:44.045443 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-j4pcc" event={"ID":"b7505e14-3560-4d0d-be2c-d6a7a5953865","Type":"ContainerStarted","Data":"9d984c90302e57f3e5a6f3ca3613b415db04ac74a5a228582f1b52ef1d59f93e"} Dec 01 08:40:44 crc kubenswrapper[4813]: I1201 08:40:44.045452 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-j4pcc" event={"ID":"b7505e14-3560-4d0d-be2c-d6a7a5953865","Type":"ContainerStarted","Data":"ce8963bf1e9ee66926907aaf3a03d18977f4d0a1983d7121b62fbc8e02a37e4d"} Dec 01 08:40:44 crc kubenswrapper[4813]: I1201 08:40:44.048494 4813 generic.go:334] "Generic (PLEG): container finished" podID="d431b613-a91b-4fff-88b9-447804766289" containerID="495593f99e92a9c28100a95e27d06ff5847008c5d996366cb474bedd9b69d5ef" exitCode=0 Dec 01 08:40:44 crc kubenswrapper[4813]: I1201 08:40:44.048704 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-h7lvd" event={"ID":"d431b613-a91b-4fff-88b9-447804766289","Type":"ContainerDied","Data":"495593f99e92a9c28100a95e27d06ff5847008c5d996366cb474bedd9b69d5ef"} Dec 01 08:40:44 crc kubenswrapper[4813]: I1201 08:40:44.069361 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-xpnlc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"22da1fcf-0c57-4708-90d8-77692af17344\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xc44d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-xpnlc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:44Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:44 crc kubenswrapper[4813]: I1201 08:40:44.084720 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6fb5a648-ae4f-4e46-a530-007e9bf84abb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e24ebfc52dcb28368a2d436e7d9ab78f07f2aec2a4c71405a6c7d3301eee96a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33898118bbf639c8e6eae5aafa10386b95795512444ed5f535db3c663f169df2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9902c00bf7738686c93f62e1772a484d05715d74b75ee98c84c8f23ed383824b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3657b47ec9f1563c83bfbe3d728c8fd123f7078e219d251b1874269e42c31eea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:16Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:44Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:44 crc kubenswrapper[4813]: I1201 08:40:44.097029 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0023b0e358d75aacc889ce364846dda89ea513832c87fa4b066aea7dac57e8a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:44Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:44 crc kubenswrapper[4813]: I1201 08:40:44.109168 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:44Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:44 crc kubenswrapper[4813]: I1201 08:40:44.121288 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c7574f892d5e9142296b26a0b484eb6759ab53f571139fb6e23b93dc3da05780\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:44Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:44 crc kubenswrapper[4813]: I1201 08:40:44.130113 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:44 crc kubenswrapper[4813]: I1201 08:40:44.130151 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:44 crc kubenswrapper[4813]: I1201 08:40:44.130199 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:44 crc kubenswrapper[4813]: I1201 08:40:44.130217 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:44 crc kubenswrapper[4813]: I1201 08:40:44.130227 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:44Z","lastTransitionTime":"2025-12-01T08:40:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:44 crc kubenswrapper[4813]: I1201 08:40:44.132769 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:44Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:44 crc kubenswrapper[4813]: I1201 08:40:44.148035 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"349332bd-6b4d-4306-84ae-07019ca3860f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d441a348087ae3cadda1ad8a7d1c4dff02cf3f6e1d346c294b278fdd9131edcc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://da67635daf6d9e32937f4d5d4ceeee3f8f09ae4fd2dd26c3e85be1286c5949de\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e97356ee274ea78d9f237ffbc756abb6ac8df0dbac244bf0a86ed215b8cbf43\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0404f5a5ed5a9750d4c479b608e68e8960b448f455d5af9d219aa5edaeb789b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a0404f5a5ed5a9750d4c479b608e68e8960b448f455d5af9d219aa5edaeb789b\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"message\\\":\\\"\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764578421\\\\\\\\\\\\\\\" (2025-12-01 08:40:21 +0000 UTC to 2025-12-31 08:40:22 +0000 UTC (now=2025-12-01 08:40:38.066068451 +0000 UTC))\\\\\\\"\\\\nI1201 08:40:38.066144 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1201 08:40:38.066166 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1201 08:40:38.066220 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1201 08:40:38.066232 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1201 08:40:38.066338 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764578432\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764578432\\\\\\\\\\\\\\\" (2025-12-01 07:40:32 +0000 UTC to 2026-12-01 07:40:32 +0000 UTC (now=2025-12-01 08:40:38.066301217 +0000 UTC))\\\\\\\"\\\\nI1201 08:40:38.066357 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1201 08:40:38.066385 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1201 08:40:38.066402 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1201 08:40:38.066414 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-63890811/tls.crt::/tmp/serving-cert-63890811/tls.key\\\\\\\"\\\\nI1201 08:40:38.066419 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1201 08:40:38.066572 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1201 08:40:38.067159 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7354f9337d42115bb1d974bf3772f2d03365acba57f2155135c7a0bbab39879e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:19Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05492df5703392cdf0932d6bf66025b18e53bd05fa0857a7cb2ad70979cbdc3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05492df5703392cdf0932d6bf66025b18e53bd05fa0857a7cb2ad70979cbdc3e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:16Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:44Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:44 crc kubenswrapper[4813]: I1201 08:40:44.161294 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3264b3af996af452df2953c59bea6deeb5ea6af1b119ce89f7334595c27e70da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9e879b1f563bd66383d58fd7b435a8a81bd50cb387dc1fabb6fc47d5a3fa468\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:44Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:44 crc kubenswrapper[4813]: I1201 08:40:44.182693 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-b6qqv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9656a03-0096-4c89-b0da-e37e4103611b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4195aebc1b2e388d480920a7f72257f6660afc99ad8c1a09478ef445739b272\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cjvnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:41Z\\\"}}\" for pod \"openshift-multus\"/\"multus-b6qqv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:44Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:44 crc kubenswrapper[4813]: I1201 08:40:44.196791 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-9wn4q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3d423402-e08f-42e9-b60c-6eb9bf2f6d48\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3049b842978c6580e1033a442a37766a2e76e278aface9a89fc0b2c57faf0df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rzxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c915733072b32a63237e2401cbf84b059aca396c2c0db884f0f439d599a35d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rzxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:41Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-9wn4q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:44Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:44 crc kubenswrapper[4813]: I1201 08:40:44.209896 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:44Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:44 crc kubenswrapper[4813]: I1201 08:40:44.220239 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-c576v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"17292cd0-1eac-40b5-a60e-67167c57409c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a42206c914d0493ca74788ea1a941de80920ab7b810fb95227cd9b8711e8f4b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7zwfx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:41Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-c576v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:44Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:44 crc kubenswrapper[4813]: I1201 08:40:44.233927 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:44 crc kubenswrapper[4813]: I1201 08:40:44.233992 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:44 crc kubenswrapper[4813]: I1201 08:40:44.234004 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:44 crc kubenswrapper[4813]: I1201 08:40:44.234045 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:44 crc kubenswrapper[4813]: I1201 08:40:44.234056 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:44Z","lastTransitionTime":"2025-12-01T08:40:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:44 crc kubenswrapper[4813]: I1201 08:40:44.242730 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-j4pcc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b7505e14-3560-4d0d-be2c-d6a7a5953865\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd67634e0c806323334495401a4e20b48fea4bc5013acd973d2abd3d70870f99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd67634e0c806323334495401a4e20b48fea4bc5013acd973d2abd3d70870f99\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:41Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-j4pcc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:44Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:44 crc kubenswrapper[4813]: I1201 08:40:44.257549 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-h7lvd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d431b613-a91b-4fff-88b9-447804766289\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://495593f99e92a9c28100a95e27d06ff5847008c5d996366cb474bedd9b69d5ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://495593f99e92a9c28100a95e27d06ff5847008c5d996366cb474bedd9b69d5ef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:41Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-h7lvd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:44Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:44 crc kubenswrapper[4813]: I1201 08:40:44.336700 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:44 crc kubenswrapper[4813]: I1201 08:40:44.336754 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:44 crc kubenswrapper[4813]: I1201 08:40:44.336767 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:44 crc kubenswrapper[4813]: I1201 08:40:44.336784 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:44 crc kubenswrapper[4813]: I1201 08:40:44.336794 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:44Z","lastTransitionTime":"2025-12-01T08:40:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:44 crc kubenswrapper[4813]: I1201 08:40:44.393595 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 08:40:44 crc kubenswrapper[4813]: I1201 08:40:44.393622 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 08:40:44 crc kubenswrapper[4813]: E1201 08:40:44.394135 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 08:40:44 crc kubenswrapper[4813]: E1201 08:40:44.394001 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 08:40:44 crc kubenswrapper[4813]: I1201 08:40:44.393669 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 08:40:44 crc kubenswrapper[4813]: E1201 08:40:44.394220 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 08:40:44 crc kubenswrapper[4813]: I1201 08:40:44.438930 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:44 crc kubenswrapper[4813]: I1201 08:40:44.438997 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:44 crc kubenswrapper[4813]: I1201 08:40:44.439009 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:44 crc kubenswrapper[4813]: I1201 08:40:44.439026 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:44 crc kubenswrapper[4813]: I1201 08:40:44.439037 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:44Z","lastTransitionTime":"2025-12-01T08:40:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:44 crc kubenswrapper[4813]: I1201 08:40:44.541283 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:44 crc kubenswrapper[4813]: I1201 08:40:44.541331 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:44 crc kubenswrapper[4813]: I1201 08:40:44.541340 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:44 crc kubenswrapper[4813]: I1201 08:40:44.541356 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:44 crc kubenswrapper[4813]: I1201 08:40:44.541367 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:44Z","lastTransitionTime":"2025-12-01T08:40:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:44 crc kubenswrapper[4813]: I1201 08:40:44.643543 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:44 crc kubenswrapper[4813]: I1201 08:40:44.643582 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:44 crc kubenswrapper[4813]: I1201 08:40:44.643591 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:44 crc kubenswrapper[4813]: I1201 08:40:44.643608 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:44 crc kubenswrapper[4813]: I1201 08:40:44.643618 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:44Z","lastTransitionTime":"2025-12-01T08:40:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:44 crc kubenswrapper[4813]: I1201 08:40:44.746200 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:44 crc kubenswrapper[4813]: I1201 08:40:44.746230 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:44 crc kubenswrapper[4813]: I1201 08:40:44.746241 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:44 crc kubenswrapper[4813]: I1201 08:40:44.746258 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:44 crc kubenswrapper[4813]: I1201 08:40:44.746269 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:44Z","lastTransitionTime":"2025-12-01T08:40:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:44 crc kubenswrapper[4813]: I1201 08:40:44.849620 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:44 crc kubenswrapper[4813]: I1201 08:40:44.849672 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:44 crc kubenswrapper[4813]: I1201 08:40:44.849683 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:44 crc kubenswrapper[4813]: I1201 08:40:44.849705 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:44 crc kubenswrapper[4813]: I1201 08:40:44.849716 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:44Z","lastTransitionTime":"2025-12-01T08:40:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:44 crc kubenswrapper[4813]: I1201 08:40:44.951946 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:44 crc kubenswrapper[4813]: I1201 08:40:44.951999 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:44 crc kubenswrapper[4813]: I1201 08:40:44.952011 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:44 crc kubenswrapper[4813]: I1201 08:40:44.952029 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:44 crc kubenswrapper[4813]: I1201 08:40:44.952040 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:44Z","lastTransitionTime":"2025-12-01T08:40:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:45 crc kubenswrapper[4813]: I1201 08:40:45.055751 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:45 crc kubenswrapper[4813]: I1201 08:40:45.055950 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:45 crc kubenswrapper[4813]: I1201 08:40:45.056033 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:45 crc kubenswrapper[4813]: I1201 08:40:45.056119 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:45 crc kubenswrapper[4813]: I1201 08:40:45.056190 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:45Z","lastTransitionTime":"2025-12-01T08:40:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:45 crc kubenswrapper[4813]: I1201 08:40:45.058521 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-j4pcc" event={"ID":"b7505e14-3560-4d0d-be2c-d6a7a5953865","Type":"ContainerStarted","Data":"dc32d930ab23f30103d8fa42291909cbec89ebb75c5c53cd783cd27fe1a3f09e"} Dec 01 08:40:45 crc kubenswrapper[4813]: I1201 08:40:45.059182 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-xpnlc" event={"ID":"22da1fcf-0c57-4708-90d8-77692af17344","Type":"ContainerStarted","Data":"93d12e4efa331a594dd87abfb249200439214260bcfd0c029e6a24e3d9929560"} Dec 01 08:40:45 crc kubenswrapper[4813]: I1201 08:40:45.161258 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:45 crc kubenswrapper[4813]: I1201 08:40:45.161312 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:45 crc kubenswrapper[4813]: I1201 08:40:45.161323 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:45 crc kubenswrapper[4813]: I1201 08:40:45.161342 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:45 crc kubenswrapper[4813]: I1201 08:40:45.161352 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:45Z","lastTransitionTime":"2025-12-01T08:40:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:45 crc kubenswrapper[4813]: I1201 08:40:45.265003 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:45 crc kubenswrapper[4813]: I1201 08:40:45.265380 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:45 crc kubenswrapper[4813]: I1201 08:40:45.265390 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:45 crc kubenswrapper[4813]: I1201 08:40:45.265410 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:45 crc kubenswrapper[4813]: I1201 08:40:45.265420 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:45Z","lastTransitionTime":"2025-12-01T08:40:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:45 crc kubenswrapper[4813]: I1201 08:40:45.369853 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:45 crc kubenswrapper[4813]: I1201 08:40:45.370423 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:45 crc kubenswrapper[4813]: I1201 08:40:45.370447 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:45 crc kubenswrapper[4813]: I1201 08:40:45.370464 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:45 crc kubenswrapper[4813]: I1201 08:40:45.370488 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:45Z","lastTransitionTime":"2025-12-01T08:40:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:45 crc kubenswrapper[4813]: I1201 08:40:45.476118 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:45 crc kubenswrapper[4813]: I1201 08:40:45.476182 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:45 crc kubenswrapper[4813]: I1201 08:40:45.476194 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:45 crc kubenswrapper[4813]: I1201 08:40:45.476213 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:45 crc kubenswrapper[4813]: I1201 08:40:45.476225 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:45Z","lastTransitionTime":"2025-12-01T08:40:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:45 crc kubenswrapper[4813]: I1201 08:40:45.578899 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:45 crc kubenswrapper[4813]: I1201 08:40:45.578945 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:45 crc kubenswrapper[4813]: I1201 08:40:45.578955 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:45 crc kubenswrapper[4813]: I1201 08:40:45.579056 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:45 crc kubenswrapper[4813]: I1201 08:40:45.579072 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:45Z","lastTransitionTime":"2025-12-01T08:40:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:45 crc kubenswrapper[4813]: I1201 08:40:45.681797 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:45 crc kubenswrapper[4813]: I1201 08:40:45.681860 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:45 crc kubenswrapper[4813]: I1201 08:40:45.681882 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:45 crc kubenswrapper[4813]: I1201 08:40:45.681913 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:45 crc kubenswrapper[4813]: I1201 08:40:45.681937 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:45Z","lastTransitionTime":"2025-12-01T08:40:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:45 crc kubenswrapper[4813]: I1201 08:40:45.784699 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:45 crc kubenswrapper[4813]: I1201 08:40:45.784731 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:45 crc kubenswrapper[4813]: I1201 08:40:45.784739 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:45 crc kubenswrapper[4813]: I1201 08:40:45.784755 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:45 crc kubenswrapper[4813]: I1201 08:40:45.784763 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:45Z","lastTransitionTime":"2025-12-01T08:40:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:45 crc kubenswrapper[4813]: I1201 08:40:45.887815 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:45 crc kubenswrapper[4813]: I1201 08:40:45.887883 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:45 crc kubenswrapper[4813]: I1201 08:40:45.887896 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:45 crc kubenswrapper[4813]: I1201 08:40:45.887915 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:45 crc kubenswrapper[4813]: I1201 08:40:45.887927 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:45Z","lastTransitionTime":"2025-12-01T08:40:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:45 crc kubenswrapper[4813]: I1201 08:40:45.990514 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:45 crc kubenswrapper[4813]: I1201 08:40:45.990557 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:45 crc kubenswrapper[4813]: I1201 08:40:45.990566 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:45 crc kubenswrapper[4813]: I1201 08:40:45.990581 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:45 crc kubenswrapper[4813]: I1201 08:40:45.990591 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:45Z","lastTransitionTime":"2025-12-01T08:40:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:46 crc kubenswrapper[4813]: I1201 08:40:46.065428 4813 generic.go:334] "Generic (PLEG): container finished" podID="d431b613-a91b-4fff-88b9-447804766289" containerID="19358f83a6c52957f959fe935a6843fdeef59905cc395e6e52a1fb4a670ac193" exitCode=0 Dec 01 08:40:46 crc kubenswrapper[4813]: I1201 08:40:46.065488 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-h7lvd" event={"ID":"d431b613-a91b-4fff-88b9-447804766289","Type":"ContainerDied","Data":"19358f83a6c52957f959fe935a6843fdeef59905cc395e6e52a1fb4a670ac193"} Dec 01 08:40:46 crc kubenswrapper[4813]: I1201 08:40:46.067853 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-xpnlc" event={"ID":"22da1fcf-0c57-4708-90d8-77692af17344","Type":"ContainerStarted","Data":"b4a36d213bdad2da91d0d33f400d5a8248befb68d5b94d19347873a93c6712a0"} Dec 01 08:40:46 crc kubenswrapper[4813]: I1201 08:40:46.076526 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 08:40:46 crc kubenswrapper[4813]: E1201 08:40:46.076628 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 08:40:54.076605291 +0000 UTC m=+38.699526877 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:40:46 crc kubenswrapper[4813]: I1201 08:40:46.076792 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 08:40:46 crc kubenswrapper[4813]: I1201 08:40:46.076843 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 08:40:46 crc kubenswrapper[4813]: E1201 08:40:46.077272 4813 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 01 08:40:46 crc kubenswrapper[4813]: E1201 08:40:46.077383 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-01 08:40:54.077354493 +0000 UTC m=+38.700276079 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 01 08:40:46 crc kubenswrapper[4813]: E1201 08:40:46.077419 4813 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 01 08:40:46 crc kubenswrapper[4813]: E1201 08:40:46.077488 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-01 08:40:54.077476216 +0000 UTC m=+38.700397962 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 01 08:40:46 crc kubenswrapper[4813]: I1201 08:40:46.087126 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"349332bd-6b4d-4306-84ae-07019ca3860f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d441a348087ae3cadda1ad8a7d1c4dff02cf3f6e1d346c294b278fdd9131edcc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://da67635daf6d9e32937f4d5d4ceeee3f8f09ae4fd2dd26c3e85be1286c5949de\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e97356ee274ea78d9f237ffbc756abb6ac8df0dbac244bf0a86ed215b8cbf43\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0404f5a5ed5a9750d4c479b608e68e8960b448f455d5af9d219aa5edaeb789b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a0404f5a5ed5a9750d4c479b608e68e8960b448f455d5af9d219aa5edaeb789b\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"message\\\":\\\"\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764578421\\\\\\\\\\\\\\\" (2025-12-01 08:40:21 +0000 UTC to 2025-12-31 08:40:22 +0000 UTC (now=2025-12-01 08:40:38.066068451 +0000 UTC))\\\\\\\"\\\\nI1201 08:40:38.066144 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1201 08:40:38.066166 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1201 08:40:38.066220 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1201 08:40:38.066232 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1201 08:40:38.066338 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764578432\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764578432\\\\\\\\\\\\\\\" (2025-12-01 07:40:32 +0000 UTC to 2026-12-01 07:40:32 +0000 UTC (now=2025-12-01 08:40:38.066301217 +0000 UTC))\\\\\\\"\\\\nI1201 08:40:38.066357 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1201 08:40:38.066385 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1201 08:40:38.066402 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1201 08:40:38.066414 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-63890811/tls.crt::/tmp/serving-cert-63890811/tls.key\\\\\\\"\\\\nI1201 08:40:38.066419 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1201 08:40:38.066572 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1201 08:40:38.067159 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7354f9337d42115bb1d974bf3772f2d03365acba57f2155135c7a0bbab39879e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:19Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05492df5703392cdf0932d6bf66025b18e53bd05fa0857a7cb2ad70979cbdc3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05492df5703392cdf0932d6bf66025b18e53bd05fa0857a7cb2ad70979cbdc3e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:16Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:46Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:46 crc kubenswrapper[4813]: I1201 08:40:46.092493 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:46 crc kubenswrapper[4813]: I1201 08:40:46.092529 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:46 crc kubenswrapper[4813]: I1201 08:40:46.092538 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:46 crc kubenswrapper[4813]: I1201 08:40:46.092554 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:46 crc kubenswrapper[4813]: I1201 08:40:46.092565 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:46Z","lastTransitionTime":"2025-12-01T08:40:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:46 crc kubenswrapper[4813]: I1201 08:40:46.104673 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3264b3af996af452df2953c59bea6deeb5ea6af1b119ce89f7334595c27e70da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9e879b1f563bd66383d58fd7b435a8a81bd50cb387dc1fabb6fc47d5a3fa468\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:46Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:46 crc kubenswrapper[4813]: I1201 08:40:46.123695 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-b6qqv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9656a03-0096-4c89-b0da-e37e4103611b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4195aebc1b2e388d480920a7f72257f6660afc99ad8c1a09478ef445739b272\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cjvnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:41Z\\\"}}\" for pod \"openshift-multus\"/\"multus-b6qqv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:46Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:46 crc kubenswrapper[4813]: I1201 08:40:46.136510 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-9wn4q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3d423402-e08f-42e9-b60c-6eb9bf2f6d48\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3049b842978c6580e1033a442a37766a2e76e278aface9a89fc0b2c57faf0df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rzxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c915733072b32a63237e2401cbf84b059aca396c2c0db884f0f439d599a35d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rzxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:41Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-9wn4q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:46Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:46 crc kubenswrapper[4813]: I1201 08:40:46.148826 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:46Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:46 crc kubenswrapper[4813]: I1201 08:40:46.161586 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-c576v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"17292cd0-1eac-40b5-a60e-67167c57409c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a42206c914d0493ca74788ea1a941de80920ab7b810fb95227cd9b8711e8f4b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7zwfx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:41Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-c576v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:46Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:46 crc kubenswrapper[4813]: I1201 08:40:46.177472 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 08:40:46 crc kubenswrapper[4813]: I1201 08:40:46.177525 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 08:40:46 crc kubenswrapper[4813]: E1201 08:40:46.177702 4813 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 01 08:40:46 crc kubenswrapper[4813]: E1201 08:40:46.177724 4813 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 01 08:40:46 crc kubenswrapper[4813]: E1201 08:40:46.177742 4813 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 08:40:46 crc kubenswrapper[4813]: E1201 08:40:46.177755 4813 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 01 08:40:46 crc kubenswrapper[4813]: E1201 08:40:46.177795 4813 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 01 08:40:46 crc kubenswrapper[4813]: E1201 08:40:46.177813 4813 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 08:40:46 crc kubenswrapper[4813]: E1201 08:40:46.177795 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-01 08:40:54.177777602 +0000 UTC m=+38.800699188 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 08:40:46 crc kubenswrapper[4813]: E1201 08:40:46.177908 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-01 08:40:54.177887106 +0000 UTC m=+38.800808752 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 08:40:46 crc kubenswrapper[4813]: I1201 08:40:46.185323 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-j4pcc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b7505e14-3560-4d0d-be2c-d6a7a5953865\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd67634e0c806323334495401a4e20b48fea4bc5013acd973d2abd3d70870f99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd67634e0c806323334495401a4e20b48fea4bc5013acd973d2abd3d70870f99\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:41Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-j4pcc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:46Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:46 crc kubenswrapper[4813]: I1201 08:40:46.195333 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:46 crc kubenswrapper[4813]: I1201 08:40:46.195371 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:46 crc kubenswrapper[4813]: I1201 08:40:46.195382 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:46 crc kubenswrapper[4813]: I1201 08:40:46.195399 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:46 crc kubenswrapper[4813]: I1201 08:40:46.195410 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:46Z","lastTransitionTime":"2025-12-01T08:40:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:46 crc kubenswrapper[4813]: I1201 08:40:46.201749 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-h7lvd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d431b613-a91b-4fff-88b9-447804766289\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://495593f99e92a9c28100a95e27d06ff5847008c5d996366cb474bedd9b69d5ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://495593f99e92a9c28100a95e27d06ff5847008c5d996366cb474bedd9b69d5ef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://19358f83a6c52957f959fe935a6843fdeef59905cc395e6e52a1fb4a670ac193\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://19358f83a6c52957f959fe935a6843fdeef59905cc395e6e52a1fb4a670ac193\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:41Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-h7lvd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:46Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:46 crc kubenswrapper[4813]: I1201 08:40:46.214456 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-xpnlc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"22da1fcf-0c57-4708-90d8-77692af17344\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xc44d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-xpnlc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:46Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:46 crc kubenswrapper[4813]: I1201 08:40:46.232803 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6fb5a648-ae4f-4e46-a530-007e9bf84abb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e24ebfc52dcb28368a2d436e7d9ab78f07f2aec2a4c71405a6c7d3301eee96a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33898118bbf639c8e6eae5aafa10386b95795512444ed5f535db3c663f169df2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9902c00bf7738686c93f62e1772a484d05715d74b75ee98c84c8f23ed383824b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3657b47ec9f1563c83bfbe3d728c8fd123f7078e219d251b1874269e42c31eea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:16Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:46Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:46 crc kubenswrapper[4813]: I1201 08:40:46.249933 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0023b0e358d75aacc889ce364846dda89ea513832c87fa4b066aea7dac57e8a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:46Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:46 crc kubenswrapper[4813]: I1201 08:40:46.264581 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:46Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:46 crc kubenswrapper[4813]: I1201 08:40:46.277053 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c7574f892d5e9142296b26a0b484eb6759ab53f571139fb6e23b93dc3da05780\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:46Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:46 crc kubenswrapper[4813]: I1201 08:40:46.288369 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:46Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:46 crc kubenswrapper[4813]: I1201 08:40:46.297916 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:46 crc kubenswrapper[4813]: I1201 08:40:46.297952 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:46 crc kubenswrapper[4813]: I1201 08:40:46.297987 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:46 crc kubenswrapper[4813]: I1201 08:40:46.298003 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:46 crc kubenswrapper[4813]: I1201 08:40:46.298013 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:46Z","lastTransitionTime":"2025-12-01T08:40:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:46 crc kubenswrapper[4813]: I1201 08:40:46.302793 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"349332bd-6b4d-4306-84ae-07019ca3860f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d441a348087ae3cadda1ad8a7d1c4dff02cf3f6e1d346c294b278fdd9131edcc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://da67635daf6d9e32937f4d5d4ceeee3f8f09ae4fd2dd26c3e85be1286c5949de\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e97356ee274ea78d9f237ffbc756abb6ac8df0dbac244bf0a86ed215b8cbf43\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0404f5a5ed5a9750d4c479b608e68e8960b448f455d5af9d219aa5edaeb789b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a0404f5a5ed5a9750d4c479b608e68e8960b448f455d5af9d219aa5edaeb789b\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"message\\\":\\\"\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764578421\\\\\\\\\\\\\\\" (2025-12-01 08:40:21 +0000 UTC to 2025-12-31 08:40:22 +0000 UTC (now=2025-12-01 08:40:38.066068451 +0000 UTC))\\\\\\\"\\\\nI1201 08:40:38.066144 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1201 08:40:38.066166 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1201 08:40:38.066220 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1201 08:40:38.066232 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1201 08:40:38.066338 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764578432\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764578432\\\\\\\\\\\\\\\" (2025-12-01 07:40:32 +0000 UTC to 2026-12-01 07:40:32 +0000 UTC (now=2025-12-01 08:40:38.066301217 +0000 UTC))\\\\\\\"\\\\nI1201 08:40:38.066357 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1201 08:40:38.066385 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1201 08:40:38.066402 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1201 08:40:38.066414 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-63890811/tls.crt::/tmp/serving-cert-63890811/tls.key\\\\\\\"\\\\nI1201 08:40:38.066419 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1201 08:40:38.066572 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1201 08:40:38.067159 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7354f9337d42115bb1d974bf3772f2d03365acba57f2155135c7a0bbab39879e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:19Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05492df5703392cdf0932d6bf66025b18e53bd05fa0857a7cb2ad70979cbdc3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05492df5703392cdf0932d6bf66025b18e53bd05fa0857a7cb2ad70979cbdc3e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:16Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:46Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:46 crc kubenswrapper[4813]: I1201 08:40:46.315147 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3264b3af996af452df2953c59bea6deeb5ea6af1b119ce89f7334595c27e70da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9e879b1f563bd66383d58fd7b435a8a81bd50cb387dc1fabb6fc47d5a3fa468\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:46Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:46 crc kubenswrapper[4813]: I1201 08:40:46.327990 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-b6qqv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9656a03-0096-4c89-b0da-e37e4103611b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4195aebc1b2e388d480920a7f72257f6660afc99ad8c1a09478ef445739b272\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cjvnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:41Z\\\"}}\" for pod \"openshift-multus\"/\"multus-b6qqv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:46Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:46 crc kubenswrapper[4813]: I1201 08:40:46.343889 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-9wn4q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3d423402-e08f-42e9-b60c-6eb9bf2f6d48\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3049b842978c6580e1033a442a37766a2e76e278aface9a89fc0b2c57faf0df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rzxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c915733072b32a63237e2401cbf84b059aca396c2c0db884f0f439d599a35d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rzxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:41Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-9wn4q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:46Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:46 crc kubenswrapper[4813]: I1201 08:40:46.360085 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-h7lvd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d431b613-a91b-4fff-88b9-447804766289\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://495593f99e92a9c28100a95e27d06ff5847008c5d996366cb474bedd9b69d5ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://495593f99e92a9c28100a95e27d06ff5847008c5d996366cb474bedd9b69d5ef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://19358f83a6c52957f959fe935a6843fdeef59905cc395e6e52a1fb4a670ac193\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://19358f83a6c52957f959fe935a6843fdeef59905cc395e6e52a1fb4a670ac193\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:41Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-h7lvd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:46Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:46 crc kubenswrapper[4813]: I1201 08:40:46.373687 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:46Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:46 crc kubenswrapper[4813]: I1201 08:40:46.385634 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-c576v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"17292cd0-1eac-40b5-a60e-67167c57409c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a42206c914d0493ca74788ea1a941de80920ab7b810fb95227cd9b8711e8f4b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7zwfx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:41Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-c576v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:46Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:46 crc kubenswrapper[4813]: I1201 08:40:46.393771 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 08:40:46 crc kubenswrapper[4813]: I1201 08:40:46.393871 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 08:40:46 crc kubenswrapper[4813]: E1201 08:40:46.393932 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 08:40:46 crc kubenswrapper[4813]: I1201 08:40:46.393869 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 08:40:46 crc kubenswrapper[4813]: E1201 08:40:46.393986 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 08:40:46 crc kubenswrapper[4813]: E1201 08:40:46.394181 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 08:40:46 crc kubenswrapper[4813]: I1201 08:40:46.402714 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:46 crc kubenswrapper[4813]: I1201 08:40:46.402757 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:46 crc kubenswrapper[4813]: I1201 08:40:46.402771 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:46 crc kubenswrapper[4813]: I1201 08:40:46.402791 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:46 crc kubenswrapper[4813]: I1201 08:40:46.402804 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:46Z","lastTransitionTime":"2025-12-01T08:40:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:46 crc kubenswrapper[4813]: I1201 08:40:46.405538 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-j4pcc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b7505e14-3560-4d0d-be2c-d6a7a5953865\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd67634e0c806323334495401a4e20b48fea4bc5013acd973d2abd3d70870f99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd67634e0c806323334495401a4e20b48fea4bc5013acd973d2abd3d70870f99\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:41Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-j4pcc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:46Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:46 crc kubenswrapper[4813]: I1201 08:40:46.422203 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:46Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:46 crc kubenswrapper[4813]: I1201 08:40:46.433296 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-xpnlc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"22da1fcf-0c57-4708-90d8-77692af17344\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4a36d213bdad2da91d0d33f400d5a8248befb68d5b94d19347873a93c6712a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xc44d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-xpnlc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:46Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:46 crc kubenswrapper[4813]: I1201 08:40:46.447314 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6fb5a648-ae4f-4e46-a530-007e9bf84abb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e24ebfc52dcb28368a2d436e7d9ab78f07f2aec2a4c71405a6c7d3301eee96a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33898118bbf639c8e6eae5aafa10386b95795512444ed5f535db3c663f169df2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9902c00bf7738686c93f62e1772a484d05715d74b75ee98c84c8f23ed383824b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3657b47ec9f1563c83bfbe3d728c8fd123f7078e219d251b1874269e42c31eea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:16Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:46Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:46 crc kubenswrapper[4813]: I1201 08:40:46.459417 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0023b0e358d75aacc889ce364846dda89ea513832c87fa4b066aea7dac57e8a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:46Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:46 crc kubenswrapper[4813]: I1201 08:40:46.470632 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:46Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:46 crc kubenswrapper[4813]: I1201 08:40:46.482825 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c7574f892d5e9142296b26a0b484eb6759ab53f571139fb6e23b93dc3da05780\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:46Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:46 crc kubenswrapper[4813]: I1201 08:40:46.495619 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:46Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:46 crc kubenswrapper[4813]: I1201 08:40:46.506527 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:46 crc kubenswrapper[4813]: I1201 08:40:46.506574 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:46 crc kubenswrapper[4813]: I1201 08:40:46.506586 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:46 crc kubenswrapper[4813]: I1201 08:40:46.506607 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:46 crc kubenswrapper[4813]: I1201 08:40:46.506621 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:46Z","lastTransitionTime":"2025-12-01T08:40:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:46 crc kubenswrapper[4813]: I1201 08:40:46.508488 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-c576v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"17292cd0-1eac-40b5-a60e-67167c57409c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a42206c914d0493ca74788ea1a941de80920ab7b810fb95227cd9b8711e8f4b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7zwfx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:41Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-c576v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:46Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:46 crc kubenswrapper[4813]: I1201 08:40:46.526604 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-j4pcc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b7505e14-3560-4d0d-be2c-d6a7a5953865\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd67634e0c806323334495401a4e20b48fea4bc5013acd973d2abd3d70870f99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd67634e0c806323334495401a4e20b48fea4bc5013acd973d2abd3d70870f99\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:41Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-j4pcc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:46Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:46 crc kubenswrapper[4813]: I1201 08:40:46.540220 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-h7lvd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d431b613-a91b-4fff-88b9-447804766289\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://495593f99e92a9c28100a95e27d06ff5847008c5d996366cb474bedd9b69d5ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://495593f99e92a9c28100a95e27d06ff5847008c5d996366cb474bedd9b69d5ef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://19358f83a6c52957f959fe935a6843fdeef59905cc395e6e52a1fb4a670ac193\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://19358f83a6c52957f959fe935a6843fdeef59905cc395e6e52a1fb4a670ac193\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:41Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-h7lvd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:46Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:46 crc kubenswrapper[4813]: I1201 08:40:46.551726 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6fb5a648-ae4f-4e46-a530-007e9bf84abb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e24ebfc52dcb28368a2d436e7d9ab78f07f2aec2a4c71405a6c7d3301eee96a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33898118bbf639c8e6eae5aafa10386b95795512444ed5f535db3c663f169df2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9902c00bf7738686c93f62e1772a484d05715d74b75ee98c84c8f23ed383824b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3657b47ec9f1563c83bfbe3d728c8fd123f7078e219d251b1874269e42c31eea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:16Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:46Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:46 crc kubenswrapper[4813]: I1201 08:40:46.563695 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0023b0e358d75aacc889ce364846dda89ea513832c87fa4b066aea7dac57e8a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:46Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:46 crc kubenswrapper[4813]: I1201 08:40:46.577264 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:46Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:46 crc kubenswrapper[4813]: I1201 08:40:46.589743 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c7574f892d5e9142296b26a0b484eb6759ab53f571139fb6e23b93dc3da05780\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:46Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:46 crc kubenswrapper[4813]: I1201 08:40:46.601590 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:46Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:46 crc kubenswrapper[4813]: I1201 08:40:46.608517 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:46 crc kubenswrapper[4813]: I1201 08:40:46.608581 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:46 crc kubenswrapper[4813]: I1201 08:40:46.608596 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:46 crc kubenswrapper[4813]: I1201 08:40:46.608617 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:46 crc kubenswrapper[4813]: I1201 08:40:46.608650 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:46Z","lastTransitionTime":"2025-12-01T08:40:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:46 crc kubenswrapper[4813]: I1201 08:40:46.613008 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-xpnlc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"22da1fcf-0c57-4708-90d8-77692af17344\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4a36d213bdad2da91d0d33f400d5a8248befb68d5b94d19347873a93c6712a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xc44d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-xpnlc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:46Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:46 crc kubenswrapper[4813]: I1201 08:40:46.628697 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"349332bd-6b4d-4306-84ae-07019ca3860f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d441a348087ae3cadda1ad8a7d1c4dff02cf3f6e1d346c294b278fdd9131edcc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://da67635daf6d9e32937f4d5d4ceeee3f8f09ae4fd2dd26c3e85be1286c5949de\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e97356ee274ea78d9f237ffbc756abb6ac8df0dbac244bf0a86ed215b8cbf43\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0404f5a5ed5a9750d4c479b608e68e8960b448f455d5af9d219aa5edaeb789b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a0404f5a5ed5a9750d4c479b608e68e8960b448f455d5af9d219aa5edaeb789b\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"message\\\":\\\"\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764578421\\\\\\\\\\\\\\\" (2025-12-01 08:40:21 +0000 UTC to 2025-12-31 08:40:22 +0000 UTC (now=2025-12-01 08:40:38.066068451 +0000 UTC))\\\\\\\"\\\\nI1201 08:40:38.066144 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1201 08:40:38.066166 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1201 08:40:38.066220 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1201 08:40:38.066232 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1201 08:40:38.066338 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764578432\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764578432\\\\\\\\\\\\\\\" (2025-12-01 07:40:32 +0000 UTC to 2026-12-01 07:40:32 +0000 UTC (now=2025-12-01 08:40:38.066301217 +0000 UTC))\\\\\\\"\\\\nI1201 08:40:38.066357 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1201 08:40:38.066385 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1201 08:40:38.066402 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1201 08:40:38.066414 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-63890811/tls.crt::/tmp/serving-cert-63890811/tls.key\\\\\\\"\\\\nI1201 08:40:38.066419 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1201 08:40:38.066572 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1201 08:40:38.067159 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7354f9337d42115bb1d974bf3772f2d03365acba57f2155135c7a0bbab39879e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:19Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05492df5703392cdf0932d6bf66025b18e53bd05fa0857a7cb2ad70979cbdc3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05492df5703392cdf0932d6bf66025b18e53bd05fa0857a7cb2ad70979cbdc3e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:16Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:46Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:46 crc kubenswrapper[4813]: I1201 08:40:46.643159 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3264b3af996af452df2953c59bea6deeb5ea6af1b119ce89f7334595c27e70da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9e879b1f563bd66383d58fd7b435a8a81bd50cb387dc1fabb6fc47d5a3fa468\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:46Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:46 crc kubenswrapper[4813]: I1201 08:40:46.655725 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-b6qqv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9656a03-0096-4c89-b0da-e37e4103611b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4195aebc1b2e388d480920a7f72257f6660afc99ad8c1a09478ef445739b272\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cjvnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:41Z\\\"}}\" for pod \"openshift-multus\"/\"multus-b6qqv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:46Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:46 crc kubenswrapper[4813]: I1201 08:40:46.666807 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-9wn4q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3d423402-e08f-42e9-b60c-6eb9bf2f6d48\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3049b842978c6580e1033a442a37766a2e76e278aface9a89fc0b2c57faf0df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rzxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c915733072b32a63237e2401cbf84b059aca396c2c0db884f0f439d599a35d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rzxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:41Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-9wn4q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:46Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:46 crc kubenswrapper[4813]: I1201 08:40:46.711237 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:46 crc kubenswrapper[4813]: I1201 08:40:46.711278 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:46 crc kubenswrapper[4813]: I1201 08:40:46.711287 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:46 crc kubenswrapper[4813]: I1201 08:40:46.711301 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:46 crc kubenswrapper[4813]: I1201 08:40:46.711311 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:46Z","lastTransitionTime":"2025-12-01T08:40:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:46 crc kubenswrapper[4813]: I1201 08:40:46.813781 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:46 crc kubenswrapper[4813]: I1201 08:40:46.813826 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:46 crc kubenswrapper[4813]: I1201 08:40:46.813835 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:46 crc kubenswrapper[4813]: I1201 08:40:46.813851 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:46 crc kubenswrapper[4813]: I1201 08:40:46.813863 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:46Z","lastTransitionTime":"2025-12-01T08:40:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:46 crc kubenswrapper[4813]: I1201 08:40:46.916278 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:46 crc kubenswrapper[4813]: I1201 08:40:46.916345 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:46 crc kubenswrapper[4813]: I1201 08:40:46.916358 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:46 crc kubenswrapper[4813]: I1201 08:40:46.916375 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:46 crc kubenswrapper[4813]: I1201 08:40:46.916386 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:46Z","lastTransitionTime":"2025-12-01T08:40:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:47 crc kubenswrapper[4813]: I1201 08:40:47.019817 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:47 crc kubenswrapper[4813]: I1201 08:40:47.019870 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:47 crc kubenswrapper[4813]: I1201 08:40:47.019883 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:47 crc kubenswrapper[4813]: I1201 08:40:47.019903 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:47 crc kubenswrapper[4813]: I1201 08:40:47.019917 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:47Z","lastTransitionTime":"2025-12-01T08:40:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:47 crc kubenswrapper[4813]: I1201 08:40:47.078650 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-j4pcc" event={"ID":"b7505e14-3560-4d0d-be2c-d6a7a5953865","Type":"ContainerStarted","Data":"b45454b440c5108fa150a6b856afa1778ba6e9f89633433cb3117b68f00f72c3"} Dec 01 08:40:47 crc kubenswrapper[4813]: I1201 08:40:47.081156 4813 generic.go:334] "Generic (PLEG): container finished" podID="d431b613-a91b-4fff-88b9-447804766289" containerID="9f089c115d916e339994879d224e3db27642b99aaabf59bcafb2979549052945" exitCode=0 Dec 01 08:40:47 crc kubenswrapper[4813]: I1201 08:40:47.081214 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-h7lvd" event={"ID":"d431b613-a91b-4fff-88b9-447804766289","Type":"ContainerDied","Data":"9f089c115d916e339994879d224e3db27642b99aaabf59bcafb2979549052945"} Dec 01 08:40:47 crc kubenswrapper[4813]: I1201 08:40:47.093774 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-9wn4q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3d423402-e08f-42e9-b60c-6eb9bf2f6d48\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3049b842978c6580e1033a442a37766a2e76e278aface9a89fc0b2c57faf0df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rzxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c915733072b32a63237e2401cbf84b059aca396c2c0db884f0f439d599a35d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rzxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:41Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-9wn4q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:47Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:47 crc kubenswrapper[4813]: I1201 08:40:47.106053 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-c576v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"17292cd0-1eac-40b5-a60e-67167c57409c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a42206c914d0493ca74788ea1a941de80920ab7b810fb95227cd9b8711e8f4b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7zwfx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:41Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-c576v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:47Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:47 crc kubenswrapper[4813]: I1201 08:40:47.124264 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:47 crc kubenswrapper[4813]: I1201 08:40:47.124304 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:47 crc kubenswrapper[4813]: I1201 08:40:47.124317 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:47 crc kubenswrapper[4813]: I1201 08:40:47.124336 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:47 crc kubenswrapper[4813]: I1201 08:40:47.124350 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:47Z","lastTransitionTime":"2025-12-01T08:40:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:47 crc kubenswrapper[4813]: I1201 08:40:47.128322 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-j4pcc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b7505e14-3560-4d0d-be2c-d6a7a5953865\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd67634e0c806323334495401a4e20b48fea4bc5013acd973d2abd3d70870f99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd67634e0c806323334495401a4e20b48fea4bc5013acd973d2abd3d70870f99\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:41Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-j4pcc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:47Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:47 crc kubenswrapper[4813]: I1201 08:40:47.149527 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-h7lvd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d431b613-a91b-4fff-88b9-447804766289\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://495593f99e92a9c28100a95e27d06ff5847008c5d996366cb474bedd9b69d5ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://495593f99e92a9c28100a95e27d06ff5847008c5d996366cb474bedd9b69d5ef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://19358f83a6c52957f959fe935a6843fdeef59905cc395e6e52a1fb4a670ac193\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://19358f83a6c52957f959fe935a6843fdeef59905cc395e6e52a1fb4a670ac193\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f089c115d916e339994879d224e3db27642b99aaabf59bcafb2979549052945\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9f089c115d916e339994879d224e3db27642b99aaabf59bcafb2979549052945\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:41Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-h7lvd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:47Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:47 crc kubenswrapper[4813]: I1201 08:40:47.162059 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:47Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:47 crc kubenswrapper[4813]: I1201 08:40:47.175148 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:47Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:47 crc kubenswrapper[4813]: I1201 08:40:47.186955 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c7574f892d5e9142296b26a0b484eb6759ab53f571139fb6e23b93dc3da05780\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:47Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:47 crc kubenswrapper[4813]: I1201 08:40:47.200974 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:47Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:47 crc kubenswrapper[4813]: I1201 08:40:47.210980 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-xpnlc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"22da1fcf-0c57-4708-90d8-77692af17344\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4a36d213bdad2da91d0d33f400d5a8248befb68d5b94d19347873a93c6712a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xc44d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-xpnlc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:47Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:47 crc kubenswrapper[4813]: I1201 08:40:47.226488 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6fb5a648-ae4f-4e46-a530-007e9bf84abb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e24ebfc52dcb28368a2d436e7d9ab78f07f2aec2a4c71405a6c7d3301eee96a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33898118bbf639c8e6eae5aafa10386b95795512444ed5f535db3c663f169df2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9902c00bf7738686c93f62e1772a484d05715d74b75ee98c84c8f23ed383824b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3657b47ec9f1563c83bfbe3d728c8fd123f7078e219d251b1874269e42c31eea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:16Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:47Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:47 crc kubenswrapper[4813]: I1201 08:40:47.227026 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:47 crc kubenswrapper[4813]: I1201 08:40:47.227067 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:47 crc kubenswrapper[4813]: I1201 08:40:47.227079 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:47 crc kubenswrapper[4813]: I1201 08:40:47.227096 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:47 crc kubenswrapper[4813]: I1201 08:40:47.227110 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:47Z","lastTransitionTime":"2025-12-01T08:40:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:47 crc kubenswrapper[4813]: I1201 08:40:47.240433 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0023b0e358d75aacc889ce364846dda89ea513832c87fa4b066aea7dac57e8a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:47Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:47 crc kubenswrapper[4813]: I1201 08:40:47.252323 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3264b3af996af452df2953c59bea6deeb5ea6af1b119ce89f7334595c27e70da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9e879b1f563bd66383d58fd7b435a8a81bd50cb387dc1fabb6fc47d5a3fa468\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:47Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:47 crc kubenswrapper[4813]: I1201 08:40:47.266362 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-b6qqv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9656a03-0096-4c89-b0da-e37e4103611b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4195aebc1b2e388d480920a7f72257f6660afc99ad8c1a09478ef445739b272\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cjvnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:41Z\\\"}}\" for pod \"openshift-multus\"/\"multus-b6qqv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:47Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:47 crc kubenswrapper[4813]: I1201 08:40:47.281368 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"349332bd-6b4d-4306-84ae-07019ca3860f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d441a348087ae3cadda1ad8a7d1c4dff02cf3f6e1d346c294b278fdd9131edcc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://da67635daf6d9e32937f4d5d4ceeee3f8f09ae4fd2dd26c3e85be1286c5949de\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e97356ee274ea78d9f237ffbc756abb6ac8df0dbac244bf0a86ed215b8cbf43\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0404f5a5ed5a9750d4c479b608e68e8960b448f455d5af9d219aa5edaeb789b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a0404f5a5ed5a9750d4c479b608e68e8960b448f455d5af9d219aa5edaeb789b\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"message\\\":\\\"\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764578421\\\\\\\\\\\\\\\" (2025-12-01 08:40:21 +0000 UTC to 2025-12-31 08:40:22 +0000 UTC (now=2025-12-01 08:40:38.066068451 +0000 UTC))\\\\\\\"\\\\nI1201 08:40:38.066144 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1201 08:40:38.066166 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1201 08:40:38.066220 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1201 08:40:38.066232 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1201 08:40:38.066338 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764578432\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764578432\\\\\\\\\\\\\\\" (2025-12-01 07:40:32 +0000 UTC to 2026-12-01 07:40:32 +0000 UTC (now=2025-12-01 08:40:38.066301217 +0000 UTC))\\\\\\\"\\\\nI1201 08:40:38.066357 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1201 08:40:38.066385 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1201 08:40:38.066402 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1201 08:40:38.066414 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-63890811/tls.crt::/tmp/serving-cert-63890811/tls.key\\\\\\\"\\\\nI1201 08:40:38.066419 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1201 08:40:38.066572 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1201 08:40:38.067159 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7354f9337d42115bb1d974bf3772f2d03365acba57f2155135c7a0bbab39879e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:19Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05492df5703392cdf0932d6bf66025b18e53bd05fa0857a7cb2ad70979cbdc3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05492df5703392cdf0932d6bf66025b18e53bd05fa0857a7cb2ad70979cbdc3e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:16Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:47Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:47 crc kubenswrapper[4813]: I1201 08:40:47.329877 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:47 crc kubenswrapper[4813]: I1201 08:40:47.329925 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:47 crc kubenswrapper[4813]: I1201 08:40:47.329936 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:47 crc kubenswrapper[4813]: I1201 08:40:47.329954 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:47 crc kubenswrapper[4813]: I1201 08:40:47.329986 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:47Z","lastTransitionTime":"2025-12-01T08:40:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:47 crc kubenswrapper[4813]: I1201 08:40:47.432979 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:47 crc kubenswrapper[4813]: I1201 08:40:47.433017 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:47 crc kubenswrapper[4813]: I1201 08:40:47.433027 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:47 crc kubenswrapper[4813]: I1201 08:40:47.433043 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:47 crc kubenswrapper[4813]: I1201 08:40:47.433054 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:47Z","lastTransitionTime":"2025-12-01T08:40:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:47 crc kubenswrapper[4813]: I1201 08:40:47.536030 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:47 crc kubenswrapper[4813]: I1201 08:40:47.536077 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:47 crc kubenswrapper[4813]: I1201 08:40:47.536088 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:47 crc kubenswrapper[4813]: I1201 08:40:47.536107 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:47 crc kubenswrapper[4813]: I1201 08:40:47.536118 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:47Z","lastTransitionTime":"2025-12-01T08:40:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:47 crc kubenswrapper[4813]: I1201 08:40:47.638717 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:47 crc kubenswrapper[4813]: I1201 08:40:47.638777 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:47 crc kubenswrapper[4813]: I1201 08:40:47.638796 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:47 crc kubenswrapper[4813]: I1201 08:40:47.638821 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:47 crc kubenswrapper[4813]: I1201 08:40:47.638838 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:47Z","lastTransitionTime":"2025-12-01T08:40:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:47 crc kubenswrapper[4813]: I1201 08:40:47.741681 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:47 crc kubenswrapper[4813]: I1201 08:40:47.741711 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:47 crc kubenswrapper[4813]: I1201 08:40:47.741719 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:47 crc kubenswrapper[4813]: I1201 08:40:47.741733 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:47 crc kubenswrapper[4813]: I1201 08:40:47.741743 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:47Z","lastTransitionTime":"2025-12-01T08:40:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:47 crc kubenswrapper[4813]: I1201 08:40:47.843929 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:47 crc kubenswrapper[4813]: I1201 08:40:47.843986 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:47 crc kubenswrapper[4813]: I1201 08:40:47.843998 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:47 crc kubenswrapper[4813]: I1201 08:40:47.844013 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:47 crc kubenswrapper[4813]: I1201 08:40:47.844023 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:47Z","lastTransitionTime":"2025-12-01T08:40:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:47 crc kubenswrapper[4813]: I1201 08:40:47.945791 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:47 crc kubenswrapper[4813]: I1201 08:40:47.945825 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:47 crc kubenswrapper[4813]: I1201 08:40:47.945834 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:47 crc kubenswrapper[4813]: I1201 08:40:47.945848 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:47 crc kubenswrapper[4813]: I1201 08:40:47.945858 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:47Z","lastTransitionTime":"2025-12-01T08:40:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:48 crc kubenswrapper[4813]: I1201 08:40:48.048078 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:48 crc kubenswrapper[4813]: I1201 08:40:48.048142 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:48 crc kubenswrapper[4813]: I1201 08:40:48.048161 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:48 crc kubenswrapper[4813]: I1201 08:40:48.048185 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:48 crc kubenswrapper[4813]: I1201 08:40:48.048200 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:48Z","lastTransitionTime":"2025-12-01T08:40:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:48 crc kubenswrapper[4813]: I1201 08:40:48.089956 4813 generic.go:334] "Generic (PLEG): container finished" podID="d431b613-a91b-4fff-88b9-447804766289" containerID="bb66f224462305c2005b31ef7bb9ed01efde8101e18518dd42d1dfb0df765bb9" exitCode=0 Dec 01 08:40:48 crc kubenswrapper[4813]: I1201 08:40:48.090052 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-h7lvd" event={"ID":"d431b613-a91b-4fff-88b9-447804766289","Type":"ContainerDied","Data":"bb66f224462305c2005b31ef7bb9ed01efde8101e18518dd42d1dfb0df765bb9"} Dec 01 08:40:48 crc kubenswrapper[4813]: I1201 08:40:48.106210 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:48Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:48 crc kubenswrapper[4813]: I1201 08:40:48.116797 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-c576v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"17292cd0-1eac-40b5-a60e-67167c57409c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a42206c914d0493ca74788ea1a941de80920ab7b810fb95227cd9b8711e8f4b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7zwfx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:41Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-c576v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:48Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:48 crc kubenswrapper[4813]: I1201 08:40:48.142536 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-j4pcc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b7505e14-3560-4d0d-be2c-d6a7a5953865\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd67634e0c806323334495401a4e20b48fea4bc5013acd973d2abd3d70870f99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd67634e0c806323334495401a4e20b48fea4bc5013acd973d2abd3d70870f99\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:41Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-j4pcc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:48Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:48 crc kubenswrapper[4813]: I1201 08:40:48.150423 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:48 crc kubenswrapper[4813]: I1201 08:40:48.150474 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:48 crc kubenswrapper[4813]: I1201 08:40:48.150483 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:48 crc kubenswrapper[4813]: I1201 08:40:48.150500 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:48 crc kubenswrapper[4813]: I1201 08:40:48.150511 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:48Z","lastTransitionTime":"2025-12-01T08:40:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:48 crc kubenswrapper[4813]: I1201 08:40:48.161824 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-h7lvd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d431b613-a91b-4fff-88b9-447804766289\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://495593f99e92a9c28100a95e27d06ff5847008c5d996366cb474bedd9b69d5ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://495593f99e92a9c28100a95e27d06ff5847008c5d996366cb474bedd9b69d5ef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://19358f83a6c52957f959fe935a6843fdeef59905cc395e6e52a1fb4a670ac193\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://19358f83a6c52957f959fe935a6843fdeef59905cc395e6e52a1fb4a670ac193\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f089c115d916e339994879d224e3db27642b99aaabf59bcafb2979549052945\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9f089c115d916e339994879d224e3db27642b99aaabf59bcafb2979549052945\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb66f224462305c2005b31ef7bb9ed01efde8101e18518dd42d1dfb0df765bb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb66f224462305c2005b31ef7bb9ed01efde8101e18518dd42d1dfb0df765bb9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:41Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-h7lvd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:48Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:48 crc kubenswrapper[4813]: I1201 08:40:48.175077 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6fb5a648-ae4f-4e46-a530-007e9bf84abb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e24ebfc52dcb28368a2d436e7d9ab78f07f2aec2a4c71405a6c7d3301eee96a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33898118bbf639c8e6eae5aafa10386b95795512444ed5f535db3c663f169df2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9902c00bf7738686c93f62e1772a484d05715d74b75ee98c84c8f23ed383824b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3657b47ec9f1563c83bfbe3d728c8fd123f7078e219d251b1874269e42c31eea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:16Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:48Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:48 crc kubenswrapper[4813]: I1201 08:40:48.186587 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0023b0e358d75aacc889ce364846dda89ea513832c87fa4b066aea7dac57e8a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:48Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:48 crc kubenswrapper[4813]: I1201 08:40:48.200262 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:48Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:48 crc kubenswrapper[4813]: I1201 08:40:48.212951 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c7574f892d5e9142296b26a0b484eb6759ab53f571139fb6e23b93dc3da05780\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:48Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:48 crc kubenswrapper[4813]: I1201 08:40:48.225352 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:48Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:48 crc kubenswrapper[4813]: I1201 08:40:48.235375 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-xpnlc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"22da1fcf-0c57-4708-90d8-77692af17344\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4a36d213bdad2da91d0d33f400d5a8248befb68d5b94d19347873a93c6712a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xc44d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-xpnlc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:48Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:48 crc kubenswrapper[4813]: I1201 08:40:48.247599 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"349332bd-6b4d-4306-84ae-07019ca3860f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d441a348087ae3cadda1ad8a7d1c4dff02cf3f6e1d346c294b278fdd9131edcc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://da67635daf6d9e32937f4d5d4ceeee3f8f09ae4fd2dd26c3e85be1286c5949de\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e97356ee274ea78d9f237ffbc756abb6ac8df0dbac244bf0a86ed215b8cbf43\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0404f5a5ed5a9750d4c479b608e68e8960b448f455d5af9d219aa5edaeb789b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a0404f5a5ed5a9750d4c479b608e68e8960b448f455d5af9d219aa5edaeb789b\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"message\\\":\\\"\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764578421\\\\\\\\\\\\\\\" (2025-12-01 08:40:21 +0000 UTC to 2025-12-31 08:40:22 +0000 UTC (now=2025-12-01 08:40:38.066068451 +0000 UTC))\\\\\\\"\\\\nI1201 08:40:38.066144 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1201 08:40:38.066166 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1201 08:40:38.066220 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1201 08:40:38.066232 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1201 08:40:38.066338 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764578432\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764578432\\\\\\\\\\\\\\\" (2025-12-01 07:40:32 +0000 UTC to 2026-12-01 07:40:32 +0000 UTC (now=2025-12-01 08:40:38.066301217 +0000 UTC))\\\\\\\"\\\\nI1201 08:40:38.066357 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1201 08:40:38.066385 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1201 08:40:38.066402 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1201 08:40:38.066414 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-63890811/tls.crt::/tmp/serving-cert-63890811/tls.key\\\\\\\"\\\\nI1201 08:40:38.066419 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1201 08:40:38.066572 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1201 08:40:38.067159 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7354f9337d42115bb1d974bf3772f2d03365acba57f2155135c7a0bbab39879e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:19Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05492df5703392cdf0932d6bf66025b18e53bd05fa0857a7cb2ad70979cbdc3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05492df5703392cdf0932d6bf66025b18e53bd05fa0857a7cb2ad70979cbdc3e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:16Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:48Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:48 crc kubenswrapper[4813]: I1201 08:40:48.252427 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:48 crc kubenswrapper[4813]: I1201 08:40:48.252465 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:48 crc kubenswrapper[4813]: I1201 08:40:48.252475 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:48 crc kubenswrapper[4813]: I1201 08:40:48.252490 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:48 crc kubenswrapper[4813]: I1201 08:40:48.252500 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:48Z","lastTransitionTime":"2025-12-01T08:40:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:48 crc kubenswrapper[4813]: I1201 08:40:48.259114 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3264b3af996af452df2953c59bea6deeb5ea6af1b119ce89f7334595c27e70da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9e879b1f563bd66383d58fd7b435a8a81bd50cb387dc1fabb6fc47d5a3fa468\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:48Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:48 crc kubenswrapper[4813]: I1201 08:40:48.272634 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-b6qqv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9656a03-0096-4c89-b0da-e37e4103611b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4195aebc1b2e388d480920a7f72257f6660afc99ad8c1a09478ef445739b272\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cjvnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:41Z\\\"}}\" for pod \"openshift-multus\"/\"multus-b6qqv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:48Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:48 crc kubenswrapper[4813]: I1201 08:40:48.283300 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-9wn4q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3d423402-e08f-42e9-b60c-6eb9bf2f6d48\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3049b842978c6580e1033a442a37766a2e76e278aface9a89fc0b2c57faf0df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rzxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c915733072b32a63237e2401cbf84b059aca396c2c0db884f0f439d599a35d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rzxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:41Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-9wn4q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:48Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:48 crc kubenswrapper[4813]: I1201 08:40:48.355307 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:48 crc kubenswrapper[4813]: I1201 08:40:48.355355 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:48 crc kubenswrapper[4813]: I1201 08:40:48.355367 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:48 crc kubenswrapper[4813]: I1201 08:40:48.355385 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:48 crc kubenswrapper[4813]: I1201 08:40:48.355395 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:48Z","lastTransitionTime":"2025-12-01T08:40:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:48 crc kubenswrapper[4813]: I1201 08:40:48.393168 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 08:40:48 crc kubenswrapper[4813]: I1201 08:40:48.393168 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 08:40:48 crc kubenswrapper[4813]: E1201 08:40:48.393337 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 08:40:48 crc kubenswrapper[4813]: E1201 08:40:48.393449 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 08:40:48 crc kubenswrapper[4813]: I1201 08:40:48.393194 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 08:40:48 crc kubenswrapper[4813]: E1201 08:40:48.393567 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 08:40:48 crc kubenswrapper[4813]: I1201 08:40:48.458516 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:48 crc kubenswrapper[4813]: I1201 08:40:48.458584 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:48 crc kubenswrapper[4813]: I1201 08:40:48.458606 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:48 crc kubenswrapper[4813]: I1201 08:40:48.458636 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:48 crc kubenswrapper[4813]: I1201 08:40:48.458660 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:48Z","lastTransitionTime":"2025-12-01T08:40:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:48 crc kubenswrapper[4813]: I1201 08:40:48.562448 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:48 crc kubenswrapper[4813]: I1201 08:40:48.562493 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:48 crc kubenswrapper[4813]: I1201 08:40:48.562504 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:48 crc kubenswrapper[4813]: I1201 08:40:48.562523 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:48 crc kubenswrapper[4813]: I1201 08:40:48.562536 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:48Z","lastTransitionTime":"2025-12-01T08:40:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:48 crc kubenswrapper[4813]: I1201 08:40:48.664797 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:48 crc kubenswrapper[4813]: I1201 08:40:48.664839 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:48 crc kubenswrapper[4813]: I1201 08:40:48.664855 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:48 crc kubenswrapper[4813]: I1201 08:40:48.664876 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:48 crc kubenswrapper[4813]: I1201 08:40:48.664890 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:48Z","lastTransitionTime":"2025-12-01T08:40:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:48 crc kubenswrapper[4813]: I1201 08:40:48.788530 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:48 crc kubenswrapper[4813]: I1201 08:40:48.788567 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:48 crc kubenswrapper[4813]: I1201 08:40:48.788576 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:48 crc kubenswrapper[4813]: I1201 08:40:48.788593 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:48 crc kubenswrapper[4813]: I1201 08:40:48.788604 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:48Z","lastTransitionTime":"2025-12-01T08:40:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:48 crc kubenswrapper[4813]: I1201 08:40:48.893731 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:48 crc kubenswrapper[4813]: I1201 08:40:48.893810 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:48 crc kubenswrapper[4813]: I1201 08:40:48.893830 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:48 crc kubenswrapper[4813]: I1201 08:40:48.893857 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:48 crc kubenswrapper[4813]: I1201 08:40:48.893871 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:48Z","lastTransitionTime":"2025-12-01T08:40:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:49 crc kubenswrapper[4813]: I1201 08:40:49.179865 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:49 crc kubenswrapper[4813]: I1201 08:40:49.180266 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:49 crc kubenswrapper[4813]: I1201 08:40:49.180281 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:49 crc kubenswrapper[4813]: I1201 08:40:49.180366 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:49 crc kubenswrapper[4813]: I1201 08:40:49.180397 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:49Z","lastTransitionTime":"2025-12-01T08:40:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:49 crc kubenswrapper[4813]: I1201 08:40:49.190988 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-j4pcc" event={"ID":"b7505e14-3560-4d0d-be2c-d6a7a5953865","Type":"ContainerStarted","Data":"4b11f49feda579bcb62f9680732033819ea11e7ae9937a06c0623a54c0ec106c"} Dec 01 08:40:49 crc kubenswrapper[4813]: I1201 08:40:49.193629 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-j4pcc" Dec 01 08:40:49 crc kubenswrapper[4813]: I1201 08:40:49.193841 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-j4pcc" Dec 01 08:40:49 crc kubenswrapper[4813]: I1201 08:40:49.205311 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-h7lvd" event={"ID":"d431b613-a91b-4fff-88b9-447804766289","Type":"ContainerStarted","Data":"1c268236c9a155c9b1f47c48d30dbd23948ceb2545592b40f356c60ea1d12e8a"} Dec 01 08:40:49 crc kubenswrapper[4813]: I1201 08:40:49.214775 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:49Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:49 crc kubenswrapper[4813]: I1201 08:40:49.267508 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-c576v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"17292cd0-1eac-40b5-a60e-67167c57409c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a42206c914d0493ca74788ea1a941de80920ab7b810fb95227cd9b8711e8f4b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7zwfx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:41Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-c576v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:49Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:49 crc kubenswrapper[4813]: I1201 08:40:49.273538 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-j4pcc" Dec 01 08:40:49 crc kubenswrapper[4813]: I1201 08:40:49.281264 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-j4pcc" Dec 01 08:40:49 crc kubenswrapper[4813]: I1201 08:40:49.286294 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:49 crc kubenswrapper[4813]: I1201 08:40:49.286327 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:49 crc kubenswrapper[4813]: I1201 08:40:49.286340 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:49 crc kubenswrapper[4813]: I1201 08:40:49.286352 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:49 crc kubenswrapper[4813]: I1201 08:40:49.286362 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:49Z","lastTransitionTime":"2025-12-01T08:40:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:49 crc kubenswrapper[4813]: I1201 08:40:49.288811 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-j4pcc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b7505e14-3560-4d0d-be2c-d6a7a5953865\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ccc30a63dc1c23ff9880fb04af1f31f0656b2696d359988e23deed23989f480a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://944977e9722025088720107530d0c91fc693593f380e4b49678f28a56da180cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc32d930ab23f30103d8fa42291909cbec89ebb75c5c53cd783cd27fe1a3f09e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f19c67b1c6c80c7f3223512102810d46ede1445464d21ae3bbea83424539b2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d984c90302e57f3e5a6f3ca3613b415db04ac74a5a228582f1b52ef1d59f93e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce8963bf1e9ee66926907aaf3a03d18977f4d0a1983d7121b62fbc8e02a37e4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b11f49feda579bcb62f9680732033819ea11e7ae9937a06c0623a54c0ec106c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b45454b440c5108fa150a6b856afa1778ba6e9f89633433cb3117b68f00f72c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd67634e0c806323334495401a4e20b48fea4bc5013acd973d2abd3d70870f99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd67634e0c806323334495401a4e20b48fea4bc5013acd973d2abd3d70870f99\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:41Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-j4pcc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:49Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:49 crc kubenswrapper[4813]: I1201 08:40:49.303508 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-h7lvd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d431b613-a91b-4fff-88b9-447804766289\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://495593f99e92a9c28100a95e27d06ff5847008c5d996366cb474bedd9b69d5ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://495593f99e92a9c28100a95e27d06ff5847008c5d996366cb474bedd9b69d5ef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://19358f83a6c52957f959fe935a6843fdeef59905cc395e6e52a1fb4a670ac193\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://19358f83a6c52957f959fe935a6843fdeef59905cc395e6e52a1fb4a670ac193\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f089c115d916e339994879d224e3db27642b99aaabf59bcafb2979549052945\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9f089c115d916e339994879d224e3db27642b99aaabf59bcafb2979549052945\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb66f224462305c2005b31ef7bb9ed01efde8101e18518dd42d1dfb0df765bb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb66f224462305c2005b31ef7bb9ed01efde8101e18518dd42d1dfb0df765bb9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:41Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-h7lvd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:49Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:49 crc kubenswrapper[4813]: I1201 08:40:49.315142 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6fb5a648-ae4f-4e46-a530-007e9bf84abb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e24ebfc52dcb28368a2d436e7d9ab78f07f2aec2a4c71405a6c7d3301eee96a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33898118bbf639c8e6eae5aafa10386b95795512444ed5f535db3c663f169df2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9902c00bf7738686c93f62e1772a484d05715d74b75ee98c84c8f23ed383824b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3657b47ec9f1563c83bfbe3d728c8fd123f7078e219d251b1874269e42c31eea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:16Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:49Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:49 crc kubenswrapper[4813]: I1201 08:40:49.328494 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0023b0e358d75aacc889ce364846dda89ea513832c87fa4b066aea7dac57e8a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:49Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:49 crc kubenswrapper[4813]: I1201 08:40:49.342441 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:49Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:49 crc kubenswrapper[4813]: I1201 08:40:49.346449 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:49 crc kubenswrapper[4813]: I1201 08:40:49.346512 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:49 crc kubenswrapper[4813]: I1201 08:40:49.346525 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:49 crc kubenswrapper[4813]: I1201 08:40:49.346538 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:49 crc kubenswrapper[4813]: I1201 08:40:49.346547 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:49Z","lastTransitionTime":"2025-12-01T08:40:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:49 crc kubenswrapper[4813]: I1201 08:40:49.358603 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c7574f892d5e9142296b26a0b484eb6759ab53f571139fb6e23b93dc3da05780\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:49Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:49 crc kubenswrapper[4813]: E1201 08:40:49.359943 4813 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:40:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:40:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:49Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:40:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:40:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:49Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cd271ba1-a5c5-44e0-b03e-cfd1f5e8cb76\\\",\\\"systemUUID\\\":\\\"906e3783-eb6d-4f21-889c-7989aeaa25ba\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:49Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:49 crc kubenswrapper[4813]: I1201 08:40:49.401881 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 08:40:49 crc kubenswrapper[4813]: I1201 08:40:49.401895 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:49Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:49 crc kubenswrapper[4813]: E1201 08:40:49.402062 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 08:40:49 crc kubenswrapper[4813]: I1201 08:40:49.404984 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:49 crc kubenswrapper[4813]: I1201 08:40:49.405018 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:49 crc kubenswrapper[4813]: I1201 08:40:49.405028 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:49 crc kubenswrapper[4813]: I1201 08:40:49.405041 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:49 crc kubenswrapper[4813]: I1201 08:40:49.405051 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:49Z","lastTransitionTime":"2025-12-01T08:40:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:49 crc kubenswrapper[4813]: E1201 08:40:49.416363 4813 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:40:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:40:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:49Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:40:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:40:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:49Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cd271ba1-a5c5-44e0-b03e-cfd1f5e8cb76\\\",\\\"systemUUID\\\":\\\"906e3783-eb6d-4f21-889c-7989aeaa25ba\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:49Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:49 crc kubenswrapper[4813]: I1201 08:40:49.417757 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-xpnlc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"22da1fcf-0c57-4708-90d8-77692af17344\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4a36d213bdad2da91d0d33f400d5a8248befb68d5b94d19347873a93c6712a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xc44d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-xpnlc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:49Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:49 crc kubenswrapper[4813]: I1201 08:40:49.421081 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:49 crc kubenswrapper[4813]: I1201 08:40:49.421113 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:49 crc kubenswrapper[4813]: I1201 08:40:49.421122 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:49 crc kubenswrapper[4813]: I1201 08:40:49.421162 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:49 crc kubenswrapper[4813]: I1201 08:40:49.421177 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:49Z","lastTransitionTime":"2025-12-01T08:40:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:49 crc kubenswrapper[4813]: I1201 08:40:49.429685 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"349332bd-6b4d-4306-84ae-07019ca3860f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d441a348087ae3cadda1ad8a7d1c4dff02cf3f6e1d346c294b278fdd9131edcc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://da67635daf6d9e32937f4d5d4ceeee3f8f09ae4fd2dd26c3e85be1286c5949de\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e97356ee274ea78d9f237ffbc756abb6ac8df0dbac244bf0a86ed215b8cbf43\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0404f5a5ed5a9750d4c479b608e68e8960b448f455d5af9d219aa5edaeb789b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a0404f5a5ed5a9750d4c479b608e68e8960b448f455d5af9d219aa5edaeb789b\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"message\\\":\\\"\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764578421\\\\\\\\\\\\\\\" (2025-12-01 08:40:21 +0000 UTC to 2025-12-31 08:40:22 +0000 UTC (now=2025-12-01 08:40:38.066068451 +0000 UTC))\\\\\\\"\\\\nI1201 08:40:38.066144 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1201 08:40:38.066166 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1201 08:40:38.066220 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1201 08:40:38.066232 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1201 08:40:38.066338 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764578432\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764578432\\\\\\\\\\\\\\\" (2025-12-01 07:40:32 +0000 UTC to 2026-12-01 07:40:32 +0000 UTC (now=2025-12-01 08:40:38.066301217 +0000 UTC))\\\\\\\"\\\\nI1201 08:40:38.066357 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1201 08:40:38.066385 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1201 08:40:38.066402 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1201 08:40:38.066414 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-63890811/tls.crt::/tmp/serving-cert-63890811/tls.key\\\\\\\"\\\\nI1201 08:40:38.066419 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1201 08:40:38.066572 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1201 08:40:38.067159 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7354f9337d42115bb1d974bf3772f2d03365acba57f2155135c7a0bbab39879e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:19Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05492df5703392cdf0932d6bf66025b18e53bd05fa0857a7cb2ad70979cbdc3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05492df5703392cdf0932d6bf66025b18e53bd05fa0857a7cb2ad70979cbdc3e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:16Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:49Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:49 crc kubenswrapper[4813]: E1201 08:40:49.573751 4813 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:40:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:40:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:49Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:40:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:40:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:49Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cd271ba1-a5c5-44e0-b03e-cfd1f5e8cb76\\\",\\\"systemUUID\\\":\\\"906e3783-eb6d-4f21-889c-7989aeaa25ba\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:49Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:49 crc kubenswrapper[4813]: I1201 08:40:49.590290 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:49 crc kubenswrapper[4813]: I1201 08:40:49.590336 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:49 crc kubenswrapper[4813]: I1201 08:40:49.590355 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:49 crc kubenswrapper[4813]: I1201 08:40:49.590382 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:49 crc kubenswrapper[4813]: I1201 08:40:49.590394 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:49Z","lastTransitionTime":"2025-12-01T08:40:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:49 crc kubenswrapper[4813]: I1201 08:40:49.596266 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3264b3af996af452df2953c59bea6deeb5ea6af1b119ce89f7334595c27e70da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9e879b1f563bd66383d58fd7b435a8a81bd50cb387dc1fabb6fc47d5a3fa468\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:49Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:49 crc kubenswrapper[4813]: I1201 08:40:49.616677 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-b6qqv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9656a03-0096-4c89-b0da-e37e4103611b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4195aebc1b2e388d480920a7f72257f6660afc99ad8c1a09478ef445739b272\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cjvnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:41Z\\\"}}\" for pod \"openshift-multus\"/\"multus-b6qqv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:49Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:49 crc kubenswrapper[4813]: E1201 08:40:49.616824 4813 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:40:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:40:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:49Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:40:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:40:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:49Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cd271ba1-a5c5-44e0-b03e-cfd1f5e8cb76\\\",\\\"systemUUID\\\":\\\"906e3783-eb6d-4f21-889c-7989aeaa25ba\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:49Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:49 crc kubenswrapper[4813]: I1201 08:40:49.627221 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:49 crc kubenswrapper[4813]: I1201 08:40:49.627874 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:49 crc kubenswrapper[4813]: I1201 08:40:49.627904 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:49 crc kubenswrapper[4813]: I1201 08:40:49.627929 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:49 crc kubenswrapper[4813]: I1201 08:40:49.627942 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:49Z","lastTransitionTime":"2025-12-01T08:40:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:49 crc kubenswrapper[4813]: I1201 08:40:49.666874 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-9wn4q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3d423402-e08f-42e9-b60c-6eb9bf2f6d48\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3049b842978c6580e1033a442a37766a2e76e278aface9a89fc0b2c57faf0df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rzxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c915733072b32a63237e2401cbf84b059aca396c2c0db884f0f439d599a35d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rzxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:41Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-9wn4q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:49Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:49 crc kubenswrapper[4813]: E1201 08:40:49.723258 4813 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:40:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:40:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:49Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:40:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:40:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:49Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cd271ba1-a5c5-44e0-b03e-cfd1f5e8cb76\\\",\\\"systemUUID\\\":\\\"906e3783-eb6d-4f21-889c-7989aeaa25ba\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:49Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:49 crc kubenswrapper[4813]: E1201 08:40:49.723451 4813 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 01 08:40:49 crc kubenswrapper[4813]: I1201 08:40:49.729071 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:49 crc kubenswrapper[4813]: I1201 08:40:49.729103 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:49 crc kubenswrapper[4813]: I1201 08:40:49.729113 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:49 crc kubenswrapper[4813]: I1201 08:40:49.729143 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:49 crc kubenswrapper[4813]: I1201 08:40:49.729151 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:49Z","lastTransitionTime":"2025-12-01T08:40:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:49 crc kubenswrapper[4813]: I1201 08:40:49.860536 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:49 crc kubenswrapper[4813]: I1201 08:40:49.860583 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:49 crc kubenswrapper[4813]: I1201 08:40:49.860601 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:49 crc kubenswrapper[4813]: I1201 08:40:49.860641 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:49 crc kubenswrapper[4813]: I1201 08:40:49.860655 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:49Z","lastTransitionTime":"2025-12-01T08:40:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:49 crc kubenswrapper[4813]: I1201 08:40:49.911806 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-9wn4q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3d423402-e08f-42e9-b60c-6eb9bf2f6d48\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3049b842978c6580e1033a442a37766a2e76e278aface9a89fc0b2c57faf0df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rzxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c915733072b32a63237e2401cbf84b059aca396c2c0db884f0f439d599a35d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rzxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:41Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-9wn4q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:49Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:49 crc kubenswrapper[4813]: I1201 08:40:49.926756 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:49Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:49 crc kubenswrapper[4813]: I1201 08:40:49.939279 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-c576v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"17292cd0-1eac-40b5-a60e-67167c57409c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a42206c914d0493ca74788ea1a941de80920ab7b810fb95227cd9b8711e8f4b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7zwfx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:41Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-c576v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:49Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:49 crc kubenswrapper[4813]: I1201 08:40:49.961411 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-j4pcc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b7505e14-3560-4d0d-be2c-d6a7a5953865\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ccc30a63dc1c23ff9880fb04af1f31f0656b2696d359988e23deed23989f480a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://944977e9722025088720107530d0c91fc693593f380e4b49678f28a56da180cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc32d930ab23f30103d8fa42291909cbec89ebb75c5c53cd783cd27fe1a3f09e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f19c67b1c6c80c7f3223512102810d46ede1445464d21ae3bbea83424539b2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d984c90302e57f3e5a6f3ca3613b415db04ac74a5a228582f1b52ef1d59f93e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce8963bf1e9ee66926907aaf3a03d18977f4d0a1983d7121b62fbc8e02a37e4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b11f49feda579bcb62f9680732033819ea11e7ae9937a06c0623a54c0ec106c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b45454b440c5108fa150a6b856afa1778ba6e9f89633433cb3117b68f00f72c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd67634e0c806323334495401a4e20b48fea4bc5013acd973d2abd3d70870f99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd67634e0c806323334495401a4e20b48fea4bc5013acd973d2abd3d70870f99\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:41Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-j4pcc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:49Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:49 crc kubenswrapper[4813]: I1201 08:40:49.974715 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-h7lvd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d431b613-a91b-4fff-88b9-447804766289\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://495593f99e92a9c28100a95e27d06ff5847008c5d996366cb474bedd9b69d5ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://495593f99e92a9c28100a95e27d06ff5847008c5d996366cb474bedd9b69d5ef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://19358f83a6c52957f959fe935a6843fdeef59905cc395e6e52a1fb4a670ac193\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://19358f83a6c52957f959fe935a6843fdeef59905cc395e6e52a1fb4a670ac193\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f089c115d916e339994879d224e3db27642b99aaabf59bcafb2979549052945\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9f089c115d916e339994879d224e3db27642b99aaabf59bcafb2979549052945\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb66f224462305c2005b31ef7bb9ed01efde8101e18518dd42d1dfb0df765bb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb66f224462305c2005b31ef7bb9ed01efde8101e18518dd42d1dfb0df765bb9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c268236c9a155c9b1f47c48d30dbd23948ceb2545592b40f356c60ea1d12e8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:41Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-h7lvd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:49Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:49 crc kubenswrapper[4813]: I1201 08:40:49.983923 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-xpnlc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"22da1fcf-0c57-4708-90d8-77692af17344\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4a36d213bdad2da91d0d33f400d5a8248befb68d5b94d19347873a93c6712a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xc44d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-xpnlc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:49Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:49 crc kubenswrapper[4813]: I1201 08:40:49.996000 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6fb5a648-ae4f-4e46-a530-007e9bf84abb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e24ebfc52dcb28368a2d436e7d9ab78f07f2aec2a4c71405a6c7d3301eee96a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33898118bbf639c8e6eae5aafa10386b95795512444ed5f535db3c663f169df2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9902c00bf7738686c93f62e1772a484d05715d74b75ee98c84c8f23ed383824b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3657b47ec9f1563c83bfbe3d728c8fd123f7078e219d251b1874269e42c31eea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:16Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:49Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:50 crc kubenswrapper[4813]: I1201 08:40:50.007753 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0023b0e358d75aacc889ce364846dda89ea513832c87fa4b066aea7dac57e8a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:50Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:50 crc kubenswrapper[4813]: I1201 08:40:50.011026 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:50 crc kubenswrapper[4813]: I1201 08:40:50.011050 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:50 crc kubenswrapper[4813]: I1201 08:40:50.011058 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:50 crc kubenswrapper[4813]: I1201 08:40:50.011072 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:50 crc kubenswrapper[4813]: I1201 08:40:50.011081 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:50Z","lastTransitionTime":"2025-12-01T08:40:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:50 crc kubenswrapper[4813]: I1201 08:40:50.021657 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:50Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:50 crc kubenswrapper[4813]: I1201 08:40:50.037050 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c7574f892d5e9142296b26a0b484eb6759ab53f571139fb6e23b93dc3da05780\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:50Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:50 crc kubenswrapper[4813]: I1201 08:40:50.049194 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:50Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:50 crc kubenswrapper[4813]: I1201 08:40:50.062995 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"349332bd-6b4d-4306-84ae-07019ca3860f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d441a348087ae3cadda1ad8a7d1c4dff02cf3f6e1d346c294b278fdd9131edcc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://da67635daf6d9e32937f4d5d4ceeee3f8f09ae4fd2dd26c3e85be1286c5949de\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e97356ee274ea78d9f237ffbc756abb6ac8df0dbac244bf0a86ed215b8cbf43\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0404f5a5ed5a9750d4c479b608e68e8960b448f455d5af9d219aa5edaeb789b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a0404f5a5ed5a9750d4c479b608e68e8960b448f455d5af9d219aa5edaeb789b\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"message\\\":\\\"\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764578421\\\\\\\\\\\\\\\" (2025-12-01 08:40:21 +0000 UTC to 2025-12-31 08:40:22 +0000 UTC (now=2025-12-01 08:40:38.066068451 +0000 UTC))\\\\\\\"\\\\nI1201 08:40:38.066144 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1201 08:40:38.066166 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1201 08:40:38.066220 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1201 08:40:38.066232 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1201 08:40:38.066338 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764578432\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764578432\\\\\\\\\\\\\\\" (2025-12-01 07:40:32 +0000 UTC to 2026-12-01 07:40:32 +0000 UTC (now=2025-12-01 08:40:38.066301217 +0000 UTC))\\\\\\\"\\\\nI1201 08:40:38.066357 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1201 08:40:38.066385 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1201 08:40:38.066402 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1201 08:40:38.066414 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-63890811/tls.crt::/tmp/serving-cert-63890811/tls.key\\\\\\\"\\\\nI1201 08:40:38.066419 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1201 08:40:38.066572 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1201 08:40:38.067159 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7354f9337d42115bb1d974bf3772f2d03365acba57f2155135c7a0bbab39879e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:19Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05492df5703392cdf0932d6bf66025b18e53bd05fa0857a7cb2ad70979cbdc3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05492df5703392cdf0932d6bf66025b18e53bd05fa0857a7cb2ad70979cbdc3e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:16Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:50Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:50 crc kubenswrapper[4813]: I1201 08:40:50.076446 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3264b3af996af452df2953c59bea6deeb5ea6af1b119ce89f7334595c27e70da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9e879b1f563bd66383d58fd7b435a8a81bd50cb387dc1fabb6fc47d5a3fa468\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:50Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:50 crc kubenswrapper[4813]: I1201 08:40:50.087178 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-b6qqv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9656a03-0096-4c89-b0da-e37e4103611b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4195aebc1b2e388d480920a7f72257f6660afc99ad8c1a09478ef445739b272\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cjvnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:41Z\\\"}}\" for pod \"openshift-multus\"/\"multus-b6qqv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:50Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:50 crc kubenswrapper[4813]: I1201 08:40:50.113376 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:50 crc kubenswrapper[4813]: I1201 08:40:50.113457 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:50 crc kubenswrapper[4813]: I1201 08:40:50.113483 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:50 crc kubenswrapper[4813]: I1201 08:40:50.113517 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:50 crc kubenswrapper[4813]: I1201 08:40:50.113541 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:50Z","lastTransitionTime":"2025-12-01T08:40:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:50 crc kubenswrapper[4813]: I1201 08:40:50.300854 4813 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 01 08:40:50 crc kubenswrapper[4813]: I1201 08:40:50.301489 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:50 crc kubenswrapper[4813]: I1201 08:40:50.301521 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:50 crc kubenswrapper[4813]: I1201 08:40:50.301529 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:50 crc kubenswrapper[4813]: I1201 08:40:50.301555 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:50 crc kubenswrapper[4813]: I1201 08:40:50.301564 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:50Z","lastTransitionTime":"2025-12-01T08:40:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:50 crc kubenswrapper[4813]: I1201 08:40:50.393397 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 08:40:50 crc kubenswrapper[4813]: I1201 08:40:50.393499 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 08:40:50 crc kubenswrapper[4813]: E1201 08:40:50.393563 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 08:40:50 crc kubenswrapper[4813]: E1201 08:40:50.393729 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 08:40:50 crc kubenswrapper[4813]: I1201 08:40:50.404142 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:50 crc kubenswrapper[4813]: I1201 08:40:50.404181 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:50 crc kubenswrapper[4813]: I1201 08:40:50.404193 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:50 crc kubenswrapper[4813]: I1201 08:40:50.404229 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:50 crc kubenswrapper[4813]: I1201 08:40:50.404243 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:50Z","lastTransitionTime":"2025-12-01T08:40:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:50 crc kubenswrapper[4813]: I1201 08:40:50.507175 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:50 crc kubenswrapper[4813]: I1201 08:40:50.507257 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:50 crc kubenswrapper[4813]: I1201 08:40:50.507274 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:50 crc kubenswrapper[4813]: I1201 08:40:50.507307 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:50 crc kubenswrapper[4813]: I1201 08:40:50.507319 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:50Z","lastTransitionTime":"2025-12-01T08:40:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:50 crc kubenswrapper[4813]: I1201 08:40:50.609947 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:50 crc kubenswrapper[4813]: I1201 08:40:50.610015 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:50 crc kubenswrapper[4813]: I1201 08:40:50.610028 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:50 crc kubenswrapper[4813]: I1201 08:40:50.610061 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:50 crc kubenswrapper[4813]: I1201 08:40:50.610073 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:50Z","lastTransitionTime":"2025-12-01T08:40:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:50 crc kubenswrapper[4813]: I1201 08:40:50.712397 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:50 crc kubenswrapper[4813]: I1201 08:40:50.712465 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:50 crc kubenswrapper[4813]: I1201 08:40:50.712478 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:50 crc kubenswrapper[4813]: I1201 08:40:50.712498 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:50 crc kubenswrapper[4813]: I1201 08:40:50.712511 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:50Z","lastTransitionTime":"2025-12-01T08:40:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:50 crc kubenswrapper[4813]: I1201 08:40:50.815174 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:50 crc kubenswrapper[4813]: I1201 08:40:50.815243 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:50 crc kubenswrapper[4813]: I1201 08:40:50.815268 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:50 crc kubenswrapper[4813]: I1201 08:40:50.815299 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:50 crc kubenswrapper[4813]: I1201 08:40:50.815323 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:50Z","lastTransitionTime":"2025-12-01T08:40:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:50 crc kubenswrapper[4813]: I1201 08:40:50.918214 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:50 crc kubenswrapper[4813]: I1201 08:40:50.918254 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:50 crc kubenswrapper[4813]: I1201 08:40:50.918265 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:50 crc kubenswrapper[4813]: I1201 08:40:50.918282 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:50 crc kubenswrapper[4813]: I1201 08:40:50.918293 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:50Z","lastTransitionTime":"2025-12-01T08:40:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:51 crc kubenswrapper[4813]: I1201 08:40:51.021759 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:51 crc kubenswrapper[4813]: I1201 08:40:51.021808 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:51 crc kubenswrapper[4813]: I1201 08:40:51.021819 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:51 crc kubenswrapper[4813]: I1201 08:40:51.021837 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:51 crc kubenswrapper[4813]: I1201 08:40:51.021849 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:51Z","lastTransitionTime":"2025-12-01T08:40:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:51 crc kubenswrapper[4813]: I1201 08:40:51.124816 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:51 crc kubenswrapper[4813]: I1201 08:40:51.124864 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:51 crc kubenswrapper[4813]: I1201 08:40:51.124874 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:51 crc kubenswrapper[4813]: I1201 08:40:51.124888 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:51 crc kubenswrapper[4813]: I1201 08:40:51.124901 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:51Z","lastTransitionTime":"2025-12-01T08:40:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:51 crc kubenswrapper[4813]: I1201 08:40:51.227094 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:51 crc kubenswrapper[4813]: I1201 08:40:51.227143 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:51 crc kubenswrapper[4813]: I1201 08:40:51.227155 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:51 crc kubenswrapper[4813]: I1201 08:40:51.227175 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:51 crc kubenswrapper[4813]: I1201 08:40:51.227186 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:51Z","lastTransitionTime":"2025-12-01T08:40:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:51 crc kubenswrapper[4813]: I1201 08:40:51.309184 4813 generic.go:334] "Generic (PLEG): container finished" podID="d431b613-a91b-4fff-88b9-447804766289" containerID="1c268236c9a155c9b1f47c48d30dbd23948ceb2545592b40f356c60ea1d12e8a" exitCode=0 Dec 01 08:40:51 crc kubenswrapper[4813]: I1201 08:40:51.309257 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-h7lvd" event={"ID":"d431b613-a91b-4fff-88b9-447804766289","Type":"ContainerDied","Data":"1c268236c9a155c9b1f47c48d30dbd23948ceb2545592b40f356c60ea1d12e8a"} Dec 01 08:40:51 crc kubenswrapper[4813]: I1201 08:40:51.309392 4813 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 01 08:40:51 crc kubenswrapper[4813]: I1201 08:40:51.329299 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:51 crc kubenswrapper[4813]: I1201 08:40:51.329339 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:51 crc kubenswrapper[4813]: I1201 08:40:51.329349 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:51 crc kubenswrapper[4813]: I1201 08:40:51.329366 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:51 crc kubenswrapper[4813]: I1201 08:40:51.329377 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:51Z","lastTransitionTime":"2025-12-01T08:40:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:51 crc kubenswrapper[4813]: I1201 08:40:51.330020 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"349332bd-6b4d-4306-84ae-07019ca3860f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d441a348087ae3cadda1ad8a7d1c4dff02cf3f6e1d346c294b278fdd9131edcc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://da67635daf6d9e32937f4d5d4ceeee3f8f09ae4fd2dd26c3e85be1286c5949de\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e97356ee274ea78d9f237ffbc756abb6ac8df0dbac244bf0a86ed215b8cbf43\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0404f5a5ed5a9750d4c479b608e68e8960b448f455d5af9d219aa5edaeb789b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a0404f5a5ed5a9750d4c479b608e68e8960b448f455d5af9d219aa5edaeb789b\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"message\\\":\\\"\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764578421\\\\\\\\\\\\\\\" (2025-12-01 08:40:21 +0000 UTC to 2025-12-31 08:40:22 +0000 UTC (now=2025-12-01 08:40:38.066068451 +0000 UTC))\\\\\\\"\\\\nI1201 08:40:38.066144 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1201 08:40:38.066166 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1201 08:40:38.066220 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1201 08:40:38.066232 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1201 08:40:38.066338 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764578432\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764578432\\\\\\\\\\\\\\\" (2025-12-01 07:40:32 +0000 UTC to 2026-12-01 07:40:32 +0000 UTC (now=2025-12-01 08:40:38.066301217 +0000 UTC))\\\\\\\"\\\\nI1201 08:40:38.066357 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1201 08:40:38.066385 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1201 08:40:38.066402 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1201 08:40:38.066414 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-63890811/tls.crt::/tmp/serving-cert-63890811/tls.key\\\\\\\"\\\\nI1201 08:40:38.066419 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1201 08:40:38.066572 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1201 08:40:38.067159 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7354f9337d42115bb1d974bf3772f2d03365acba57f2155135c7a0bbab39879e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:19Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05492df5703392cdf0932d6bf66025b18e53bd05fa0857a7cb2ad70979cbdc3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05492df5703392cdf0932d6bf66025b18e53bd05fa0857a7cb2ad70979cbdc3e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:16Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:51Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:51 crc kubenswrapper[4813]: I1201 08:40:51.344227 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3264b3af996af452df2953c59bea6deeb5ea6af1b119ce89f7334595c27e70da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9e879b1f563bd66383d58fd7b435a8a81bd50cb387dc1fabb6fc47d5a3fa468\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:51Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:51 crc kubenswrapper[4813]: I1201 08:40:51.359654 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-b6qqv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9656a03-0096-4c89-b0da-e37e4103611b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4195aebc1b2e388d480920a7f72257f6660afc99ad8c1a09478ef445739b272\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cjvnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:41Z\\\"}}\" for pod \"openshift-multus\"/\"multus-b6qqv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:51Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:51 crc kubenswrapper[4813]: I1201 08:40:51.375955 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-9wn4q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3d423402-e08f-42e9-b60c-6eb9bf2f6d48\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3049b842978c6580e1033a442a37766a2e76e278aface9a89fc0b2c57faf0df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rzxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c915733072b32a63237e2401cbf84b059aca396c2c0db884f0f439d599a35d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rzxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:41Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-9wn4q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:51Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:51 crc kubenswrapper[4813]: I1201 08:40:51.390116 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:51Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:51 crc kubenswrapper[4813]: I1201 08:40:51.392589 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 08:40:51 crc kubenswrapper[4813]: E1201 08:40:51.392755 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 08:40:51 crc kubenswrapper[4813]: I1201 08:40:51.393581 4813 scope.go:117] "RemoveContainer" containerID="a0404f5a5ed5a9750d4c479b608e68e8960b448f455d5af9d219aa5edaeb789b" Dec 01 08:40:51 crc kubenswrapper[4813]: I1201 08:40:51.402031 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-c576v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"17292cd0-1eac-40b5-a60e-67167c57409c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a42206c914d0493ca74788ea1a941de80920ab7b810fb95227cd9b8711e8f4b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7zwfx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:41Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-c576v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:51Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:51 crc kubenswrapper[4813]: I1201 08:40:51.431343 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-j4pcc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b7505e14-3560-4d0d-be2c-d6a7a5953865\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ccc30a63dc1c23ff9880fb04af1f31f0656b2696d359988e23deed23989f480a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://944977e9722025088720107530d0c91fc693593f380e4b49678f28a56da180cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc32d930ab23f30103d8fa42291909cbec89ebb75c5c53cd783cd27fe1a3f09e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f19c67b1c6c80c7f3223512102810d46ede1445464d21ae3bbea83424539b2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d984c90302e57f3e5a6f3ca3613b415db04ac74a5a228582f1b52ef1d59f93e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce8963bf1e9ee66926907aaf3a03d18977f4d0a1983d7121b62fbc8e02a37e4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b11f49feda579bcb62f9680732033819ea11e7ae9937a06c0623a54c0ec106c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b45454b440c5108fa150a6b856afa1778ba6e9f89633433cb3117b68f00f72c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd67634e0c806323334495401a4e20b48fea4bc5013acd973d2abd3d70870f99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd67634e0c806323334495401a4e20b48fea4bc5013acd973d2abd3d70870f99\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:41Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-j4pcc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:51Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:51 crc kubenswrapper[4813]: I1201 08:40:51.434194 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:51 crc kubenswrapper[4813]: I1201 08:40:51.434228 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:51 crc kubenswrapper[4813]: I1201 08:40:51.434239 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:51 crc kubenswrapper[4813]: I1201 08:40:51.434260 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:51 crc kubenswrapper[4813]: I1201 08:40:51.434272 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:51Z","lastTransitionTime":"2025-12-01T08:40:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:51 crc kubenswrapper[4813]: I1201 08:40:51.452308 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-h7lvd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d431b613-a91b-4fff-88b9-447804766289\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://495593f99e92a9c28100a95e27d06ff5847008c5d996366cb474bedd9b69d5ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://495593f99e92a9c28100a95e27d06ff5847008c5d996366cb474bedd9b69d5ef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://19358f83a6c52957f959fe935a6843fdeef59905cc395e6e52a1fb4a670ac193\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://19358f83a6c52957f959fe935a6843fdeef59905cc395e6e52a1fb4a670ac193\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f089c115d916e339994879d224e3db27642b99aaabf59bcafb2979549052945\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9f089c115d916e339994879d224e3db27642b99aaabf59bcafb2979549052945\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb66f224462305c2005b31ef7bb9ed01efde8101e18518dd42d1dfb0df765bb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb66f224462305c2005b31ef7bb9ed01efde8101e18518dd42d1dfb0df765bb9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c268236c9a155c9b1f47c48d30dbd23948ceb2545592b40f356c60ea1d12e8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c268236c9a155c9b1f47c48d30dbd23948ceb2545592b40f356c60ea1d12e8a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:41Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-h7lvd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:51Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:51 crc kubenswrapper[4813]: I1201 08:40:51.466486 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6fb5a648-ae4f-4e46-a530-007e9bf84abb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e24ebfc52dcb28368a2d436e7d9ab78f07f2aec2a4c71405a6c7d3301eee96a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33898118bbf639c8e6eae5aafa10386b95795512444ed5f535db3c663f169df2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9902c00bf7738686c93f62e1772a484d05715d74b75ee98c84c8f23ed383824b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3657b47ec9f1563c83bfbe3d728c8fd123f7078e219d251b1874269e42c31eea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:16Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:51Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:51 crc kubenswrapper[4813]: I1201 08:40:51.481824 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0023b0e358d75aacc889ce364846dda89ea513832c87fa4b066aea7dac57e8a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:51Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:51 crc kubenswrapper[4813]: I1201 08:40:51.585690 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:51Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:51 crc kubenswrapper[4813]: I1201 08:40:51.589476 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:51 crc kubenswrapper[4813]: I1201 08:40:51.589508 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:51 crc kubenswrapper[4813]: I1201 08:40:51.589518 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:51 crc kubenswrapper[4813]: I1201 08:40:51.589553 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:51 crc kubenswrapper[4813]: I1201 08:40:51.589563 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:51Z","lastTransitionTime":"2025-12-01T08:40:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:51 crc kubenswrapper[4813]: I1201 08:40:51.599767 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c7574f892d5e9142296b26a0b484eb6759ab53f571139fb6e23b93dc3da05780\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:51Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:51 crc kubenswrapper[4813]: I1201 08:40:51.616639 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:51Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:51 crc kubenswrapper[4813]: I1201 08:40:51.629310 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-xpnlc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"22da1fcf-0c57-4708-90d8-77692af17344\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4a36d213bdad2da91d0d33f400d5a8248befb68d5b94d19347873a93c6712a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xc44d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-xpnlc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:51Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:51 crc kubenswrapper[4813]: I1201 08:40:51.725704 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:51 crc kubenswrapper[4813]: I1201 08:40:51.725764 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:51 crc kubenswrapper[4813]: I1201 08:40:51.725779 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:51 crc kubenswrapper[4813]: I1201 08:40:51.725811 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:51 crc kubenswrapper[4813]: I1201 08:40:51.725828 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:51Z","lastTransitionTime":"2025-12-01T08:40:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:51 crc kubenswrapper[4813]: I1201 08:40:51.828903 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:51 crc kubenswrapper[4813]: I1201 08:40:51.828948 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:51 crc kubenswrapper[4813]: I1201 08:40:51.828986 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:51 crc kubenswrapper[4813]: I1201 08:40:51.829009 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:51 crc kubenswrapper[4813]: I1201 08:40:51.829026 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:51Z","lastTransitionTime":"2025-12-01T08:40:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:51 crc kubenswrapper[4813]: I1201 08:40:51.932551 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:51 crc kubenswrapper[4813]: I1201 08:40:51.932598 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:51 crc kubenswrapper[4813]: I1201 08:40:51.932615 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:51 crc kubenswrapper[4813]: I1201 08:40:51.932637 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:51 crc kubenswrapper[4813]: I1201 08:40:51.932654 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:51Z","lastTransitionTime":"2025-12-01T08:40:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:52 crc kubenswrapper[4813]: I1201 08:40:52.034803 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:52 crc kubenswrapper[4813]: I1201 08:40:52.034849 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:52 crc kubenswrapper[4813]: I1201 08:40:52.034862 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:52 crc kubenswrapper[4813]: I1201 08:40:52.034879 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:52 crc kubenswrapper[4813]: I1201 08:40:52.034890 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:52Z","lastTransitionTime":"2025-12-01T08:40:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:52 crc kubenswrapper[4813]: I1201 08:40:52.137012 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:52 crc kubenswrapper[4813]: I1201 08:40:52.137075 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:52 crc kubenswrapper[4813]: I1201 08:40:52.137091 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:52 crc kubenswrapper[4813]: I1201 08:40:52.137123 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:52 crc kubenswrapper[4813]: I1201 08:40:52.137142 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:52Z","lastTransitionTime":"2025-12-01T08:40:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:52 crc kubenswrapper[4813]: I1201 08:40:52.239843 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:52 crc kubenswrapper[4813]: I1201 08:40:52.239914 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:52 crc kubenswrapper[4813]: I1201 08:40:52.239938 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:52 crc kubenswrapper[4813]: I1201 08:40:52.239999 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:52 crc kubenswrapper[4813]: I1201 08:40:52.240024 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:52Z","lastTransitionTime":"2025-12-01T08:40:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:52 crc kubenswrapper[4813]: I1201 08:40:52.318450 4813 generic.go:334] "Generic (PLEG): container finished" podID="d431b613-a91b-4fff-88b9-447804766289" containerID="458aedca5f186e9aaed7e4ee29fd5247221491888b130f55c1fe48d88d565f8f" exitCode=0 Dec 01 08:40:52 crc kubenswrapper[4813]: I1201 08:40:52.318547 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-h7lvd" event={"ID":"d431b613-a91b-4fff-88b9-447804766289","Type":"ContainerDied","Data":"458aedca5f186e9aaed7e4ee29fd5247221491888b130f55c1fe48d88d565f8f"} Dec 01 08:40:52 crc kubenswrapper[4813]: I1201 08:40:52.321867 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Dec 01 08:40:52 crc kubenswrapper[4813]: I1201 08:40:52.328124 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"8dc6feb9817f4f356b8e36af43ec3070774f312fbbb4909fa05b8b3e8dd53582"} Dec 01 08:40:52 crc kubenswrapper[4813]: I1201 08:40:52.328780 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 08:40:52 crc kubenswrapper[4813]: I1201 08:40:52.342327 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:52 crc kubenswrapper[4813]: I1201 08:40:52.342390 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:52 crc kubenswrapper[4813]: I1201 08:40:52.342410 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:52 crc kubenswrapper[4813]: I1201 08:40:52.342434 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:52 crc kubenswrapper[4813]: I1201 08:40:52.342450 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:52Z","lastTransitionTime":"2025-12-01T08:40:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:52 crc kubenswrapper[4813]: I1201 08:40:52.350237 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-j4pcc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b7505e14-3560-4d0d-be2c-d6a7a5953865\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ccc30a63dc1c23ff9880fb04af1f31f0656b2696d359988e23deed23989f480a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://944977e9722025088720107530d0c91fc693593f380e4b49678f28a56da180cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc32d930ab23f30103d8fa42291909cbec89ebb75c5c53cd783cd27fe1a3f09e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f19c67b1c6c80c7f3223512102810d46ede1445464d21ae3bbea83424539b2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d984c90302e57f3e5a6f3ca3613b415db04ac74a5a228582f1b52ef1d59f93e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce8963bf1e9ee66926907aaf3a03d18977f4d0a1983d7121b62fbc8e02a37e4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b11f49feda579bcb62f9680732033819ea11e7ae9937a06c0623a54c0ec106c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b45454b440c5108fa150a6b856afa1778ba6e9f89633433cb3117b68f00f72c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd67634e0c806323334495401a4e20b48fea4bc5013acd973d2abd3d70870f99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd67634e0c806323334495401a4e20b48fea4bc5013acd973d2abd3d70870f99\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:41Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-j4pcc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:52Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:52 crc kubenswrapper[4813]: I1201 08:40:52.378157 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-h7lvd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d431b613-a91b-4fff-88b9-447804766289\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://495593f99e92a9c28100a95e27d06ff5847008c5d996366cb474bedd9b69d5ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://495593f99e92a9c28100a95e27d06ff5847008c5d996366cb474bedd9b69d5ef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://19358f83a6c52957f959fe935a6843fdeef59905cc395e6e52a1fb4a670ac193\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://19358f83a6c52957f959fe935a6843fdeef59905cc395e6e52a1fb4a670ac193\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f089c115d916e339994879d224e3db27642b99aaabf59bcafb2979549052945\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9f089c115d916e339994879d224e3db27642b99aaabf59bcafb2979549052945\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb66f224462305c2005b31ef7bb9ed01efde8101e18518dd42d1dfb0df765bb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb66f224462305c2005b31ef7bb9ed01efde8101e18518dd42d1dfb0df765bb9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c268236c9a155c9b1f47c48d30dbd23948ceb2545592b40f356c60ea1d12e8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c268236c9a155c9b1f47c48d30dbd23948ceb2545592b40f356c60ea1d12e8a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://458aedca5f186e9aaed7e4ee29fd5247221491888b130f55c1fe48d88d565f8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://458aedca5f186e9aaed7e4ee29fd5247221491888b130f55c1fe48d88d565f8f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:41Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-h7lvd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:52Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:52 crc kubenswrapper[4813]: I1201 08:40:52.392659 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 08:40:52 crc kubenswrapper[4813]: I1201 08:40:52.392707 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 08:40:52 crc kubenswrapper[4813]: E1201 08:40:52.392810 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 08:40:52 crc kubenswrapper[4813]: E1201 08:40:52.392899 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 08:40:52 crc kubenswrapper[4813]: I1201 08:40:52.397519 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:52Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:52 crc kubenswrapper[4813]: I1201 08:40:52.412877 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-c576v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"17292cd0-1eac-40b5-a60e-67167c57409c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a42206c914d0493ca74788ea1a941de80920ab7b810fb95227cd9b8711e8f4b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7zwfx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:41Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-c576v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:52Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:52 crc kubenswrapper[4813]: I1201 08:40:52.430662 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c7574f892d5e9142296b26a0b484eb6759ab53f571139fb6e23b93dc3da05780\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:52Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:52 crc kubenswrapper[4813]: I1201 08:40:52.444566 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:52Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:52 crc kubenswrapper[4813]: I1201 08:40:52.445575 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:52 crc kubenswrapper[4813]: I1201 08:40:52.445626 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:52 crc kubenswrapper[4813]: I1201 08:40:52.445641 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:52 crc kubenswrapper[4813]: I1201 08:40:52.445668 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:52 crc kubenswrapper[4813]: I1201 08:40:52.445683 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:52Z","lastTransitionTime":"2025-12-01T08:40:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:52 crc kubenswrapper[4813]: I1201 08:40:52.473688 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-xpnlc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"22da1fcf-0c57-4708-90d8-77692af17344\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4a36d213bdad2da91d0d33f400d5a8248befb68d5b94d19347873a93c6712a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xc44d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-xpnlc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:52Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:52 crc kubenswrapper[4813]: I1201 08:40:52.491683 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6fb5a648-ae4f-4e46-a530-007e9bf84abb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e24ebfc52dcb28368a2d436e7d9ab78f07f2aec2a4c71405a6c7d3301eee96a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33898118bbf639c8e6eae5aafa10386b95795512444ed5f535db3c663f169df2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9902c00bf7738686c93f62e1772a484d05715d74b75ee98c84c8f23ed383824b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3657b47ec9f1563c83bfbe3d728c8fd123f7078e219d251b1874269e42c31eea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:16Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:52Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:52 crc kubenswrapper[4813]: I1201 08:40:52.509577 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0023b0e358d75aacc889ce364846dda89ea513832c87fa4b066aea7dac57e8a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:52Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:52 crc kubenswrapper[4813]: I1201 08:40:52.521669 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:52Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:52 crc kubenswrapper[4813]: I1201 08:40:52.536367 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-b6qqv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9656a03-0096-4c89-b0da-e37e4103611b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4195aebc1b2e388d480920a7f72257f6660afc99ad8c1a09478ef445739b272\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cjvnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:41Z\\\"}}\" for pod \"openshift-multus\"/\"multus-b6qqv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:52Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:52 crc kubenswrapper[4813]: I1201 08:40:52.548565 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"349332bd-6b4d-4306-84ae-07019ca3860f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d441a348087ae3cadda1ad8a7d1c4dff02cf3f6e1d346c294b278fdd9131edcc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://da67635daf6d9e32937f4d5d4ceeee3f8f09ae4fd2dd26c3e85be1286c5949de\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e97356ee274ea78d9f237ffbc756abb6ac8df0dbac244bf0a86ed215b8cbf43\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0404f5a5ed5a9750d4c479b608e68e8960b448f455d5af9d219aa5edaeb789b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a0404f5a5ed5a9750d4c479b608e68e8960b448f455d5af9d219aa5edaeb789b\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"message\\\":\\\"\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764578421\\\\\\\\\\\\\\\" (2025-12-01 08:40:21 +0000 UTC to 2025-12-31 08:40:22 +0000 UTC (now=2025-12-01 08:40:38.066068451 +0000 UTC))\\\\\\\"\\\\nI1201 08:40:38.066144 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1201 08:40:38.066166 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1201 08:40:38.066220 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1201 08:40:38.066232 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1201 08:40:38.066338 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764578432\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764578432\\\\\\\\\\\\\\\" (2025-12-01 07:40:32 +0000 UTC to 2026-12-01 07:40:32 +0000 UTC (now=2025-12-01 08:40:38.066301217 +0000 UTC))\\\\\\\"\\\\nI1201 08:40:38.066357 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1201 08:40:38.066385 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1201 08:40:38.066402 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1201 08:40:38.066414 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-63890811/tls.crt::/tmp/serving-cert-63890811/tls.key\\\\\\\"\\\\nI1201 08:40:38.066419 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1201 08:40:38.066572 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1201 08:40:38.067159 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7354f9337d42115bb1d974bf3772f2d03365acba57f2155135c7a0bbab39879e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:19Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05492df5703392cdf0932d6bf66025b18e53bd05fa0857a7cb2ad70979cbdc3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05492df5703392cdf0932d6bf66025b18e53bd05fa0857a7cb2ad70979cbdc3e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:16Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:52Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:52 crc kubenswrapper[4813]: I1201 08:40:52.549452 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:52 crc kubenswrapper[4813]: I1201 08:40:52.549478 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:52 crc kubenswrapper[4813]: I1201 08:40:52.549487 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:52 crc kubenswrapper[4813]: I1201 08:40:52.549503 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:52 crc kubenswrapper[4813]: I1201 08:40:52.549514 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:52Z","lastTransitionTime":"2025-12-01T08:40:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:52 crc kubenswrapper[4813]: I1201 08:40:52.561005 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3264b3af996af452df2953c59bea6deeb5ea6af1b119ce89f7334595c27e70da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9e879b1f563bd66383d58fd7b435a8a81bd50cb387dc1fabb6fc47d5a3fa468\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:52Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:52 crc kubenswrapper[4813]: I1201 08:40:52.573360 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-9wn4q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3d423402-e08f-42e9-b60c-6eb9bf2f6d48\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3049b842978c6580e1033a442a37766a2e76e278aface9a89fc0b2c57faf0df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rzxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c915733072b32a63237e2401cbf84b059aca396c2c0db884f0f439d599a35d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rzxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:41Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-9wn4q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:52Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:52 crc kubenswrapper[4813]: I1201 08:40:52.591719 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"349332bd-6b4d-4306-84ae-07019ca3860f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d441a348087ae3cadda1ad8a7d1c4dff02cf3f6e1d346c294b278fdd9131edcc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://da67635daf6d9e32937f4d5d4ceeee3f8f09ae4fd2dd26c3e85be1286c5949de\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e97356ee274ea78d9f237ffbc756abb6ac8df0dbac244bf0a86ed215b8cbf43\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8dc6feb9817f4f356b8e36af43ec3070774f312fbbb4909fa05b8b3e8dd53582\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a0404f5a5ed5a9750d4c479b608e68e8960b448f455d5af9d219aa5edaeb789b\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"message\\\":\\\"\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764578421\\\\\\\\\\\\\\\" (2025-12-01 08:40:21 +0000 UTC to 2025-12-31 08:40:22 +0000 UTC (now=2025-12-01 08:40:38.066068451 +0000 UTC))\\\\\\\"\\\\nI1201 08:40:38.066144 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1201 08:40:38.066166 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1201 08:40:38.066220 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1201 08:40:38.066232 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1201 08:40:38.066338 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764578432\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764578432\\\\\\\\\\\\\\\" (2025-12-01 07:40:32 +0000 UTC to 2026-12-01 07:40:32 +0000 UTC (now=2025-12-01 08:40:38.066301217 +0000 UTC))\\\\\\\"\\\\nI1201 08:40:38.066357 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1201 08:40:38.066385 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1201 08:40:38.066402 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1201 08:40:38.066414 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-63890811/tls.crt::/tmp/serving-cert-63890811/tls.key\\\\\\\"\\\\nI1201 08:40:38.066419 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1201 08:40:38.066572 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1201 08:40:38.067159 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7354f9337d42115bb1d974bf3772f2d03365acba57f2155135c7a0bbab39879e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:19Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05492df5703392cdf0932d6bf66025b18e53bd05fa0857a7cb2ad70979cbdc3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05492df5703392cdf0932d6bf66025b18e53bd05fa0857a7cb2ad70979cbdc3e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:16Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:52Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:52 crc kubenswrapper[4813]: I1201 08:40:52.605745 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3264b3af996af452df2953c59bea6deeb5ea6af1b119ce89f7334595c27e70da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9e879b1f563bd66383d58fd7b435a8a81bd50cb387dc1fabb6fc47d5a3fa468\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:52Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:52 crc kubenswrapper[4813]: I1201 08:40:52.618904 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-b6qqv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9656a03-0096-4c89-b0da-e37e4103611b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4195aebc1b2e388d480920a7f72257f6660afc99ad8c1a09478ef445739b272\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cjvnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:41Z\\\"}}\" for pod \"openshift-multus\"/\"multus-b6qqv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:52Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:52 crc kubenswrapper[4813]: I1201 08:40:52.629411 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-9wn4q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3d423402-e08f-42e9-b60c-6eb9bf2f6d48\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3049b842978c6580e1033a442a37766a2e76e278aface9a89fc0b2c57faf0df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rzxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c915733072b32a63237e2401cbf84b059aca396c2c0db884f0f439d599a35d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rzxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:41Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-9wn4q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:52Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:52 crc kubenswrapper[4813]: I1201 08:40:52.640892 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:52Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:52 crc kubenswrapper[4813]: I1201 08:40:52.652375 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-c576v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"17292cd0-1eac-40b5-a60e-67167c57409c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a42206c914d0493ca74788ea1a941de80920ab7b810fb95227cd9b8711e8f4b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7zwfx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:41Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-c576v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:52Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:52 crc kubenswrapper[4813]: I1201 08:40:52.652530 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:52 crc kubenswrapper[4813]: I1201 08:40:52.652560 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:52 crc kubenswrapper[4813]: I1201 08:40:52.652637 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:52 crc kubenswrapper[4813]: I1201 08:40:52.652671 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:52 crc kubenswrapper[4813]: I1201 08:40:52.652739 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:52Z","lastTransitionTime":"2025-12-01T08:40:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:52 crc kubenswrapper[4813]: I1201 08:40:52.705419 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-j4pcc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b7505e14-3560-4d0d-be2c-d6a7a5953865\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ccc30a63dc1c23ff9880fb04af1f31f0656b2696d359988e23deed23989f480a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://944977e9722025088720107530d0c91fc693593f380e4b49678f28a56da180cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc32d930ab23f30103d8fa42291909cbec89ebb75c5c53cd783cd27fe1a3f09e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f19c67b1c6c80c7f3223512102810d46ede1445464d21ae3bbea83424539b2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d984c90302e57f3e5a6f3ca3613b415db04ac74a5a228582f1b52ef1d59f93e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce8963bf1e9ee66926907aaf3a03d18977f4d0a1983d7121b62fbc8e02a37e4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b11f49feda579bcb62f9680732033819ea11e7ae9937a06c0623a54c0ec106c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b45454b440c5108fa150a6b856afa1778ba6e9f89633433cb3117b68f00f72c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd67634e0c806323334495401a4e20b48fea4bc5013acd973d2abd3d70870f99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd67634e0c806323334495401a4e20b48fea4bc5013acd973d2abd3d70870f99\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:41Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-j4pcc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:52Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:52 crc kubenswrapper[4813]: I1201 08:40:52.725941 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-h7lvd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d431b613-a91b-4fff-88b9-447804766289\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://495593f99e92a9c28100a95e27d06ff5847008c5d996366cb474bedd9b69d5ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://495593f99e92a9c28100a95e27d06ff5847008c5d996366cb474bedd9b69d5ef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://19358f83a6c52957f959fe935a6843fdeef59905cc395e6e52a1fb4a670ac193\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://19358f83a6c52957f959fe935a6843fdeef59905cc395e6e52a1fb4a670ac193\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f089c115d916e339994879d224e3db27642b99aaabf59bcafb2979549052945\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9f089c115d916e339994879d224e3db27642b99aaabf59bcafb2979549052945\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb66f224462305c2005b31ef7bb9ed01efde8101e18518dd42d1dfb0df765bb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb66f224462305c2005b31ef7bb9ed01efde8101e18518dd42d1dfb0df765bb9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c268236c9a155c9b1f47c48d30dbd23948ceb2545592b40f356c60ea1d12e8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c268236c9a155c9b1f47c48d30dbd23948ceb2545592b40f356c60ea1d12e8a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://458aedca5f186e9aaed7e4ee29fd5247221491888b130f55c1fe48d88d565f8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://458aedca5f186e9aaed7e4ee29fd5247221491888b130f55c1fe48d88d565f8f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:41Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-h7lvd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:52Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:52 crc kubenswrapper[4813]: I1201 08:40:52.738212 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0023b0e358d75aacc889ce364846dda89ea513832c87fa4b066aea7dac57e8a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:52Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:52 crc kubenswrapper[4813]: I1201 08:40:52.748287 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:52Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:52 crc kubenswrapper[4813]: I1201 08:40:52.757325 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c7574f892d5e9142296b26a0b484eb6759ab53f571139fb6e23b93dc3da05780\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:52Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:52 crc kubenswrapper[4813]: I1201 08:40:52.782096 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:52Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:52 crc kubenswrapper[4813]: I1201 08:40:52.782657 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:52 crc kubenswrapper[4813]: I1201 08:40:52.782703 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:52 crc kubenswrapper[4813]: I1201 08:40:52.782720 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:52 crc kubenswrapper[4813]: I1201 08:40:52.782743 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:52 crc kubenswrapper[4813]: I1201 08:40:52.782762 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:52Z","lastTransitionTime":"2025-12-01T08:40:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:52 crc kubenswrapper[4813]: I1201 08:40:52.793051 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-xpnlc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"22da1fcf-0c57-4708-90d8-77692af17344\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4a36d213bdad2da91d0d33f400d5a8248befb68d5b94d19347873a93c6712a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xc44d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-xpnlc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:52Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:52 crc kubenswrapper[4813]: I1201 08:40:52.809141 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6fb5a648-ae4f-4e46-a530-007e9bf84abb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e24ebfc52dcb28368a2d436e7d9ab78f07f2aec2a4c71405a6c7d3301eee96a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33898118bbf639c8e6eae5aafa10386b95795512444ed5f535db3c663f169df2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9902c00bf7738686c93f62e1772a484d05715d74b75ee98c84c8f23ed383824b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3657b47ec9f1563c83bfbe3d728c8fd123f7078e219d251b1874269e42c31eea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:16Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:52Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:52 crc kubenswrapper[4813]: I1201 08:40:52.885442 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:52 crc kubenswrapper[4813]: I1201 08:40:52.885500 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:52 crc kubenswrapper[4813]: I1201 08:40:52.885517 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:52 crc kubenswrapper[4813]: I1201 08:40:52.885544 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:52 crc kubenswrapper[4813]: I1201 08:40:52.885564 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:52Z","lastTransitionTime":"2025-12-01T08:40:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:52 crc kubenswrapper[4813]: I1201 08:40:52.987361 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:52 crc kubenswrapper[4813]: I1201 08:40:52.987401 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:52 crc kubenswrapper[4813]: I1201 08:40:52.987413 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:52 crc kubenswrapper[4813]: I1201 08:40:52.987430 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:52 crc kubenswrapper[4813]: I1201 08:40:52.987444 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:52Z","lastTransitionTime":"2025-12-01T08:40:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:53 crc kubenswrapper[4813]: I1201 08:40:53.013628 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-8pg2p"] Dec 01 08:40:53 crc kubenswrapper[4813]: I1201 08:40:53.014552 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-8pg2p" Dec 01 08:40:53 crc kubenswrapper[4813]: I1201 08:40:53.017531 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Dec 01 08:40:53 crc kubenswrapper[4813]: I1201 08:40:53.018064 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Dec 01 08:40:53 crc kubenswrapper[4813]: I1201 08:40:53.038133 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:53Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:53 crc kubenswrapper[4813]: I1201 08:40:53.051427 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-c576v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"17292cd0-1eac-40b5-a60e-67167c57409c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a42206c914d0493ca74788ea1a941de80920ab7b810fb95227cd9b8711e8f4b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7zwfx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:41Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-c576v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:53Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:53 crc kubenswrapper[4813]: I1201 08:40:53.071807 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-j4pcc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b7505e14-3560-4d0d-be2c-d6a7a5953865\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ccc30a63dc1c23ff9880fb04af1f31f0656b2696d359988e23deed23989f480a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://944977e9722025088720107530d0c91fc693593f380e4b49678f28a56da180cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc32d930ab23f30103d8fa42291909cbec89ebb75c5c53cd783cd27fe1a3f09e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f19c67b1c6c80c7f3223512102810d46ede1445464d21ae3bbea83424539b2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d984c90302e57f3e5a6f3ca3613b415db04ac74a5a228582f1b52ef1d59f93e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce8963bf1e9ee66926907aaf3a03d18977f4d0a1983d7121b62fbc8e02a37e4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b11f49feda579bcb62f9680732033819ea11e7ae9937a06c0623a54c0ec106c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b45454b440c5108fa150a6b856afa1778ba6e9f89633433cb3117b68f00f72c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd67634e0c806323334495401a4e20b48fea4bc5013acd973d2abd3d70870f99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd67634e0c806323334495401a4e20b48fea4bc5013acd973d2abd3d70870f99\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:41Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-j4pcc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:53Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:53 crc kubenswrapper[4813]: I1201 08:40:53.086350 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-h7lvd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d431b613-a91b-4fff-88b9-447804766289\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://495593f99e92a9c28100a95e27d06ff5847008c5d996366cb474bedd9b69d5ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://495593f99e92a9c28100a95e27d06ff5847008c5d996366cb474bedd9b69d5ef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://19358f83a6c52957f959fe935a6843fdeef59905cc395e6e52a1fb4a670ac193\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://19358f83a6c52957f959fe935a6843fdeef59905cc395e6e52a1fb4a670ac193\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f089c115d916e339994879d224e3db27642b99aaabf59bcafb2979549052945\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9f089c115d916e339994879d224e3db27642b99aaabf59bcafb2979549052945\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb66f224462305c2005b31ef7bb9ed01efde8101e18518dd42d1dfb0df765bb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb66f224462305c2005b31ef7bb9ed01efde8101e18518dd42d1dfb0df765bb9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c268236c9a155c9b1f47c48d30dbd23948ceb2545592b40f356c60ea1d12e8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c268236c9a155c9b1f47c48d30dbd23948ceb2545592b40f356c60ea1d12e8a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://458aedca5f186e9aaed7e4ee29fd5247221491888b130f55c1fe48d88d565f8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://458aedca5f186e9aaed7e4ee29fd5247221491888b130f55c1fe48d88d565f8f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:41Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-h7lvd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:53Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:53 crc kubenswrapper[4813]: I1201 08:40:53.089945 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:53 crc kubenswrapper[4813]: I1201 08:40:53.090048 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:53 crc kubenswrapper[4813]: I1201 08:40:53.090067 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:53 crc kubenswrapper[4813]: I1201 08:40:53.090087 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:53 crc kubenswrapper[4813]: I1201 08:40:53.090100 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:53Z","lastTransitionTime":"2025-12-01T08:40:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:53 crc kubenswrapper[4813]: I1201 08:40:53.115382 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-xpnlc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"22da1fcf-0c57-4708-90d8-77692af17344\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4a36d213bdad2da91d0d33f400d5a8248befb68d5b94d19347873a93c6712a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xc44d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-xpnlc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:53Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:53 crc kubenswrapper[4813]: I1201 08:40:53.139424 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vnjhq\" (UniqueName: \"kubernetes.io/projected/dd53b38d-d9dd-4f98-9d06-dd95f1a4e3a3-kube-api-access-vnjhq\") pod \"ovnkube-control-plane-749d76644c-8pg2p\" (UID: \"dd53b38d-d9dd-4f98-9d06-dd95f1a4e3a3\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-8pg2p" Dec 01 08:40:53 crc kubenswrapper[4813]: I1201 08:40:53.139490 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/dd53b38d-d9dd-4f98-9d06-dd95f1a4e3a3-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-8pg2p\" (UID: \"dd53b38d-d9dd-4f98-9d06-dd95f1a4e3a3\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-8pg2p" Dec 01 08:40:53 crc kubenswrapper[4813]: I1201 08:40:53.139645 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/dd53b38d-d9dd-4f98-9d06-dd95f1a4e3a3-env-overrides\") pod \"ovnkube-control-plane-749d76644c-8pg2p\" (UID: \"dd53b38d-d9dd-4f98-9d06-dd95f1a4e3a3\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-8pg2p" Dec 01 08:40:53 crc kubenswrapper[4813]: I1201 08:40:53.139708 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/dd53b38d-d9dd-4f98-9d06-dd95f1a4e3a3-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-8pg2p\" (UID: \"dd53b38d-d9dd-4f98-9d06-dd95f1a4e3a3\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-8pg2p" Dec 01 08:40:53 crc kubenswrapper[4813]: I1201 08:40:53.154716 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6fb5a648-ae4f-4e46-a530-007e9bf84abb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e24ebfc52dcb28368a2d436e7d9ab78f07f2aec2a4c71405a6c7d3301eee96a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33898118bbf639c8e6eae5aafa10386b95795512444ed5f535db3c663f169df2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9902c00bf7738686c93f62e1772a484d05715d74b75ee98c84c8f23ed383824b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3657b47ec9f1563c83bfbe3d728c8fd123f7078e219d251b1874269e42c31eea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:16Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:53Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:53 crc kubenswrapper[4813]: I1201 08:40:53.170168 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0023b0e358d75aacc889ce364846dda89ea513832c87fa4b066aea7dac57e8a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:53Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:53 crc kubenswrapper[4813]: I1201 08:40:53.184866 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:53Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:53 crc kubenswrapper[4813]: I1201 08:40:53.193175 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:53 crc kubenswrapper[4813]: I1201 08:40:53.193214 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:53 crc kubenswrapper[4813]: I1201 08:40:53.193222 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:53 crc kubenswrapper[4813]: I1201 08:40:53.193236 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:53 crc kubenswrapper[4813]: I1201 08:40:53.193245 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:53Z","lastTransitionTime":"2025-12-01T08:40:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:53 crc kubenswrapper[4813]: I1201 08:40:53.201921 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c7574f892d5e9142296b26a0b484eb6759ab53f571139fb6e23b93dc3da05780\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:53Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:53 crc kubenswrapper[4813]: I1201 08:40:53.214441 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:53Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:53 crc kubenswrapper[4813]: I1201 08:40:53.226086 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"349332bd-6b4d-4306-84ae-07019ca3860f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d441a348087ae3cadda1ad8a7d1c4dff02cf3f6e1d346c294b278fdd9131edcc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://da67635daf6d9e32937f4d5d4ceeee3f8f09ae4fd2dd26c3e85be1286c5949de\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e97356ee274ea78d9f237ffbc756abb6ac8df0dbac244bf0a86ed215b8cbf43\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8dc6feb9817f4f356b8e36af43ec3070774f312fbbb4909fa05b8b3e8dd53582\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a0404f5a5ed5a9750d4c479b608e68e8960b448f455d5af9d219aa5edaeb789b\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"message\\\":\\\"\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764578421\\\\\\\\\\\\\\\" (2025-12-01 08:40:21 +0000 UTC to 2025-12-31 08:40:22 +0000 UTC (now=2025-12-01 08:40:38.066068451 +0000 UTC))\\\\\\\"\\\\nI1201 08:40:38.066144 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1201 08:40:38.066166 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1201 08:40:38.066220 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1201 08:40:38.066232 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1201 08:40:38.066338 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764578432\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764578432\\\\\\\\\\\\\\\" (2025-12-01 07:40:32 +0000 UTC to 2026-12-01 07:40:32 +0000 UTC (now=2025-12-01 08:40:38.066301217 +0000 UTC))\\\\\\\"\\\\nI1201 08:40:38.066357 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1201 08:40:38.066385 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1201 08:40:38.066402 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1201 08:40:38.066414 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-63890811/tls.crt::/tmp/serving-cert-63890811/tls.key\\\\\\\"\\\\nI1201 08:40:38.066419 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1201 08:40:38.066572 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1201 08:40:38.067159 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7354f9337d42115bb1d974bf3772f2d03365acba57f2155135c7a0bbab39879e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:19Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05492df5703392cdf0932d6bf66025b18e53bd05fa0857a7cb2ad70979cbdc3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05492df5703392cdf0932d6bf66025b18e53bd05fa0857a7cb2ad70979cbdc3e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:16Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:53Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:53 crc kubenswrapper[4813]: I1201 08:40:53.236812 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3264b3af996af452df2953c59bea6deeb5ea6af1b119ce89f7334595c27e70da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9e879b1f563bd66383d58fd7b435a8a81bd50cb387dc1fabb6fc47d5a3fa468\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:53Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:53 crc kubenswrapper[4813]: I1201 08:40:53.245431 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vnjhq\" (UniqueName: \"kubernetes.io/projected/dd53b38d-d9dd-4f98-9d06-dd95f1a4e3a3-kube-api-access-vnjhq\") pod \"ovnkube-control-plane-749d76644c-8pg2p\" (UID: \"dd53b38d-d9dd-4f98-9d06-dd95f1a4e3a3\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-8pg2p" Dec 01 08:40:53 crc kubenswrapper[4813]: I1201 08:40:53.245484 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/dd53b38d-d9dd-4f98-9d06-dd95f1a4e3a3-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-8pg2p\" (UID: \"dd53b38d-d9dd-4f98-9d06-dd95f1a4e3a3\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-8pg2p" Dec 01 08:40:53 crc kubenswrapper[4813]: I1201 08:40:53.245516 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/dd53b38d-d9dd-4f98-9d06-dd95f1a4e3a3-env-overrides\") pod \"ovnkube-control-plane-749d76644c-8pg2p\" (UID: \"dd53b38d-d9dd-4f98-9d06-dd95f1a4e3a3\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-8pg2p" Dec 01 08:40:53 crc kubenswrapper[4813]: I1201 08:40:53.245539 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/dd53b38d-d9dd-4f98-9d06-dd95f1a4e3a3-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-8pg2p\" (UID: \"dd53b38d-d9dd-4f98-9d06-dd95f1a4e3a3\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-8pg2p" Dec 01 08:40:53 crc kubenswrapper[4813]: I1201 08:40:53.246563 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/dd53b38d-d9dd-4f98-9d06-dd95f1a4e3a3-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-8pg2p\" (UID: \"dd53b38d-d9dd-4f98-9d06-dd95f1a4e3a3\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-8pg2p" Dec 01 08:40:53 crc kubenswrapper[4813]: I1201 08:40:53.246763 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/dd53b38d-d9dd-4f98-9d06-dd95f1a4e3a3-env-overrides\") pod \"ovnkube-control-plane-749d76644c-8pg2p\" (UID: \"dd53b38d-d9dd-4f98-9d06-dd95f1a4e3a3\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-8pg2p" Dec 01 08:40:53 crc kubenswrapper[4813]: I1201 08:40:53.251060 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-b6qqv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9656a03-0096-4c89-b0da-e37e4103611b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4195aebc1b2e388d480920a7f72257f6660afc99ad8c1a09478ef445739b272\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cjvnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:41Z\\\"}}\" for pod \"openshift-multus\"/\"multus-b6qqv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:53Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:53 crc kubenswrapper[4813]: I1201 08:40:53.255766 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/dd53b38d-d9dd-4f98-9d06-dd95f1a4e3a3-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-8pg2p\" (UID: \"dd53b38d-d9dd-4f98-9d06-dd95f1a4e3a3\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-8pg2p" Dec 01 08:40:53 crc kubenswrapper[4813]: I1201 08:40:53.262136 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vnjhq\" (UniqueName: \"kubernetes.io/projected/dd53b38d-d9dd-4f98-9d06-dd95f1a4e3a3-kube-api-access-vnjhq\") pod \"ovnkube-control-plane-749d76644c-8pg2p\" (UID: \"dd53b38d-d9dd-4f98-9d06-dd95f1a4e3a3\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-8pg2p" Dec 01 08:40:53 crc kubenswrapper[4813]: I1201 08:40:53.265550 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-9wn4q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3d423402-e08f-42e9-b60c-6eb9bf2f6d48\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3049b842978c6580e1033a442a37766a2e76e278aface9a89fc0b2c57faf0df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rzxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c915733072b32a63237e2401cbf84b059aca396c2c0db884f0f439d599a35d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rzxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:41Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-9wn4q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:53Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:53 crc kubenswrapper[4813]: I1201 08:40:53.276705 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-8pg2p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dd53b38d-d9dd-4f98-9d06-dd95f1a4e3a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnjhq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnjhq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:53Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-8pg2p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:53Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:53 crc kubenswrapper[4813]: I1201 08:40:53.295529 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:53 crc kubenswrapper[4813]: I1201 08:40:53.295668 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:53 crc kubenswrapper[4813]: I1201 08:40:53.295734 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:53 crc kubenswrapper[4813]: I1201 08:40:53.295821 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:53 crc kubenswrapper[4813]: I1201 08:40:53.295892 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:53Z","lastTransitionTime":"2025-12-01T08:40:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:53 crc kubenswrapper[4813]: I1201 08:40:53.329921 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-8pg2p" Dec 01 08:40:53 crc kubenswrapper[4813]: I1201 08:40:53.335826 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-h7lvd" event={"ID":"d431b613-a91b-4fff-88b9-447804766289","Type":"ContainerStarted","Data":"1a32c51c943fcf6ada7e3f614667eb023e29d94c11128d26c1fecca37b03963b"} Dec 01 08:40:53 crc kubenswrapper[4813]: W1201 08:40:53.345656 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddd53b38d_d9dd_4f98_9d06_dd95f1a4e3a3.slice/crio-662cc6f21ad408ee136abe2518df33da3728196b83c8cd5dc625715244a2f7f6 WatchSource:0}: Error finding container 662cc6f21ad408ee136abe2518df33da3728196b83c8cd5dc625715244a2f7f6: Status 404 returned error can't find the container with id 662cc6f21ad408ee136abe2518df33da3728196b83c8cd5dc625715244a2f7f6 Dec 01 08:40:53 crc kubenswrapper[4813]: I1201 08:40:53.350335 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-9wn4q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3d423402-e08f-42e9-b60c-6eb9bf2f6d48\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3049b842978c6580e1033a442a37766a2e76e278aface9a89fc0b2c57faf0df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rzxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c915733072b32a63237e2401cbf84b059aca396c2c0db884f0f439d599a35d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rzxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:41Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-9wn4q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:53Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:53 crc kubenswrapper[4813]: I1201 08:40:53.366475 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-8pg2p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dd53b38d-d9dd-4f98-9d06-dd95f1a4e3a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnjhq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnjhq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:53Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-8pg2p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:53Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:53 crc kubenswrapper[4813]: I1201 08:40:53.382133 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-h7lvd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d431b613-a91b-4fff-88b9-447804766289\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a32c51c943fcf6ada7e3f614667eb023e29d94c11128d26c1fecca37b03963b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://495593f99e92a9c28100a95e27d06ff5847008c5d996366cb474bedd9b69d5ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://495593f99e92a9c28100a95e27d06ff5847008c5d996366cb474bedd9b69d5ef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://19358f83a6c52957f959fe935a6843fdeef59905cc395e6e52a1fb4a670ac193\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://19358f83a6c52957f959fe935a6843fdeef59905cc395e6e52a1fb4a670ac193\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f089c115d916e339994879d224e3db27642b99aaabf59bcafb2979549052945\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9f089c115d916e339994879d224e3db27642b99aaabf59bcafb2979549052945\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb66f224462305c2005b31ef7bb9ed01efde8101e18518dd42d1dfb0df765bb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb66f224462305c2005b31ef7bb9ed01efde8101e18518dd42d1dfb0df765bb9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c268236c9a155c9b1f47c48d30dbd23948ceb2545592b40f356c60ea1d12e8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c268236c9a155c9b1f47c48d30dbd23948ceb2545592b40f356c60ea1d12e8a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://458aedca5f186e9aaed7e4ee29fd5247221491888b130f55c1fe48d88d565f8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://458aedca5f186e9aaed7e4ee29fd5247221491888b130f55c1fe48d88d565f8f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:41Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-h7lvd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:53Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:53 crc kubenswrapper[4813]: I1201 08:40:53.392694 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 08:40:53 crc kubenswrapper[4813]: E1201 08:40:53.392798 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 08:40:53 crc kubenswrapper[4813]: I1201 08:40:53.394019 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:53Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:53 crc kubenswrapper[4813]: I1201 08:40:53.457728 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:53 crc kubenswrapper[4813]: I1201 08:40:53.457751 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:53 crc kubenswrapper[4813]: I1201 08:40:53.457759 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:53 crc kubenswrapper[4813]: I1201 08:40:53.457772 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:53 crc kubenswrapper[4813]: I1201 08:40:53.457781 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:53Z","lastTransitionTime":"2025-12-01T08:40:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:53 crc kubenswrapper[4813]: I1201 08:40:53.464539 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-c576v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"17292cd0-1eac-40b5-a60e-67167c57409c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a42206c914d0493ca74788ea1a941de80920ab7b810fb95227cd9b8711e8f4b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7zwfx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:41Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-c576v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:53Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:53 crc kubenswrapper[4813]: I1201 08:40:53.495090 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-j4pcc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b7505e14-3560-4d0d-be2c-d6a7a5953865\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ccc30a63dc1c23ff9880fb04af1f31f0656b2696d359988e23deed23989f480a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://944977e9722025088720107530d0c91fc693593f380e4b49678f28a56da180cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc32d930ab23f30103d8fa42291909cbec89ebb75c5c53cd783cd27fe1a3f09e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f19c67b1c6c80c7f3223512102810d46ede1445464d21ae3bbea83424539b2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d984c90302e57f3e5a6f3ca3613b415db04ac74a5a228582f1b52ef1d59f93e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce8963bf1e9ee66926907aaf3a03d18977f4d0a1983d7121b62fbc8e02a37e4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b11f49feda579bcb62f9680732033819ea11e7ae9937a06c0623a54c0ec106c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b45454b440c5108fa150a6b856afa1778ba6e9f89633433cb3117b68f00f72c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd67634e0c806323334495401a4e20b48fea4bc5013acd973d2abd3d70870f99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd67634e0c806323334495401a4e20b48fea4bc5013acd973d2abd3d70870f99\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:41Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-j4pcc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:53Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:53 crc kubenswrapper[4813]: I1201 08:40:53.513983 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:53Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:53 crc kubenswrapper[4813]: I1201 08:40:53.527422 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-xpnlc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"22da1fcf-0c57-4708-90d8-77692af17344\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4a36d213bdad2da91d0d33f400d5a8248befb68d5b94d19347873a93c6712a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xc44d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-xpnlc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:53Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:53 crc kubenswrapper[4813]: I1201 08:40:53.543088 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6fb5a648-ae4f-4e46-a530-007e9bf84abb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e24ebfc52dcb28368a2d436e7d9ab78f07f2aec2a4c71405a6c7d3301eee96a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33898118bbf639c8e6eae5aafa10386b95795512444ed5f535db3c663f169df2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9902c00bf7738686c93f62e1772a484d05715d74b75ee98c84c8f23ed383824b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3657b47ec9f1563c83bfbe3d728c8fd123f7078e219d251b1874269e42c31eea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:16Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:53Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:53 crc kubenswrapper[4813]: I1201 08:40:53.560257 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:53 crc kubenswrapper[4813]: I1201 08:40:53.560292 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:53 crc kubenswrapper[4813]: I1201 08:40:53.560300 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:53 crc kubenswrapper[4813]: I1201 08:40:53.560313 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:53 crc kubenswrapper[4813]: I1201 08:40:53.560323 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:53Z","lastTransitionTime":"2025-12-01T08:40:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:53 crc kubenswrapper[4813]: I1201 08:40:53.563181 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0023b0e358d75aacc889ce364846dda89ea513832c87fa4b066aea7dac57e8a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:53Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:53 crc kubenswrapper[4813]: I1201 08:40:53.601211 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:53Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:53 crc kubenswrapper[4813]: I1201 08:40:53.614160 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c7574f892d5e9142296b26a0b484eb6759ab53f571139fb6e23b93dc3da05780\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:53Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:53 crc kubenswrapper[4813]: I1201 08:40:53.629113 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"349332bd-6b4d-4306-84ae-07019ca3860f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d441a348087ae3cadda1ad8a7d1c4dff02cf3f6e1d346c294b278fdd9131edcc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://da67635daf6d9e32937f4d5d4ceeee3f8f09ae4fd2dd26c3e85be1286c5949de\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e97356ee274ea78d9f237ffbc756abb6ac8df0dbac244bf0a86ed215b8cbf43\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8dc6feb9817f4f356b8e36af43ec3070774f312fbbb4909fa05b8b3e8dd53582\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a0404f5a5ed5a9750d4c479b608e68e8960b448f455d5af9d219aa5edaeb789b\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"message\\\":\\\"\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764578421\\\\\\\\\\\\\\\" (2025-12-01 08:40:21 +0000 UTC to 2025-12-31 08:40:22 +0000 UTC (now=2025-12-01 08:40:38.066068451 +0000 UTC))\\\\\\\"\\\\nI1201 08:40:38.066144 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1201 08:40:38.066166 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1201 08:40:38.066220 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1201 08:40:38.066232 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1201 08:40:38.066338 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764578432\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764578432\\\\\\\\\\\\\\\" (2025-12-01 07:40:32 +0000 UTC to 2026-12-01 07:40:32 +0000 UTC (now=2025-12-01 08:40:38.066301217 +0000 UTC))\\\\\\\"\\\\nI1201 08:40:38.066357 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1201 08:40:38.066385 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1201 08:40:38.066402 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1201 08:40:38.066414 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-63890811/tls.crt::/tmp/serving-cert-63890811/tls.key\\\\\\\"\\\\nI1201 08:40:38.066419 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1201 08:40:38.066572 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1201 08:40:38.067159 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7354f9337d42115bb1d974bf3772f2d03365acba57f2155135c7a0bbab39879e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:19Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05492df5703392cdf0932d6bf66025b18e53bd05fa0857a7cb2ad70979cbdc3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05492df5703392cdf0932d6bf66025b18e53bd05fa0857a7cb2ad70979cbdc3e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:16Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:53Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:53 crc kubenswrapper[4813]: I1201 08:40:53.641268 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3264b3af996af452df2953c59bea6deeb5ea6af1b119ce89f7334595c27e70da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9e879b1f563bd66383d58fd7b435a8a81bd50cb387dc1fabb6fc47d5a3fa468\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:53Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:53 crc kubenswrapper[4813]: I1201 08:40:53.654709 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-b6qqv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9656a03-0096-4c89-b0da-e37e4103611b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4195aebc1b2e388d480920a7f72257f6660afc99ad8c1a09478ef445739b272\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cjvnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:41Z\\\"}}\" for pod \"openshift-multus\"/\"multus-b6qqv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:53Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:53 crc kubenswrapper[4813]: I1201 08:40:53.662527 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:53 crc kubenswrapper[4813]: I1201 08:40:53.662561 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:53 crc kubenswrapper[4813]: I1201 08:40:53.662573 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:53 crc kubenswrapper[4813]: I1201 08:40:53.662588 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:53 crc kubenswrapper[4813]: I1201 08:40:53.662599 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:53Z","lastTransitionTime":"2025-12-01T08:40:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:53 crc kubenswrapper[4813]: I1201 08:40:53.765010 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:53 crc kubenswrapper[4813]: I1201 08:40:53.765044 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:53 crc kubenswrapper[4813]: I1201 08:40:53.765054 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:53 crc kubenswrapper[4813]: I1201 08:40:53.765068 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:53 crc kubenswrapper[4813]: I1201 08:40:53.765078 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:53Z","lastTransitionTime":"2025-12-01T08:40:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:53 crc kubenswrapper[4813]: I1201 08:40:53.869283 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:53 crc kubenswrapper[4813]: I1201 08:40:53.869337 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:53 crc kubenswrapper[4813]: I1201 08:40:53.869357 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:53 crc kubenswrapper[4813]: I1201 08:40:53.869380 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:53 crc kubenswrapper[4813]: I1201 08:40:53.869397 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:53Z","lastTransitionTime":"2025-12-01T08:40:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:53 crc kubenswrapper[4813]: I1201 08:40:53.972059 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:53 crc kubenswrapper[4813]: I1201 08:40:53.972097 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:53 crc kubenswrapper[4813]: I1201 08:40:53.972108 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:53 crc kubenswrapper[4813]: I1201 08:40:53.972127 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:53 crc kubenswrapper[4813]: I1201 08:40:53.972137 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:53Z","lastTransitionTime":"2025-12-01T08:40:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:54 crc kubenswrapper[4813]: I1201 08:40:54.074524 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:54 crc kubenswrapper[4813]: I1201 08:40:54.074563 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:54 crc kubenswrapper[4813]: I1201 08:40:54.074571 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:54 crc kubenswrapper[4813]: I1201 08:40:54.074586 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:54 crc kubenswrapper[4813]: I1201 08:40:54.074596 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:54Z","lastTransitionTime":"2025-12-01T08:40:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:54 crc kubenswrapper[4813]: I1201 08:40:54.171398 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 08:40:54 crc kubenswrapper[4813]: I1201 08:40:54.171501 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 08:40:54 crc kubenswrapper[4813]: I1201 08:40:54.171527 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 08:40:54 crc kubenswrapper[4813]: E1201 08:40:54.171672 4813 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 01 08:40:54 crc kubenswrapper[4813]: E1201 08:40:54.171752 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-01 08:41:10.17172283 +0000 UTC m=+54.794644406 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 01 08:40:54 crc kubenswrapper[4813]: E1201 08:40:54.171858 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 08:41:10.171852354 +0000 UTC m=+54.794773940 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:40:54 crc kubenswrapper[4813]: E1201 08:40:54.171903 4813 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 01 08:40:54 crc kubenswrapper[4813]: E1201 08:40:54.171932 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-01 08:41:10.171926856 +0000 UTC m=+54.794848442 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 01 08:40:54 crc kubenswrapper[4813]: I1201 08:40:54.176908 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:54 crc kubenswrapper[4813]: I1201 08:40:54.176942 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:54 crc kubenswrapper[4813]: I1201 08:40:54.176950 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:54 crc kubenswrapper[4813]: I1201 08:40:54.176978 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:54 crc kubenswrapper[4813]: I1201 08:40:54.176987 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:54Z","lastTransitionTime":"2025-12-01T08:40:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:54 crc kubenswrapper[4813]: I1201 08:40:54.273007 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 08:40:54 crc kubenswrapper[4813]: I1201 08:40:54.273072 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 08:40:54 crc kubenswrapper[4813]: E1201 08:40:54.273209 4813 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 01 08:40:54 crc kubenswrapper[4813]: E1201 08:40:54.273227 4813 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 01 08:40:54 crc kubenswrapper[4813]: E1201 08:40:54.273243 4813 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 08:40:54 crc kubenswrapper[4813]: E1201 08:40:54.273275 4813 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 01 08:40:54 crc kubenswrapper[4813]: E1201 08:40:54.273311 4813 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 01 08:40:54 crc kubenswrapper[4813]: E1201 08:40:54.273323 4813 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 08:40:54 crc kubenswrapper[4813]: E1201 08:40:54.273295 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-01 08:41:10.273281242 +0000 UTC m=+54.896202828 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 08:40:54 crc kubenswrapper[4813]: E1201 08:40:54.273465 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-01 08:41:10.273408056 +0000 UTC m=+54.896329662 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 08:40:54 crc kubenswrapper[4813]: I1201 08:40:54.279139 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:54 crc kubenswrapper[4813]: I1201 08:40:54.279195 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:54 crc kubenswrapper[4813]: I1201 08:40:54.279207 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:54 crc kubenswrapper[4813]: I1201 08:40:54.279225 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:54 crc kubenswrapper[4813]: I1201 08:40:54.279237 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:54Z","lastTransitionTime":"2025-12-01T08:40:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:54 crc kubenswrapper[4813]: I1201 08:40:54.348230 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-8pg2p" event={"ID":"dd53b38d-d9dd-4f98-9d06-dd95f1a4e3a3","Type":"ContainerStarted","Data":"662cc6f21ad408ee136abe2518df33da3728196b83c8cd5dc625715244a2f7f6"} Dec 01 08:40:54 crc kubenswrapper[4813]: I1201 08:40:54.381685 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:54 crc kubenswrapper[4813]: I1201 08:40:54.381730 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:54 crc kubenswrapper[4813]: I1201 08:40:54.381738 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:54 crc kubenswrapper[4813]: I1201 08:40:54.381753 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:54 crc kubenswrapper[4813]: I1201 08:40:54.381763 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:54Z","lastTransitionTime":"2025-12-01T08:40:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:54 crc kubenswrapper[4813]: I1201 08:40:54.393337 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 08:40:54 crc kubenswrapper[4813]: E1201 08:40:54.393466 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 08:40:54 crc kubenswrapper[4813]: I1201 08:40:54.393337 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 08:40:54 crc kubenswrapper[4813]: E1201 08:40:54.393565 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 08:40:54 crc kubenswrapper[4813]: I1201 08:40:54.487757 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:54 crc kubenswrapper[4813]: I1201 08:40:54.487819 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:54 crc kubenswrapper[4813]: I1201 08:40:54.487829 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:54 crc kubenswrapper[4813]: I1201 08:40:54.487855 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:54 crc kubenswrapper[4813]: I1201 08:40:54.487866 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:54Z","lastTransitionTime":"2025-12-01T08:40:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:54 crc kubenswrapper[4813]: I1201 08:40:54.591979 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:54 crc kubenswrapper[4813]: I1201 08:40:54.592019 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:54 crc kubenswrapper[4813]: I1201 08:40:54.592136 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:54 crc kubenswrapper[4813]: I1201 08:40:54.592152 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:54 crc kubenswrapper[4813]: I1201 08:40:54.592172 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:54Z","lastTransitionTime":"2025-12-01T08:40:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:54 crc kubenswrapper[4813]: I1201 08:40:54.696589 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:54 crc kubenswrapper[4813]: I1201 08:40:54.696627 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:54 crc kubenswrapper[4813]: I1201 08:40:54.696636 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:54 crc kubenswrapper[4813]: I1201 08:40:54.696651 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:54 crc kubenswrapper[4813]: I1201 08:40:54.696659 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:54Z","lastTransitionTime":"2025-12-01T08:40:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:54 crc kubenswrapper[4813]: I1201 08:40:54.799665 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:54 crc kubenswrapper[4813]: I1201 08:40:54.800109 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:54 crc kubenswrapper[4813]: I1201 08:40:54.800303 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:54 crc kubenswrapper[4813]: I1201 08:40:54.800491 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:54 crc kubenswrapper[4813]: I1201 08:40:54.800668 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:54Z","lastTransitionTime":"2025-12-01T08:40:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:54 crc kubenswrapper[4813]: I1201 08:40:54.903606 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:54 crc kubenswrapper[4813]: I1201 08:40:54.903897 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:54 crc kubenswrapper[4813]: I1201 08:40:54.904026 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:54 crc kubenswrapper[4813]: I1201 08:40:54.904117 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:54 crc kubenswrapper[4813]: I1201 08:40:54.904206 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:54Z","lastTransitionTime":"2025-12-01T08:40:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:55 crc kubenswrapper[4813]: I1201 08:40:55.007804 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:55 crc kubenswrapper[4813]: I1201 08:40:55.007871 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:55 crc kubenswrapper[4813]: I1201 08:40:55.007888 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:55 crc kubenswrapper[4813]: I1201 08:40:55.007909 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:55 crc kubenswrapper[4813]: I1201 08:40:55.007925 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:55Z","lastTransitionTime":"2025-12-01T08:40:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:55 crc kubenswrapper[4813]: I1201 08:40:55.109797 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:55 crc kubenswrapper[4813]: I1201 08:40:55.109838 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:55 crc kubenswrapper[4813]: I1201 08:40:55.109851 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:55 crc kubenswrapper[4813]: I1201 08:40:55.109868 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:55 crc kubenswrapper[4813]: I1201 08:40:55.109879 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:55Z","lastTransitionTime":"2025-12-01T08:40:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:55 crc kubenswrapper[4813]: I1201 08:40:55.212500 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:55 crc kubenswrapper[4813]: I1201 08:40:55.212715 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:55 crc kubenswrapper[4813]: I1201 08:40:55.212825 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:55 crc kubenswrapper[4813]: I1201 08:40:55.212914 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:55 crc kubenswrapper[4813]: I1201 08:40:55.212994 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:55Z","lastTransitionTime":"2025-12-01T08:40:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:55 crc kubenswrapper[4813]: I1201 08:40:55.315892 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:55 crc kubenswrapper[4813]: I1201 08:40:55.315939 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:55 crc kubenswrapper[4813]: I1201 08:40:55.315956 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:55 crc kubenswrapper[4813]: I1201 08:40:55.315994 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:55 crc kubenswrapper[4813]: I1201 08:40:55.316007 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:55Z","lastTransitionTime":"2025-12-01T08:40:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:55 crc kubenswrapper[4813]: I1201 08:40:55.353795 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-j4pcc_b7505e14-3560-4d0d-be2c-d6a7a5953865/ovnkube-controller/0.log" Dec 01 08:40:55 crc kubenswrapper[4813]: I1201 08:40:55.357015 4813 generic.go:334] "Generic (PLEG): container finished" podID="b7505e14-3560-4d0d-be2c-d6a7a5953865" containerID="4b11f49feda579bcb62f9680732033819ea11e7ae9937a06c0623a54c0ec106c" exitCode=1 Dec 01 08:40:55 crc kubenswrapper[4813]: I1201 08:40:55.357087 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-j4pcc" event={"ID":"b7505e14-3560-4d0d-be2c-d6a7a5953865","Type":"ContainerDied","Data":"4b11f49feda579bcb62f9680732033819ea11e7ae9937a06c0623a54c0ec106c"} Dec 01 08:40:55 crc kubenswrapper[4813]: I1201 08:40:55.357828 4813 scope.go:117] "RemoveContainer" containerID="4b11f49feda579bcb62f9680732033819ea11e7ae9937a06c0623a54c0ec106c" Dec 01 08:40:55 crc kubenswrapper[4813]: I1201 08:40:55.359602 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-8pg2p" event={"ID":"dd53b38d-d9dd-4f98-9d06-dd95f1a4e3a3","Type":"ContainerStarted","Data":"2ddf9d5989e4195b3074de08f99d53b48066f53fe93ce204145b3c714996fc57"} Dec 01 08:40:55 crc kubenswrapper[4813]: I1201 08:40:55.359637 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-8pg2p" event={"ID":"dd53b38d-d9dd-4f98-9d06-dd95f1a4e3a3","Type":"ContainerStarted","Data":"2a214d71293e4e09573c665524d880fc319dfb1c0c6d95d11d78f253865660fa"} Dec 01 08:40:55 crc kubenswrapper[4813]: I1201 08:40:55.379245 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:55Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:55 crc kubenswrapper[4813]: I1201 08:40:55.392587 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 08:40:55 crc kubenswrapper[4813]: E1201 08:40:55.392721 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 08:40:55 crc kubenswrapper[4813]: I1201 08:40:55.393375 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-c576v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"17292cd0-1eac-40b5-a60e-67167c57409c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a42206c914d0493ca74788ea1a941de80920ab7b810fb95227cd9b8711e8f4b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7zwfx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:41Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-c576v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:55Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:55 crc kubenswrapper[4813]: I1201 08:40:55.411630 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-j4pcc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b7505e14-3560-4d0d-be2c-d6a7a5953865\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ccc30a63dc1c23ff9880fb04af1f31f0656b2696d359988e23deed23989f480a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://944977e9722025088720107530d0c91fc693593f380e4b49678f28a56da180cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc32d930ab23f30103d8fa42291909cbec89ebb75c5c53cd783cd27fe1a3f09e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f19c67b1c6c80c7f3223512102810d46ede1445464d21ae3bbea83424539b2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d984c90302e57f3e5a6f3ca3613b415db04ac74a5a228582f1b52ef1d59f93e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce8963bf1e9ee66926907aaf3a03d18977f4d0a1983d7121b62fbc8e02a37e4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b11f49feda579bcb62f9680732033819ea11e7ae9937a06c0623a54c0ec106c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4b11f49feda579bcb62f9680732033819ea11e7ae9937a06c0623a54c0ec106c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T08:40:54Z\\\",\\\"message\\\":\\\" *v1.Pod event handler 3\\\\nI1201 08:40:53.558214 6038 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1201 08:40:53.559209 6038 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1201 08:40:53.559251 6038 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1201 08:40:53.559264 6038 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1201 08:40:53.559268 6038 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1201 08:40:53.559286 6038 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1201 08:40:53.559299 6038 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1201 08:40:53.559315 6038 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1201 08:40:53.559320 6038 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1201 08:40:53.559327 6038 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1201 08:40:53.559329 6038 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1201 08:40:53.559336 6038 handler.go:208] Removed *v1.Node event handler 7\\\\nI1201 08:40:53.559354 6038 factory.go:656] Stopping watch factory\\\\nI1201 08:40:53.559365 6038 handler.go:208] Removed *v1.Node ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b45454b440c5108fa150a6b856afa1778ba6e9f89633433cb3117b68f00f72c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd67634e0c806323334495401a4e20b48fea4bc5013acd973d2abd3d70870f99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd67634e0c806323334495401a4e20b48fea4bc5013acd973d2abd3d70870f99\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:41Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-j4pcc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:55Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:55 crc kubenswrapper[4813]: I1201 08:40:55.418907 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:55 crc kubenswrapper[4813]: I1201 08:40:55.419172 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:55 crc kubenswrapper[4813]: I1201 08:40:55.419193 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:55 crc kubenswrapper[4813]: I1201 08:40:55.419210 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:55 crc kubenswrapper[4813]: I1201 08:40:55.419221 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:55Z","lastTransitionTime":"2025-12-01T08:40:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:55 crc kubenswrapper[4813]: I1201 08:40:55.427608 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-h7lvd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d431b613-a91b-4fff-88b9-447804766289\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a32c51c943fcf6ada7e3f614667eb023e29d94c11128d26c1fecca37b03963b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://495593f99e92a9c28100a95e27d06ff5847008c5d996366cb474bedd9b69d5ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://495593f99e92a9c28100a95e27d06ff5847008c5d996366cb474bedd9b69d5ef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://19358f83a6c52957f959fe935a6843fdeef59905cc395e6e52a1fb4a670ac193\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://19358f83a6c52957f959fe935a6843fdeef59905cc395e6e52a1fb4a670ac193\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f089c115d916e339994879d224e3db27642b99aaabf59bcafb2979549052945\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9f089c115d916e339994879d224e3db27642b99aaabf59bcafb2979549052945\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb66f224462305c2005b31ef7bb9ed01efde8101e18518dd42d1dfb0df765bb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb66f224462305c2005b31ef7bb9ed01efde8101e18518dd42d1dfb0df765bb9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c268236c9a155c9b1f47c48d30dbd23948ceb2545592b40f356c60ea1d12e8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c268236c9a155c9b1f47c48d30dbd23948ceb2545592b40f356c60ea1d12e8a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://458aedca5f186e9aaed7e4ee29fd5247221491888b130f55c1fe48d88d565f8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://458aedca5f186e9aaed7e4ee29fd5247221491888b130f55c1fe48d88d565f8f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:41Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-h7lvd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:55Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:55 crc kubenswrapper[4813]: I1201 08:40:55.442078 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6fb5a648-ae4f-4e46-a530-007e9bf84abb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e24ebfc52dcb28368a2d436e7d9ab78f07f2aec2a4c71405a6c7d3301eee96a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33898118bbf639c8e6eae5aafa10386b95795512444ed5f535db3c663f169df2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9902c00bf7738686c93f62e1772a484d05715d74b75ee98c84c8f23ed383824b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3657b47ec9f1563c83bfbe3d728c8fd123f7078e219d251b1874269e42c31eea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:16Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:55Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:55 crc kubenswrapper[4813]: I1201 08:40:55.455628 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0023b0e358d75aacc889ce364846dda89ea513832c87fa4b066aea7dac57e8a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:55Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:55 crc kubenswrapper[4813]: I1201 08:40:55.469432 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:55Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:55 crc kubenswrapper[4813]: I1201 08:40:55.484459 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c7574f892d5e9142296b26a0b484eb6759ab53f571139fb6e23b93dc3da05780\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:55Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:55 crc kubenswrapper[4813]: I1201 08:40:55.499074 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:55Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:55 crc kubenswrapper[4813]: I1201 08:40:55.514431 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-xpnlc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"22da1fcf-0c57-4708-90d8-77692af17344\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4a36d213bdad2da91d0d33f400d5a8248befb68d5b94d19347873a93c6712a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xc44d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-xpnlc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:55Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:55 crc kubenswrapper[4813]: I1201 08:40:55.520944 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:55 crc kubenswrapper[4813]: I1201 08:40:55.521011 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:55 crc kubenswrapper[4813]: I1201 08:40:55.521029 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:55 crc kubenswrapper[4813]: I1201 08:40:55.521068 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:55 crc kubenswrapper[4813]: I1201 08:40:55.521091 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:55Z","lastTransitionTime":"2025-12-01T08:40:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:55 crc kubenswrapper[4813]: I1201 08:40:55.530955 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"349332bd-6b4d-4306-84ae-07019ca3860f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d441a348087ae3cadda1ad8a7d1c4dff02cf3f6e1d346c294b278fdd9131edcc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://da67635daf6d9e32937f4d5d4ceeee3f8f09ae4fd2dd26c3e85be1286c5949de\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e97356ee274ea78d9f237ffbc756abb6ac8df0dbac244bf0a86ed215b8cbf43\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8dc6feb9817f4f356b8e36af43ec3070774f312fbbb4909fa05b8b3e8dd53582\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a0404f5a5ed5a9750d4c479b608e68e8960b448f455d5af9d219aa5edaeb789b\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"message\\\":\\\"\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764578421\\\\\\\\\\\\\\\" (2025-12-01 08:40:21 +0000 UTC to 2025-12-31 08:40:22 +0000 UTC (now=2025-12-01 08:40:38.066068451 +0000 UTC))\\\\\\\"\\\\nI1201 08:40:38.066144 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1201 08:40:38.066166 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1201 08:40:38.066220 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1201 08:40:38.066232 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1201 08:40:38.066338 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764578432\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764578432\\\\\\\\\\\\\\\" (2025-12-01 07:40:32 +0000 UTC to 2026-12-01 07:40:32 +0000 UTC (now=2025-12-01 08:40:38.066301217 +0000 UTC))\\\\\\\"\\\\nI1201 08:40:38.066357 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1201 08:40:38.066385 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1201 08:40:38.066402 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1201 08:40:38.066414 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-63890811/tls.crt::/tmp/serving-cert-63890811/tls.key\\\\\\\"\\\\nI1201 08:40:38.066419 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1201 08:40:38.066572 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1201 08:40:38.067159 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7354f9337d42115bb1d974bf3772f2d03365acba57f2155135c7a0bbab39879e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:19Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05492df5703392cdf0932d6bf66025b18e53bd05fa0857a7cb2ad70979cbdc3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05492df5703392cdf0932d6bf66025b18e53bd05fa0857a7cb2ad70979cbdc3e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:16Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:55Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:55 crc kubenswrapper[4813]: I1201 08:40:55.543306 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3264b3af996af452df2953c59bea6deeb5ea6af1b119ce89f7334595c27e70da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9e879b1f563bd66383d58fd7b435a8a81bd50cb387dc1fabb6fc47d5a3fa468\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:55Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:55 crc kubenswrapper[4813]: I1201 08:40:55.554710 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-b6qqv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9656a03-0096-4c89-b0da-e37e4103611b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4195aebc1b2e388d480920a7f72257f6660afc99ad8c1a09478ef445739b272\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cjvnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:41Z\\\"}}\" for pod \"openshift-multus\"/\"multus-b6qqv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:55Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:55 crc kubenswrapper[4813]: I1201 08:40:55.564673 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-9wn4q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3d423402-e08f-42e9-b60c-6eb9bf2f6d48\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3049b842978c6580e1033a442a37766a2e76e278aface9a89fc0b2c57faf0df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rzxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c915733072b32a63237e2401cbf84b059aca396c2c0db884f0f439d599a35d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rzxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:41Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-9wn4q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:55Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:55 crc kubenswrapper[4813]: I1201 08:40:55.574824 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-8pg2p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dd53b38d-d9dd-4f98-9d06-dd95f1a4e3a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnjhq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnjhq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:53Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-8pg2p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:55Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:55 crc kubenswrapper[4813]: I1201 08:40:55.587571 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-h7lvd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d431b613-a91b-4fff-88b9-447804766289\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a32c51c943fcf6ada7e3f614667eb023e29d94c11128d26c1fecca37b03963b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://495593f99e92a9c28100a95e27d06ff5847008c5d996366cb474bedd9b69d5ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://495593f99e92a9c28100a95e27d06ff5847008c5d996366cb474bedd9b69d5ef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://19358f83a6c52957f959fe935a6843fdeef59905cc395e6e52a1fb4a670ac193\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://19358f83a6c52957f959fe935a6843fdeef59905cc395e6e52a1fb4a670ac193\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f089c115d916e339994879d224e3db27642b99aaabf59bcafb2979549052945\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9f089c115d916e339994879d224e3db27642b99aaabf59bcafb2979549052945\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb66f224462305c2005b31ef7bb9ed01efde8101e18518dd42d1dfb0df765bb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb66f224462305c2005b31ef7bb9ed01efde8101e18518dd42d1dfb0df765bb9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c268236c9a155c9b1f47c48d30dbd23948ceb2545592b40f356c60ea1d12e8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c268236c9a155c9b1f47c48d30dbd23948ceb2545592b40f356c60ea1d12e8a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://458aedca5f186e9aaed7e4ee29fd5247221491888b130f55c1fe48d88d565f8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://458aedca5f186e9aaed7e4ee29fd5247221491888b130f55c1fe48d88d565f8f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:41Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-h7lvd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:55Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:55 crc kubenswrapper[4813]: I1201 08:40:55.598051 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:55Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:55 crc kubenswrapper[4813]: I1201 08:40:55.609278 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-c576v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"17292cd0-1eac-40b5-a60e-67167c57409c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a42206c914d0493ca74788ea1a941de80920ab7b810fb95227cd9b8711e8f4b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7zwfx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:41Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-c576v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:55Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:55 crc kubenswrapper[4813]: I1201 08:40:55.623815 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:55 crc kubenswrapper[4813]: I1201 08:40:55.623864 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:55 crc kubenswrapper[4813]: I1201 08:40:55.623876 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:55 crc kubenswrapper[4813]: I1201 08:40:55.623894 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:55 crc kubenswrapper[4813]: I1201 08:40:55.623906 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:55Z","lastTransitionTime":"2025-12-01T08:40:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:55 crc kubenswrapper[4813]: I1201 08:40:55.627724 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-j4pcc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b7505e14-3560-4d0d-be2c-d6a7a5953865\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ccc30a63dc1c23ff9880fb04af1f31f0656b2696d359988e23deed23989f480a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://944977e9722025088720107530d0c91fc693593f380e4b49678f28a56da180cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc32d930ab23f30103d8fa42291909cbec89ebb75c5c53cd783cd27fe1a3f09e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f19c67b1c6c80c7f3223512102810d46ede1445464d21ae3bbea83424539b2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d984c90302e57f3e5a6f3ca3613b415db04ac74a5a228582f1b52ef1d59f93e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce8963bf1e9ee66926907aaf3a03d18977f4d0a1983d7121b62fbc8e02a37e4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b11f49feda579bcb62f9680732033819ea11e7ae9937a06c0623a54c0ec106c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4b11f49feda579bcb62f9680732033819ea11e7ae9937a06c0623a54c0ec106c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T08:40:54Z\\\",\\\"message\\\":\\\" *v1.Pod event handler 3\\\\nI1201 08:40:53.558214 6038 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1201 08:40:53.559209 6038 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1201 08:40:53.559251 6038 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1201 08:40:53.559264 6038 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1201 08:40:53.559268 6038 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1201 08:40:53.559286 6038 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1201 08:40:53.559299 6038 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1201 08:40:53.559315 6038 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1201 08:40:53.559320 6038 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1201 08:40:53.559327 6038 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1201 08:40:53.559329 6038 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1201 08:40:53.559336 6038 handler.go:208] Removed *v1.Node event handler 7\\\\nI1201 08:40:53.559354 6038 factory.go:656] Stopping watch factory\\\\nI1201 08:40:53.559365 6038 handler.go:208] Removed *v1.Node ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b45454b440c5108fa150a6b856afa1778ba6e9f89633433cb3117b68f00f72c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd67634e0c806323334495401a4e20b48fea4bc5013acd973d2abd3d70870f99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd67634e0c806323334495401a4e20b48fea4bc5013acd973d2abd3d70870f99\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:41Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-j4pcc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:55Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:55 crc kubenswrapper[4813]: I1201 08:40:55.640748 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:55Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:55 crc kubenswrapper[4813]: I1201 08:40:55.661133 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-xpnlc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"22da1fcf-0c57-4708-90d8-77692af17344\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4a36d213bdad2da91d0d33f400d5a8248befb68d5b94d19347873a93c6712a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xc44d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-xpnlc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:55Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:55 crc kubenswrapper[4813]: I1201 08:40:55.675495 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6fb5a648-ae4f-4e46-a530-007e9bf84abb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e24ebfc52dcb28368a2d436e7d9ab78f07f2aec2a4c71405a6c7d3301eee96a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33898118bbf639c8e6eae5aafa10386b95795512444ed5f535db3c663f169df2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9902c00bf7738686c93f62e1772a484d05715d74b75ee98c84c8f23ed383824b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3657b47ec9f1563c83bfbe3d728c8fd123f7078e219d251b1874269e42c31eea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:16Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:55Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:55 crc kubenswrapper[4813]: I1201 08:40:55.688869 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0023b0e358d75aacc889ce364846dda89ea513832c87fa4b066aea7dac57e8a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:55Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:55 crc kubenswrapper[4813]: I1201 08:40:55.700327 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:55Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:55 crc kubenswrapper[4813]: I1201 08:40:55.709562 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c7574f892d5e9142296b26a0b484eb6759ab53f571139fb6e23b93dc3da05780\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:55Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:55 crc kubenswrapper[4813]: I1201 08:40:55.721402 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"349332bd-6b4d-4306-84ae-07019ca3860f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d441a348087ae3cadda1ad8a7d1c4dff02cf3f6e1d346c294b278fdd9131edcc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://da67635daf6d9e32937f4d5d4ceeee3f8f09ae4fd2dd26c3e85be1286c5949de\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e97356ee274ea78d9f237ffbc756abb6ac8df0dbac244bf0a86ed215b8cbf43\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8dc6feb9817f4f356b8e36af43ec3070774f312fbbb4909fa05b8b3e8dd53582\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a0404f5a5ed5a9750d4c479b608e68e8960b448f455d5af9d219aa5edaeb789b\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"message\\\":\\\"\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764578421\\\\\\\\\\\\\\\" (2025-12-01 08:40:21 +0000 UTC to 2025-12-31 08:40:22 +0000 UTC (now=2025-12-01 08:40:38.066068451 +0000 UTC))\\\\\\\"\\\\nI1201 08:40:38.066144 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1201 08:40:38.066166 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1201 08:40:38.066220 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1201 08:40:38.066232 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1201 08:40:38.066338 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764578432\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764578432\\\\\\\\\\\\\\\" (2025-12-01 07:40:32 +0000 UTC to 2026-12-01 07:40:32 +0000 UTC (now=2025-12-01 08:40:38.066301217 +0000 UTC))\\\\\\\"\\\\nI1201 08:40:38.066357 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1201 08:40:38.066385 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1201 08:40:38.066402 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1201 08:40:38.066414 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-63890811/tls.crt::/tmp/serving-cert-63890811/tls.key\\\\\\\"\\\\nI1201 08:40:38.066419 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1201 08:40:38.066572 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1201 08:40:38.067159 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7354f9337d42115bb1d974bf3772f2d03365acba57f2155135c7a0bbab39879e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:19Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05492df5703392cdf0932d6bf66025b18e53bd05fa0857a7cb2ad70979cbdc3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05492df5703392cdf0932d6bf66025b18e53bd05fa0857a7cb2ad70979cbdc3e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:16Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:55Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:55 crc kubenswrapper[4813]: I1201 08:40:55.732387 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3264b3af996af452df2953c59bea6deeb5ea6af1b119ce89f7334595c27e70da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9e879b1f563bd66383d58fd7b435a8a81bd50cb387dc1fabb6fc47d5a3fa468\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:55Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:55 crc kubenswrapper[4813]: I1201 08:40:55.744521 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-b6qqv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9656a03-0096-4c89-b0da-e37e4103611b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4195aebc1b2e388d480920a7f72257f6660afc99ad8c1a09478ef445739b272\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cjvnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:41Z\\\"}}\" for pod \"openshift-multus\"/\"multus-b6qqv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:55Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:55 crc kubenswrapper[4813]: I1201 08:40:55.754181 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-9wn4q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3d423402-e08f-42e9-b60c-6eb9bf2f6d48\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3049b842978c6580e1033a442a37766a2e76e278aface9a89fc0b2c57faf0df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rzxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c915733072b32a63237e2401cbf84b059aca396c2c0db884f0f439d599a35d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rzxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:41Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-9wn4q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:55Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:55 crc kubenswrapper[4813]: I1201 08:40:55.763861 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-8pg2p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dd53b38d-d9dd-4f98-9d06-dd95f1a4e3a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a214d71293e4e09573c665524d880fc319dfb1c0c6d95d11d78f253865660fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnjhq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ddf9d5989e4195b3074de08f99d53b48066f53fe93ce204145b3c714996fc57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnjhq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:53Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-8pg2p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:55Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:55 crc kubenswrapper[4813]: I1201 08:40:55.783910 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:55 crc kubenswrapper[4813]: I1201 08:40:55.783934 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:55 crc kubenswrapper[4813]: I1201 08:40:55.783942 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:55 crc kubenswrapper[4813]: I1201 08:40:55.783955 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:55 crc kubenswrapper[4813]: I1201 08:40:55.783980 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:55Z","lastTransitionTime":"2025-12-01T08:40:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:55 crc kubenswrapper[4813]: I1201 08:40:55.806933 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/network-metrics-daemon-96mbr"] Dec 01 08:40:55 crc kubenswrapper[4813]: I1201 08:40:55.807491 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-96mbr" Dec 01 08:40:55 crc kubenswrapper[4813]: E1201 08:40:55.807558 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-96mbr" podUID="0789e3f7-5b08-4c04-8596-9ba29f50cdd0" Dec 01 08:40:55 crc kubenswrapper[4813]: I1201 08:40:55.822485 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-9wn4q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3d423402-e08f-42e9-b60c-6eb9bf2f6d48\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3049b842978c6580e1033a442a37766a2e76e278aface9a89fc0b2c57faf0df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rzxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c915733072b32a63237e2401cbf84b059aca396c2c0db884f0f439d599a35d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rzxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:41Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-9wn4q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:55Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:55 crc kubenswrapper[4813]: I1201 08:40:55.842195 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-8pg2p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dd53b38d-d9dd-4f98-9d06-dd95f1a4e3a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a214d71293e4e09573c665524d880fc319dfb1c0c6d95d11d78f253865660fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnjhq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ddf9d5989e4195b3074de08f99d53b48066f53fe93ce204145b3c714996fc57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnjhq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:53Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-8pg2p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:55Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:55 crc kubenswrapper[4813]: I1201 08:40:55.857377 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:55Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:55 crc kubenswrapper[4813]: I1201 08:40:55.869336 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-c576v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"17292cd0-1eac-40b5-a60e-67167c57409c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a42206c914d0493ca74788ea1a941de80920ab7b810fb95227cd9b8711e8f4b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7zwfx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:41Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-c576v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:55Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:55 crc kubenswrapper[4813]: I1201 08:40:55.886578 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:55 crc kubenswrapper[4813]: I1201 08:40:55.886617 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:55 crc kubenswrapper[4813]: I1201 08:40:55.886627 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:55 crc kubenswrapper[4813]: I1201 08:40:55.886647 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:55 crc kubenswrapper[4813]: I1201 08:40:55.886658 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:55Z","lastTransitionTime":"2025-12-01T08:40:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:55 crc kubenswrapper[4813]: I1201 08:40:55.889102 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0789e3f7-5b08-4c04-8596-9ba29f50cdd0-metrics-certs\") pod \"network-metrics-daemon-96mbr\" (UID: \"0789e3f7-5b08-4c04-8596-9ba29f50cdd0\") " pod="openshift-multus/network-metrics-daemon-96mbr" Dec 01 08:40:55 crc kubenswrapper[4813]: I1201 08:40:55.889230 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xt4fj\" (UniqueName: \"kubernetes.io/projected/0789e3f7-5b08-4c04-8596-9ba29f50cdd0-kube-api-access-xt4fj\") pod \"network-metrics-daemon-96mbr\" (UID: \"0789e3f7-5b08-4c04-8596-9ba29f50cdd0\") " pod="openshift-multus/network-metrics-daemon-96mbr" Dec 01 08:40:55 crc kubenswrapper[4813]: I1201 08:40:55.922994 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-j4pcc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b7505e14-3560-4d0d-be2c-d6a7a5953865\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ccc30a63dc1c23ff9880fb04af1f31f0656b2696d359988e23deed23989f480a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://944977e9722025088720107530d0c91fc693593f380e4b49678f28a56da180cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc32d930ab23f30103d8fa42291909cbec89ebb75c5c53cd783cd27fe1a3f09e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f19c67b1c6c80c7f3223512102810d46ede1445464d21ae3bbea83424539b2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d984c90302e57f3e5a6f3ca3613b415db04ac74a5a228582f1b52ef1d59f93e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce8963bf1e9ee66926907aaf3a03d18977f4d0a1983d7121b62fbc8e02a37e4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b11f49feda579bcb62f9680732033819ea11e7ae9937a06c0623a54c0ec106c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4b11f49feda579bcb62f9680732033819ea11e7ae9937a06c0623a54c0ec106c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T08:40:54Z\\\",\\\"message\\\":\\\" *v1.Pod event handler 3\\\\nI1201 08:40:53.558214 6038 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1201 08:40:53.559209 6038 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1201 08:40:53.559251 6038 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1201 08:40:53.559264 6038 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1201 08:40:53.559268 6038 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1201 08:40:53.559286 6038 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1201 08:40:53.559299 6038 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1201 08:40:53.559315 6038 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1201 08:40:53.559320 6038 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1201 08:40:53.559327 6038 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1201 08:40:53.559329 6038 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1201 08:40:53.559336 6038 handler.go:208] Removed *v1.Node event handler 7\\\\nI1201 08:40:53.559354 6038 factory.go:656] Stopping watch factory\\\\nI1201 08:40:53.559365 6038 handler.go:208] Removed *v1.Node ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b45454b440c5108fa150a6b856afa1778ba6e9f89633433cb3117b68f00f72c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd67634e0c806323334495401a4e20b48fea4bc5013acd973d2abd3d70870f99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd67634e0c806323334495401a4e20b48fea4bc5013acd973d2abd3d70870f99\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:41Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-j4pcc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:55Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:55 crc kubenswrapper[4813]: I1201 08:40:55.942860 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-h7lvd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d431b613-a91b-4fff-88b9-447804766289\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a32c51c943fcf6ada7e3f614667eb023e29d94c11128d26c1fecca37b03963b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://495593f99e92a9c28100a95e27d06ff5847008c5d996366cb474bedd9b69d5ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://495593f99e92a9c28100a95e27d06ff5847008c5d996366cb474bedd9b69d5ef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://19358f83a6c52957f959fe935a6843fdeef59905cc395e6e52a1fb4a670ac193\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://19358f83a6c52957f959fe935a6843fdeef59905cc395e6e52a1fb4a670ac193\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f089c115d916e339994879d224e3db27642b99aaabf59bcafb2979549052945\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9f089c115d916e339994879d224e3db27642b99aaabf59bcafb2979549052945\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb66f224462305c2005b31ef7bb9ed01efde8101e18518dd42d1dfb0df765bb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb66f224462305c2005b31ef7bb9ed01efde8101e18518dd42d1dfb0df765bb9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c268236c9a155c9b1f47c48d30dbd23948ceb2545592b40f356c60ea1d12e8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c268236c9a155c9b1f47c48d30dbd23948ceb2545592b40f356c60ea1d12e8a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://458aedca5f186e9aaed7e4ee29fd5247221491888b130f55c1fe48d88d565f8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://458aedca5f186e9aaed7e4ee29fd5247221491888b130f55c1fe48d88d565f8f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:41Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-h7lvd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:55Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:55 crc kubenswrapper[4813]: I1201 08:40:55.957181 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6fb5a648-ae4f-4e46-a530-007e9bf84abb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e24ebfc52dcb28368a2d436e7d9ab78f07f2aec2a4c71405a6c7d3301eee96a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33898118bbf639c8e6eae5aafa10386b95795512444ed5f535db3c663f169df2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9902c00bf7738686c93f62e1772a484d05715d74b75ee98c84c8f23ed383824b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3657b47ec9f1563c83bfbe3d728c8fd123f7078e219d251b1874269e42c31eea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:16Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:55Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:55 crc kubenswrapper[4813]: I1201 08:40:55.970794 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0023b0e358d75aacc889ce364846dda89ea513832c87fa4b066aea7dac57e8a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:55Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:55 crc kubenswrapper[4813]: I1201 08:40:55.983290 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:55Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:55 crc kubenswrapper[4813]: I1201 08:40:55.988744 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:55 crc kubenswrapper[4813]: I1201 08:40:55.988778 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:55 crc kubenswrapper[4813]: I1201 08:40:55.988788 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:55 crc kubenswrapper[4813]: I1201 08:40:55.988802 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:55 crc kubenswrapper[4813]: I1201 08:40:55.988811 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:55Z","lastTransitionTime":"2025-12-01T08:40:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:55 crc kubenswrapper[4813]: I1201 08:40:55.990330 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0789e3f7-5b08-4c04-8596-9ba29f50cdd0-metrics-certs\") pod \"network-metrics-daemon-96mbr\" (UID: \"0789e3f7-5b08-4c04-8596-9ba29f50cdd0\") " pod="openshift-multus/network-metrics-daemon-96mbr" Dec 01 08:40:55 crc kubenswrapper[4813]: I1201 08:40:55.990399 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xt4fj\" (UniqueName: \"kubernetes.io/projected/0789e3f7-5b08-4c04-8596-9ba29f50cdd0-kube-api-access-xt4fj\") pod \"network-metrics-daemon-96mbr\" (UID: \"0789e3f7-5b08-4c04-8596-9ba29f50cdd0\") " pod="openshift-multus/network-metrics-daemon-96mbr" Dec 01 08:40:55 crc kubenswrapper[4813]: E1201 08:40:55.990676 4813 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 01 08:40:55 crc kubenswrapper[4813]: E1201 08:40:55.990716 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/0789e3f7-5b08-4c04-8596-9ba29f50cdd0-metrics-certs podName:0789e3f7-5b08-4c04-8596-9ba29f50cdd0 nodeName:}" failed. No retries permitted until 2025-12-01 08:40:56.490704288 +0000 UTC m=+41.113625864 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/0789e3f7-5b08-4c04-8596-9ba29f50cdd0-metrics-certs") pod "network-metrics-daemon-96mbr" (UID: "0789e3f7-5b08-4c04-8596-9ba29f50cdd0") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 01 08:40:55 crc kubenswrapper[4813]: I1201 08:40:55.994763 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c7574f892d5e9142296b26a0b484eb6759ab53f571139fb6e23b93dc3da05780\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:55Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:56 crc kubenswrapper[4813]: I1201 08:40:56.006591 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:56Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:56 crc kubenswrapper[4813]: I1201 08:40:56.008197 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xt4fj\" (UniqueName: \"kubernetes.io/projected/0789e3f7-5b08-4c04-8596-9ba29f50cdd0-kube-api-access-xt4fj\") pod \"network-metrics-daemon-96mbr\" (UID: \"0789e3f7-5b08-4c04-8596-9ba29f50cdd0\") " pod="openshift-multus/network-metrics-daemon-96mbr" Dec 01 08:40:56 crc kubenswrapper[4813]: I1201 08:40:56.022510 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-xpnlc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"22da1fcf-0c57-4708-90d8-77692af17344\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4a36d213bdad2da91d0d33f400d5a8248befb68d5b94d19347873a93c6712a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xc44d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-xpnlc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:56Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:56 crc kubenswrapper[4813]: I1201 08:40:56.034583 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"349332bd-6b4d-4306-84ae-07019ca3860f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d441a348087ae3cadda1ad8a7d1c4dff02cf3f6e1d346c294b278fdd9131edcc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://da67635daf6d9e32937f4d5d4ceeee3f8f09ae4fd2dd26c3e85be1286c5949de\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e97356ee274ea78d9f237ffbc756abb6ac8df0dbac244bf0a86ed215b8cbf43\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8dc6feb9817f4f356b8e36af43ec3070774f312fbbb4909fa05b8b3e8dd53582\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a0404f5a5ed5a9750d4c479b608e68e8960b448f455d5af9d219aa5edaeb789b\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"message\\\":\\\"\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764578421\\\\\\\\\\\\\\\" (2025-12-01 08:40:21 +0000 UTC to 2025-12-31 08:40:22 +0000 UTC (now=2025-12-01 08:40:38.066068451 +0000 UTC))\\\\\\\"\\\\nI1201 08:40:38.066144 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1201 08:40:38.066166 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1201 08:40:38.066220 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1201 08:40:38.066232 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1201 08:40:38.066338 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764578432\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764578432\\\\\\\\\\\\\\\" (2025-12-01 07:40:32 +0000 UTC to 2026-12-01 07:40:32 +0000 UTC (now=2025-12-01 08:40:38.066301217 +0000 UTC))\\\\\\\"\\\\nI1201 08:40:38.066357 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1201 08:40:38.066385 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1201 08:40:38.066402 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1201 08:40:38.066414 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-63890811/tls.crt::/tmp/serving-cert-63890811/tls.key\\\\\\\"\\\\nI1201 08:40:38.066419 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1201 08:40:38.066572 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1201 08:40:38.067159 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7354f9337d42115bb1d974bf3772f2d03365acba57f2155135c7a0bbab39879e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:19Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05492df5703392cdf0932d6bf66025b18e53bd05fa0857a7cb2ad70979cbdc3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05492df5703392cdf0932d6bf66025b18e53bd05fa0857a7cb2ad70979cbdc3e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:16Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:56Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:56 crc kubenswrapper[4813]: I1201 08:40:56.046442 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3264b3af996af452df2953c59bea6deeb5ea6af1b119ce89f7334595c27e70da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9e879b1f563bd66383d58fd7b435a8a81bd50cb387dc1fabb6fc47d5a3fa468\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:56Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:56 crc kubenswrapper[4813]: I1201 08:40:56.057298 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-b6qqv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9656a03-0096-4c89-b0da-e37e4103611b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4195aebc1b2e388d480920a7f72257f6660afc99ad8c1a09478ef445739b272\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cjvnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:41Z\\\"}}\" for pod \"openshift-multus\"/\"multus-b6qqv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:56Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:56 crc kubenswrapper[4813]: I1201 08:40:56.066246 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-96mbr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0789e3f7-5b08-4c04-8596-9ba29f50cdd0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xt4fj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xt4fj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:55Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-96mbr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:56Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:56 crc kubenswrapper[4813]: I1201 08:40:56.091854 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:56 crc kubenswrapper[4813]: I1201 08:40:56.091893 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:56 crc kubenswrapper[4813]: I1201 08:40:56.091905 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:56 crc kubenswrapper[4813]: I1201 08:40:56.091922 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:56 crc kubenswrapper[4813]: I1201 08:40:56.091934 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:56Z","lastTransitionTime":"2025-12-01T08:40:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:56 crc kubenswrapper[4813]: I1201 08:40:56.194913 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:56 crc kubenswrapper[4813]: I1201 08:40:56.194952 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:56 crc kubenswrapper[4813]: I1201 08:40:56.194981 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:56 crc kubenswrapper[4813]: I1201 08:40:56.194999 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:56 crc kubenswrapper[4813]: I1201 08:40:56.195011 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:56Z","lastTransitionTime":"2025-12-01T08:40:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:56 crc kubenswrapper[4813]: I1201 08:40:56.297183 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:56 crc kubenswrapper[4813]: I1201 08:40:56.297218 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:56 crc kubenswrapper[4813]: I1201 08:40:56.297229 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:56 crc kubenswrapper[4813]: I1201 08:40:56.297242 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:56 crc kubenswrapper[4813]: I1201 08:40:56.297251 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:56Z","lastTransitionTime":"2025-12-01T08:40:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:56 crc kubenswrapper[4813]: I1201 08:40:56.378740 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-j4pcc_b7505e14-3560-4d0d-be2c-d6a7a5953865/ovnkube-controller/0.log" Dec 01 08:40:56 crc kubenswrapper[4813]: I1201 08:40:56.382589 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-j4pcc" event={"ID":"b7505e14-3560-4d0d-be2c-d6a7a5953865","Type":"ContainerStarted","Data":"8790f4730881303e7743498b66fc17b4b0df5b6620c8ca640ad7b9c90e6898d2"} Dec 01 08:40:56 crc kubenswrapper[4813]: I1201 08:40:56.382815 4813 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 01 08:40:56 crc kubenswrapper[4813]: I1201 08:40:56.394179 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 08:40:56 crc kubenswrapper[4813]: I1201 08:40:56.394270 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 08:40:56 crc kubenswrapper[4813]: E1201 08:40:56.394343 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 08:40:56 crc kubenswrapper[4813]: E1201 08:40:56.394412 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 08:40:56 crc kubenswrapper[4813]: I1201 08:40:56.397730 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6fb5a648-ae4f-4e46-a530-007e9bf84abb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e24ebfc52dcb28368a2d436e7d9ab78f07f2aec2a4c71405a6c7d3301eee96a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33898118bbf639c8e6eae5aafa10386b95795512444ed5f535db3c663f169df2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9902c00bf7738686c93f62e1772a484d05715d74b75ee98c84c8f23ed383824b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3657b47ec9f1563c83bfbe3d728c8fd123f7078e219d251b1874269e42c31eea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:16Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:56Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:56 crc kubenswrapper[4813]: I1201 08:40:56.398821 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:56 crc kubenswrapper[4813]: I1201 08:40:56.398854 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:56 crc kubenswrapper[4813]: I1201 08:40:56.398865 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:56 crc kubenswrapper[4813]: I1201 08:40:56.398878 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:56 crc kubenswrapper[4813]: I1201 08:40:56.398887 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:56Z","lastTransitionTime":"2025-12-01T08:40:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:56 crc kubenswrapper[4813]: I1201 08:40:56.417389 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0023b0e358d75aacc889ce364846dda89ea513832c87fa4b066aea7dac57e8a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:56Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:56 crc kubenswrapper[4813]: I1201 08:40:56.428755 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:56Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:56 crc kubenswrapper[4813]: I1201 08:40:56.439556 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c7574f892d5e9142296b26a0b484eb6759ab53f571139fb6e23b93dc3da05780\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:56Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:56 crc kubenswrapper[4813]: I1201 08:40:56.453419 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:56Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:56 crc kubenswrapper[4813]: I1201 08:40:56.462782 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-xpnlc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"22da1fcf-0c57-4708-90d8-77692af17344\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4a36d213bdad2da91d0d33f400d5a8248befb68d5b94d19347873a93c6712a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xc44d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-xpnlc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:56Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:56 crc kubenswrapper[4813]: I1201 08:40:56.474304 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"349332bd-6b4d-4306-84ae-07019ca3860f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d441a348087ae3cadda1ad8a7d1c4dff02cf3f6e1d346c294b278fdd9131edcc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://da67635daf6d9e32937f4d5d4ceeee3f8f09ae4fd2dd26c3e85be1286c5949de\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e97356ee274ea78d9f237ffbc756abb6ac8df0dbac244bf0a86ed215b8cbf43\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8dc6feb9817f4f356b8e36af43ec3070774f312fbbb4909fa05b8b3e8dd53582\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a0404f5a5ed5a9750d4c479b608e68e8960b448f455d5af9d219aa5edaeb789b\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"message\\\":\\\"\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764578421\\\\\\\\\\\\\\\" (2025-12-01 08:40:21 +0000 UTC to 2025-12-31 08:40:22 +0000 UTC (now=2025-12-01 08:40:38.066068451 +0000 UTC))\\\\\\\"\\\\nI1201 08:40:38.066144 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1201 08:40:38.066166 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1201 08:40:38.066220 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1201 08:40:38.066232 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1201 08:40:38.066338 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764578432\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764578432\\\\\\\\\\\\\\\" (2025-12-01 07:40:32 +0000 UTC to 2026-12-01 07:40:32 +0000 UTC (now=2025-12-01 08:40:38.066301217 +0000 UTC))\\\\\\\"\\\\nI1201 08:40:38.066357 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1201 08:40:38.066385 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1201 08:40:38.066402 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1201 08:40:38.066414 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-63890811/tls.crt::/tmp/serving-cert-63890811/tls.key\\\\\\\"\\\\nI1201 08:40:38.066419 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1201 08:40:38.066572 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1201 08:40:38.067159 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7354f9337d42115bb1d974bf3772f2d03365acba57f2155135c7a0bbab39879e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:19Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05492df5703392cdf0932d6bf66025b18e53bd05fa0857a7cb2ad70979cbdc3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05492df5703392cdf0932d6bf66025b18e53bd05fa0857a7cb2ad70979cbdc3e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:16Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:56Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:56 crc kubenswrapper[4813]: I1201 08:40:56.488895 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3264b3af996af452df2953c59bea6deeb5ea6af1b119ce89f7334595c27e70da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9e879b1f563bd66383d58fd7b435a8a81bd50cb387dc1fabb6fc47d5a3fa468\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:56Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:56 crc kubenswrapper[4813]: I1201 08:40:56.495726 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0789e3f7-5b08-4c04-8596-9ba29f50cdd0-metrics-certs\") pod \"network-metrics-daemon-96mbr\" (UID: \"0789e3f7-5b08-4c04-8596-9ba29f50cdd0\") " pod="openshift-multus/network-metrics-daemon-96mbr" Dec 01 08:40:56 crc kubenswrapper[4813]: E1201 08:40:56.495889 4813 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 01 08:40:56 crc kubenswrapper[4813]: E1201 08:40:56.495944 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/0789e3f7-5b08-4c04-8596-9ba29f50cdd0-metrics-certs podName:0789e3f7-5b08-4c04-8596-9ba29f50cdd0 nodeName:}" failed. No retries permitted until 2025-12-01 08:40:57.495928945 +0000 UTC m=+42.118850531 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/0789e3f7-5b08-4c04-8596-9ba29f50cdd0-metrics-certs") pod "network-metrics-daemon-96mbr" (UID: "0789e3f7-5b08-4c04-8596-9ba29f50cdd0") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 01 08:40:56 crc kubenswrapper[4813]: I1201 08:40:56.500914 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:56 crc kubenswrapper[4813]: I1201 08:40:56.501082 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:56 crc kubenswrapper[4813]: I1201 08:40:56.501115 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:56 crc kubenswrapper[4813]: I1201 08:40:56.501142 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:56 crc kubenswrapper[4813]: I1201 08:40:56.501156 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:56Z","lastTransitionTime":"2025-12-01T08:40:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:56 crc kubenswrapper[4813]: I1201 08:40:56.503672 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-b6qqv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9656a03-0096-4c89-b0da-e37e4103611b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4195aebc1b2e388d480920a7f72257f6660afc99ad8c1a09478ef445739b272\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cjvnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:41Z\\\"}}\" for pod \"openshift-multus\"/\"multus-b6qqv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:56Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:56 crc kubenswrapper[4813]: I1201 08:40:56.514302 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-96mbr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0789e3f7-5b08-4c04-8596-9ba29f50cdd0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xt4fj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xt4fj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:55Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-96mbr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:56Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:56 crc kubenswrapper[4813]: I1201 08:40:56.524659 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-9wn4q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3d423402-e08f-42e9-b60c-6eb9bf2f6d48\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3049b842978c6580e1033a442a37766a2e76e278aface9a89fc0b2c57faf0df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rzxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c915733072b32a63237e2401cbf84b059aca396c2c0db884f0f439d599a35d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rzxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:41Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-9wn4q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:56Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:56 crc kubenswrapper[4813]: I1201 08:40:56.541879 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-8pg2p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dd53b38d-d9dd-4f98-9d06-dd95f1a4e3a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a214d71293e4e09573c665524d880fc319dfb1c0c6d95d11d78f253865660fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnjhq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ddf9d5989e4195b3074de08f99d53b48066f53fe93ce204145b3c714996fc57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnjhq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:53Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-8pg2p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:56Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:56 crc kubenswrapper[4813]: I1201 08:40:56.554697 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:56Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:56 crc kubenswrapper[4813]: I1201 08:40:56.564288 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-c576v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"17292cd0-1eac-40b5-a60e-67167c57409c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a42206c914d0493ca74788ea1a941de80920ab7b810fb95227cd9b8711e8f4b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7zwfx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:41Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-c576v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:56Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:56 crc kubenswrapper[4813]: I1201 08:40:56.581634 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-j4pcc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b7505e14-3560-4d0d-be2c-d6a7a5953865\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ccc30a63dc1c23ff9880fb04af1f31f0656b2696d359988e23deed23989f480a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://944977e9722025088720107530d0c91fc693593f380e4b49678f28a56da180cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc32d930ab23f30103d8fa42291909cbec89ebb75c5c53cd783cd27fe1a3f09e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f19c67b1c6c80c7f3223512102810d46ede1445464d21ae3bbea83424539b2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d984c90302e57f3e5a6f3ca3613b415db04ac74a5a228582f1b52ef1d59f93e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce8963bf1e9ee66926907aaf3a03d18977f4d0a1983d7121b62fbc8e02a37e4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8790f4730881303e7743498b66fc17b4b0df5b6620c8ca640ad7b9c90e6898d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4b11f49feda579bcb62f9680732033819ea11e7ae9937a06c0623a54c0ec106c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T08:40:54Z\\\",\\\"message\\\":\\\" *v1.Pod event handler 3\\\\nI1201 08:40:53.558214 6038 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1201 08:40:53.559209 6038 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1201 08:40:53.559251 6038 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1201 08:40:53.559264 6038 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1201 08:40:53.559268 6038 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1201 08:40:53.559286 6038 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1201 08:40:53.559299 6038 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1201 08:40:53.559315 6038 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1201 08:40:53.559320 6038 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1201 08:40:53.559327 6038 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1201 08:40:53.559329 6038 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1201 08:40:53.559336 6038 handler.go:208] Removed *v1.Node event handler 7\\\\nI1201 08:40:53.559354 6038 factory.go:656] Stopping watch factory\\\\nI1201 08:40:53.559365 6038 handler.go:208] Removed *v1.Node ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:48Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b45454b440c5108fa150a6b856afa1778ba6e9f89633433cb3117b68f00f72c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd67634e0c806323334495401a4e20b48fea4bc5013acd973d2abd3d70870f99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd67634e0c806323334495401a4e20b48fea4bc5013acd973d2abd3d70870f99\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:41Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-j4pcc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:56Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:56 crc kubenswrapper[4813]: I1201 08:40:56.596127 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-h7lvd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d431b613-a91b-4fff-88b9-447804766289\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a32c51c943fcf6ada7e3f614667eb023e29d94c11128d26c1fecca37b03963b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://495593f99e92a9c28100a95e27d06ff5847008c5d996366cb474bedd9b69d5ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://495593f99e92a9c28100a95e27d06ff5847008c5d996366cb474bedd9b69d5ef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://19358f83a6c52957f959fe935a6843fdeef59905cc395e6e52a1fb4a670ac193\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://19358f83a6c52957f959fe935a6843fdeef59905cc395e6e52a1fb4a670ac193\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f089c115d916e339994879d224e3db27642b99aaabf59bcafb2979549052945\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9f089c115d916e339994879d224e3db27642b99aaabf59bcafb2979549052945\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb66f224462305c2005b31ef7bb9ed01efde8101e18518dd42d1dfb0df765bb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb66f224462305c2005b31ef7bb9ed01efde8101e18518dd42d1dfb0df765bb9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c268236c9a155c9b1f47c48d30dbd23948ceb2545592b40f356c60ea1d12e8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c268236c9a155c9b1f47c48d30dbd23948ceb2545592b40f356c60ea1d12e8a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://458aedca5f186e9aaed7e4ee29fd5247221491888b130f55c1fe48d88d565f8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://458aedca5f186e9aaed7e4ee29fd5247221491888b130f55c1fe48d88d565f8f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:41Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-h7lvd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:56Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:56 crc kubenswrapper[4813]: I1201 08:40:56.604859 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:56 crc kubenswrapper[4813]: I1201 08:40:56.604896 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:56 crc kubenswrapper[4813]: I1201 08:40:56.604908 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:56 crc kubenswrapper[4813]: I1201 08:40:56.604924 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:56 crc kubenswrapper[4813]: I1201 08:40:56.604937 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:56Z","lastTransitionTime":"2025-12-01T08:40:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:56 crc kubenswrapper[4813]: I1201 08:40:56.616833 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-j4pcc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b7505e14-3560-4d0d-be2c-d6a7a5953865\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ccc30a63dc1c23ff9880fb04af1f31f0656b2696d359988e23deed23989f480a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://944977e9722025088720107530d0c91fc693593f380e4b49678f28a56da180cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc32d930ab23f30103d8fa42291909cbec89ebb75c5c53cd783cd27fe1a3f09e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f19c67b1c6c80c7f3223512102810d46ede1445464d21ae3bbea83424539b2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d984c90302e57f3e5a6f3ca3613b415db04ac74a5a228582f1b52ef1d59f93e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce8963bf1e9ee66926907aaf3a03d18977f4d0a1983d7121b62fbc8e02a37e4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8790f4730881303e7743498b66fc17b4b0df5b6620c8ca640ad7b9c90e6898d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4b11f49feda579bcb62f9680732033819ea11e7ae9937a06c0623a54c0ec106c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T08:40:54Z\\\",\\\"message\\\":\\\" *v1.Pod event handler 3\\\\nI1201 08:40:53.558214 6038 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1201 08:40:53.559209 6038 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1201 08:40:53.559251 6038 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1201 08:40:53.559264 6038 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1201 08:40:53.559268 6038 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1201 08:40:53.559286 6038 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1201 08:40:53.559299 6038 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1201 08:40:53.559315 6038 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1201 08:40:53.559320 6038 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1201 08:40:53.559327 6038 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1201 08:40:53.559329 6038 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1201 08:40:53.559336 6038 handler.go:208] Removed *v1.Node event handler 7\\\\nI1201 08:40:53.559354 6038 factory.go:656] Stopping watch factory\\\\nI1201 08:40:53.559365 6038 handler.go:208] Removed *v1.Node ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:48Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b45454b440c5108fa150a6b856afa1778ba6e9f89633433cb3117b68f00f72c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd67634e0c806323334495401a4e20b48fea4bc5013acd973d2abd3d70870f99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd67634e0c806323334495401a4e20b48fea4bc5013acd973d2abd3d70870f99\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:41Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-j4pcc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:56Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:56 crc kubenswrapper[4813]: I1201 08:40:56.629730 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-h7lvd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d431b613-a91b-4fff-88b9-447804766289\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a32c51c943fcf6ada7e3f614667eb023e29d94c11128d26c1fecca37b03963b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://495593f99e92a9c28100a95e27d06ff5847008c5d996366cb474bedd9b69d5ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://495593f99e92a9c28100a95e27d06ff5847008c5d996366cb474bedd9b69d5ef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://19358f83a6c52957f959fe935a6843fdeef59905cc395e6e52a1fb4a670ac193\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://19358f83a6c52957f959fe935a6843fdeef59905cc395e6e52a1fb4a670ac193\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f089c115d916e339994879d224e3db27642b99aaabf59bcafb2979549052945\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9f089c115d916e339994879d224e3db27642b99aaabf59bcafb2979549052945\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb66f224462305c2005b31ef7bb9ed01efde8101e18518dd42d1dfb0df765bb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb66f224462305c2005b31ef7bb9ed01efde8101e18518dd42d1dfb0df765bb9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c268236c9a155c9b1f47c48d30dbd23948ceb2545592b40f356c60ea1d12e8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c268236c9a155c9b1f47c48d30dbd23948ceb2545592b40f356c60ea1d12e8a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://458aedca5f186e9aaed7e4ee29fd5247221491888b130f55c1fe48d88d565f8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://458aedca5f186e9aaed7e4ee29fd5247221491888b130f55c1fe48d88d565f8f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:41Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-h7lvd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:56Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:56 crc kubenswrapper[4813]: I1201 08:40:56.640899 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:56Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:56 crc kubenswrapper[4813]: I1201 08:40:56.651742 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-c576v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"17292cd0-1eac-40b5-a60e-67167c57409c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a42206c914d0493ca74788ea1a941de80920ab7b810fb95227cd9b8711e8f4b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7zwfx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:41Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-c576v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:56Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:56 crc kubenswrapper[4813]: I1201 08:40:56.661892 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c7574f892d5e9142296b26a0b484eb6759ab53f571139fb6e23b93dc3da05780\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:56Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:56 crc kubenswrapper[4813]: I1201 08:40:56.672804 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:56Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:56 crc kubenswrapper[4813]: I1201 08:40:56.681787 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-xpnlc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"22da1fcf-0c57-4708-90d8-77692af17344\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4a36d213bdad2da91d0d33f400d5a8248befb68d5b94d19347873a93c6712a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xc44d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-xpnlc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:56Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:56 crc kubenswrapper[4813]: I1201 08:40:56.692309 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6fb5a648-ae4f-4e46-a530-007e9bf84abb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e24ebfc52dcb28368a2d436e7d9ab78f07f2aec2a4c71405a6c7d3301eee96a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33898118bbf639c8e6eae5aafa10386b95795512444ed5f535db3c663f169df2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9902c00bf7738686c93f62e1772a484d05715d74b75ee98c84c8f23ed383824b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3657b47ec9f1563c83bfbe3d728c8fd123f7078e219d251b1874269e42c31eea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:16Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:56Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:56 crc kubenswrapper[4813]: I1201 08:40:56.703155 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0023b0e358d75aacc889ce364846dda89ea513832c87fa4b066aea7dac57e8a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:56Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:56 crc kubenswrapper[4813]: I1201 08:40:56.706627 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:56 crc kubenswrapper[4813]: I1201 08:40:56.706663 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:56 crc kubenswrapper[4813]: I1201 08:40:56.706674 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:56 crc kubenswrapper[4813]: I1201 08:40:56.706689 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:56 crc kubenswrapper[4813]: I1201 08:40:56.706700 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:56Z","lastTransitionTime":"2025-12-01T08:40:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:56 crc kubenswrapper[4813]: I1201 08:40:56.714227 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:56Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:56 crc kubenswrapper[4813]: I1201 08:40:56.727827 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-b6qqv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9656a03-0096-4c89-b0da-e37e4103611b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4195aebc1b2e388d480920a7f72257f6660afc99ad8c1a09478ef445739b272\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cjvnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:41Z\\\"}}\" for pod \"openshift-multus\"/\"multus-b6qqv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:56Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:56 crc kubenswrapper[4813]: I1201 08:40:56.742017 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-96mbr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0789e3f7-5b08-4c04-8596-9ba29f50cdd0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xt4fj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xt4fj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:55Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-96mbr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:56Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:56 crc kubenswrapper[4813]: I1201 08:40:56.755672 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"349332bd-6b4d-4306-84ae-07019ca3860f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d441a348087ae3cadda1ad8a7d1c4dff02cf3f6e1d346c294b278fdd9131edcc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://da67635daf6d9e32937f4d5d4ceeee3f8f09ae4fd2dd26c3e85be1286c5949de\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e97356ee274ea78d9f237ffbc756abb6ac8df0dbac244bf0a86ed215b8cbf43\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8dc6feb9817f4f356b8e36af43ec3070774f312fbbb4909fa05b8b3e8dd53582\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a0404f5a5ed5a9750d4c479b608e68e8960b448f455d5af9d219aa5edaeb789b\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"message\\\":\\\"\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764578421\\\\\\\\\\\\\\\" (2025-12-01 08:40:21 +0000 UTC to 2025-12-31 08:40:22 +0000 UTC (now=2025-12-01 08:40:38.066068451 +0000 UTC))\\\\\\\"\\\\nI1201 08:40:38.066144 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1201 08:40:38.066166 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1201 08:40:38.066220 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1201 08:40:38.066232 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1201 08:40:38.066338 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764578432\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764578432\\\\\\\\\\\\\\\" (2025-12-01 07:40:32 +0000 UTC to 2026-12-01 07:40:32 +0000 UTC (now=2025-12-01 08:40:38.066301217 +0000 UTC))\\\\\\\"\\\\nI1201 08:40:38.066357 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1201 08:40:38.066385 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1201 08:40:38.066402 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1201 08:40:38.066414 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-63890811/tls.crt::/tmp/serving-cert-63890811/tls.key\\\\\\\"\\\\nI1201 08:40:38.066419 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1201 08:40:38.066572 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1201 08:40:38.067159 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7354f9337d42115bb1d974bf3772f2d03365acba57f2155135c7a0bbab39879e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:19Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05492df5703392cdf0932d6bf66025b18e53bd05fa0857a7cb2ad70979cbdc3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05492df5703392cdf0932d6bf66025b18e53bd05fa0857a7cb2ad70979cbdc3e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:16Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:56Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:56 crc kubenswrapper[4813]: I1201 08:40:56.771475 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3264b3af996af452df2953c59bea6deeb5ea6af1b119ce89f7334595c27e70da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9e879b1f563bd66383d58fd7b435a8a81bd50cb387dc1fabb6fc47d5a3fa468\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:56Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:56 crc kubenswrapper[4813]: I1201 08:40:56.784261 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-9wn4q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3d423402-e08f-42e9-b60c-6eb9bf2f6d48\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3049b842978c6580e1033a442a37766a2e76e278aface9a89fc0b2c57faf0df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rzxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c915733072b32a63237e2401cbf84b059aca396c2c0db884f0f439d599a35d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rzxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:41Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-9wn4q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:56Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:56 crc kubenswrapper[4813]: I1201 08:40:56.795224 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-8pg2p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dd53b38d-d9dd-4f98-9d06-dd95f1a4e3a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a214d71293e4e09573c665524d880fc319dfb1c0c6d95d11d78f253865660fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnjhq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ddf9d5989e4195b3074de08f99d53b48066f53fe93ce204145b3c714996fc57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnjhq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:53Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-8pg2p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:56Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:56 crc kubenswrapper[4813]: I1201 08:40:56.809121 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:56 crc kubenswrapper[4813]: I1201 08:40:56.809148 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:56 crc kubenswrapper[4813]: I1201 08:40:56.809155 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:56 crc kubenswrapper[4813]: I1201 08:40:56.809169 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:56 crc kubenswrapper[4813]: I1201 08:40:56.809178 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:56Z","lastTransitionTime":"2025-12-01T08:40:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:56 crc kubenswrapper[4813]: I1201 08:40:56.911485 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:56 crc kubenswrapper[4813]: I1201 08:40:56.911617 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:56 crc kubenswrapper[4813]: I1201 08:40:56.911649 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:56 crc kubenswrapper[4813]: I1201 08:40:56.911709 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:56 crc kubenswrapper[4813]: I1201 08:40:56.911767 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:56Z","lastTransitionTime":"2025-12-01T08:40:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:57 crc kubenswrapper[4813]: I1201 08:40:57.017305 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:57 crc kubenswrapper[4813]: I1201 08:40:57.017390 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:57 crc kubenswrapper[4813]: I1201 08:40:57.017420 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:57 crc kubenswrapper[4813]: I1201 08:40:57.017458 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:57 crc kubenswrapper[4813]: I1201 08:40:57.017493 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:57Z","lastTransitionTime":"2025-12-01T08:40:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:57 crc kubenswrapper[4813]: I1201 08:40:57.121770 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:57 crc kubenswrapper[4813]: I1201 08:40:57.121825 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:57 crc kubenswrapper[4813]: I1201 08:40:57.121840 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:57 crc kubenswrapper[4813]: I1201 08:40:57.121859 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:57 crc kubenswrapper[4813]: I1201 08:40:57.121870 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:57Z","lastTransitionTime":"2025-12-01T08:40:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:57 crc kubenswrapper[4813]: I1201 08:40:57.225099 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:57 crc kubenswrapper[4813]: I1201 08:40:57.225216 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:57 crc kubenswrapper[4813]: I1201 08:40:57.225240 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:57 crc kubenswrapper[4813]: I1201 08:40:57.225274 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:57 crc kubenswrapper[4813]: I1201 08:40:57.225300 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:57Z","lastTransitionTime":"2025-12-01T08:40:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:57 crc kubenswrapper[4813]: I1201 08:40:57.329251 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:57 crc kubenswrapper[4813]: I1201 08:40:57.329321 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:57 crc kubenswrapper[4813]: I1201 08:40:57.329338 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:57 crc kubenswrapper[4813]: I1201 08:40:57.329393 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:57 crc kubenswrapper[4813]: I1201 08:40:57.329415 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:57Z","lastTransitionTime":"2025-12-01T08:40:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:57 crc kubenswrapper[4813]: I1201 08:40:57.393260 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 08:40:57 crc kubenswrapper[4813]: I1201 08:40:57.393260 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-96mbr" Dec 01 08:40:57 crc kubenswrapper[4813]: E1201 08:40:57.394017 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-96mbr" podUID="0789e3f7-5b08-4c04-8596-9ba29f50cdd0" Dec 01 08:40:57 crc kubenswrapper[4813]: I1201 08:40:57.394658 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-j4pcc_b7505e14-3560-4d0d-be2c-d6a7a5953865/ovnkube-controller/1.log" Dec 01 08:40:57 crc kubenswrapper[4813]: E1201 08:40:57.394914 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 08:40:57 crc kubenswrapper[4813]: I1201 08:40:57.396177 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-j4pcc_b7505e14-3560-4d0d-be2c-d6a7a5953865/ovnkube-controller/0.log" Dec 01 08:40:57 crc kubenswrapper[4813]: I1201 08:40:57.403152 4813 generic.go:334] "Generic (PLEG): container finished" podID="b7505e14-3560-4d0d-be2c-d6a7a5953865" containerID="8790f4730881303e7743498b66fc17b4b0df5b6620c8ca640ad7b9c90e6898d2" exitCode=1 Dec 01 08:40:57 crc kubenswrapper[4813]: I1201 08:40:57.403412 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-j4pcc" event={"ID":"b7505e14-3560-4d0d-be2c-d6a7a5953865","Type":"ContainerDied","Data":"8790f4730881303e7743498b66fc17b4b0df5b6620c8ca640ad7b9c90e6898d2"} Dec 01 08:40:57 crc kubenswrapper[4813]: I1201 08:40:57.403960 4813 scope.go:117] "RemoveContainer" containerID="4b11f49feda579bcb62f9680732033819ea11e7ae9937a06c0623a54c0ec106c" Dec 01 08:40:57 crc kubenswrapper[4813]: I1201 08:40:57.405743 4813 scope.go:117] "RemoveContainer" containerID="8790f4730881303e7743498b66fc17b4b0df5b6620c8ca640ad7b9c90e6898d2" Dec 01 08:40:57 crc kubenswrapper[4813]: E1201 08:40:57.406232 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-j4pcc_openshift-ovn-kubernetes(b7505e14-3560-4d0d-be2c-d6a7a5953865)\"" pod="openshift-ovn-kubernetes/ovnkube-node-j4pcc" podUID="b7505e14-3560-4d0d-be2c-d6a7a5953865" Dec 01 08:40:57 crc kubenswrapper[4813]: I1201 08:40:57.602257 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0789e3f7-5b08-4c04-8596-9ba29f50cdd0-metrics-certs\") pod \"network-metrics-daemon-96mbr\" (UID: \"0789e3f7-5b08-4c04-8596-9ba29f50cdd0\") " pod="openshift-multus/network-metrics-daemon-96mbr" Dec 01 08:40:57 crc kubenswrapper[4813]: E1201 08:40:57.604038 4813 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 01 08:40:57 crc kubenswrapper[4813]: E1201 08:40:57.604327 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/0789e3f7-5b08-4c04-8596-9ba29f50cdd0-metrics-certs podName:0789e3f7-5b08-4c04-8596-9ba29f50cdd0 nodeName:}" failed. No retries permitted until 2025-12-01 08:40:59.604219446 +0000 UTC m=+44.227141102 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/0789e3f7-5b08-4c04-8596-9ba29f50cdd0-metrics-certs") pod "network-metrics-daemon-96mbr" (UID: "0789e3f7-5b08-4c04-8596-9ba29f50cdd0") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 01 08:40:57 crc kubenswrapper[4813]: I1201 08:40:57.605508 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:57 crc kubenswrapper[4813]: I1201 08:40:57.605544 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:57 crc kubenswrapper[4813]: I1201 08:40:57.605563 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:57 crc kubenswrapper[4813]: I1201 08:40:57.605622 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:57 crc kubenswrapper[4813]: I1201 08:40:57.605646 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:57Z","lastTransitionTime":"2025-12-01T08:40:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:57 crc kubenswrapper[4813]: I1201 08:40:57.614172 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-9wn4q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3d423402-e08f-42e9-b60c-6eb9bf2f6d48\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3049b842978c6580e1033a442a37766a2e76e278aface9a89fc0b2c57faf0df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rzxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c915733072b32a63237e2401cbf84b059aca396c2c0db884f0f439d599a35d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rzxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:41Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-9wn4q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:57Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:57 crc kubenswrapper[4813]: I1201 08:40:57.627108 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-8pg2p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dd53b38d-d9dd-4f98-9d06-dd95f1a4e3a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a214d71293e4e09573c665524d880fc319dfb1c0c6d95d11d78f253865660fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnjhq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ddf9d5989e4195b3074de08f99d53b48066f53fe93ce204145b3c714996fc57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnjhq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:53Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-8pg2p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:57Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:57 crc kubenswrapper[4813]: I1201 08:40:57.639428 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:57Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:57 crc kubenswrapper[4813]: I1201 08:40:57.650420 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-c576v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"17292cd0-1eac-40b5-a60e-67167c57409c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a42206c914d0493ca74788ea1a941de80920ab7b810fb95227cd9b8711e8f4b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7zwfx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:41Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-c576v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:57Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:57 crc kubenswrapper[4813]: I1201 08:40:57.667392 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-j4pcc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b7505e14-3560-4d0d-be2c-d6a7a5953865\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ccc30a63dc1c23ff9880fb04af1f31f0656b2696d359988e23deed23989f480a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://944977e9722025088720107530d0c91fc693593f380e4b49678f28a56da180cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc32d930ab23f30103d8fa42291909cbec89ebb75c5c53cd783cd27fe1a3f09e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f19c67b1c6c80c7f3223512102810d46ede1445464d21ae3bbea83424539b2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d984c90302e57f3e5a6f3ca3613b415db04ac74a5a228582f1b52ef1d59f93e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce8963bf1e9ee66926907aaf3a03d18977f4d0a1983d7121b62fbc8e02a37e4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8790f4730881303e7743498b66fc17b4b0df5b6620c8ca640ad7b9c90e6898d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4b11f49feda579bcb62f9680732033819ea11e7ae9937a06c0623a54c0ec106c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T08:40:54Z\\\",\\\"message\\\":\\\" *v1.Pod event handler 3\\\\nI1201 08:40:53.558214 6038 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1201 08:40:53.559209 6038 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1201 08:40:53.559251 6038 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1201 08:40:53.559264 6038 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1201 08:40:53.559268 6038 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1201 08:40:53.559286 6038 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1201 08:40:53.559299 6038 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1201 08:40:53.559315 6038 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1201 08:40:53.559320 6038 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1201 08:40:53.559327 6038 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1201 08:40:53.559329 6038 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1201 08:40:53.559336 6038 handler.go:208] Removed *v1.Node event handler 7\\\\nI1201 08:40:53.559354 6038 factory.go:656] Stopping watch factory\\\\nI1201 08:40:53.559365 6038 handler.go:208] Removed *v1.Node ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:48Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8790f4730881303e7743498b66fc17b4b0df5b6620c8ca640ad7b9c90e6898d2\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T08:40:56Z\\\",\\\"message\\\":\\\"I1201 08:40:56.782175 6307 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1201 08:40:56.782219 6307 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1201 08:40:56.782251 6307 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1201 08:40:56.782302 6307 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1201 08:40:56.782342 6307 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1201 08:40:56.782379 6307 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1201 08:40:56.782414 6307 factory.go:656] Stopping watch factory\\\\nI1201 08:40:56.782444 6307 ovnkube.go:599] Stopped ovnkube\\\\nI1201 08:40:56.782487 6307 handler.go:208] Removed *v1.Node event handler 7\\\\nI1201 08:40:56.782514 6307 handler.go:208] Removed *v1.Node event handler 2\\\\nI1201 08:40:56.782539 6307 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1201 08:40:56.782564 6307 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1201 08:40:56.782591 6307 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1201 08:40:56.782650 6307 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1201 08:40:56.782670 6307 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1201 08:40:56.782663 6307 handler.go:208] Removed *v1.EgressFirewa\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b45454b440c5108fa150a6b856afa1778ba6e9f89633433cb3117b68f00f72c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd67634e0c806323334495401a4e20b48fea4bc5013acd973d2abd3d70870f99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd67634e0c806323334495401a4e20b48fea4bc5013acd973d2abd3d70870f99\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:41Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-j4pcc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:57Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:57 crc kubenswrapper[4813]: I1201 08:40:57.683002 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-h7lvd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d431b613-a91b-4fff-88b9-447804766289\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a32c51c943fcf6ada7e3f614667eb023e29d94c11128d26c1fecca37b03963b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://495593f99e92a9c28100a95e27d06ff5847008c5d996366cb474bedd9b69d5ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://495593f99e92a9c28100a95e27d06ff5847008c5d996366cb474bedd9b69d5ef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://19358f83a6c52957f959fe935a6843fdeef59905cc395e6e52a1fb4a670ac193\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://19358f83a6c52957f959fe935a6843fdeef59905cc395e6e52a1fb4a670ac193\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f089c115d916e339994879d224e3db27642b99aaabf59bcafb2979549052945\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9f089c115d916e339994879d224e3db27642b99aaabf59bcafb2979549052945\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb66f224462305c2005b31ef7bb9ed01efde8101e18518dd42d1dfb0df765bb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb66f224462305c2005b31ef7bb9ed01efde8101e18518dd42d1dfb0df765bb9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c268236c9a155c9b1f47c48d30dbd23948ceb2545592b40f356c60ea1d12e8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c268236c9a155c9b1f47c48d30dbd23948ceb2545592b40f356c60ea1d12e8a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://458aedca5f186e9aaed7e4ee29fd5247221491888b130f55c1fe48d88d565f8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://458aedca5f186e9aaed7e4ee29fd5247221491888b130f55c1fe48d88d565f8f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:41Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-h7lvd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:57Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:57 crc kubenswrapper[4813]: I1201 08:40:57.695006 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6fb5a648-ae4f-4e46-a530-007e9bf84abb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e24ebfc52dcb28368a2d436e7d9ab78f07f2aec2a4c71405a6c7d3301eee96a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33898118bbf639c8e6eae5aafa10386b95795512444ed5f535db3c663f169df2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9902c00bf7738686c93f62e1772a484d05715d74b75ee98c84c8f23ed383824b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3657b47ec9f1563c83bfbe3d728c8fd123f7078e219d251b1874269e42c31eea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:16Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:57Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:57 crc kubenswrapper[4813]: I1201 08:40:57.707755 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0023b0e358d75aacc889ce364846dda89ea513832c87fa4b066aea7dac57e8a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:57Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:57 crc kubenswrapper[4813]: I1201 08:40:57.708706 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:57 crc kubenswrapper[4813]: I1201 08:40:57.708783 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:57 crc kubenswrapper[4813]: I1201 08:40:57.708796 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:57 crc kubenswrapper[4813]: I1201 08:40:57.708815 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:57 crc kubenswrapper[4813]: I1201 08:40:57.708828 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:57Z","lastTransitionTime":"2025-12-01T08:40:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:57 crc kubenswrapper[4813]: I1201 08:40:57.720511 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:57Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:57 crc kubenswrapper[4813]: I1201 08:40:57.732803 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c7574f892d5e9142296b26a0b484eb6759ab53f571139fb6e23b93dc3da05780\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:57Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:57 crc kubenswrapper[4813]: I1201 08:40:57.745105 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:57Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:57 crc kubenswrapper[4813]: I1201 08:40:57.754843 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-xpnlc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"22da1fcf-0c57-4708-90d8-77692af17344\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4a36d213bdad2da91d0d33f400d5a8248befb68d5b94d19347873a93c6712a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xc44d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-xpnlc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:57Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:57 crc kubenswrapper[4813]: I1201 08:40:57.770814 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"349332bd-6b4d-4306-84ae-07019ca3860f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d441a348087ae3cadda1ad8a7d1c4dff02cf3f6e1d346c294b278fdd9131edcc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://da67635daf6d9e32937f4d5d4ceeee3f8f09ae4fd2dd26c3e85be1286c5949de\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e97356ee274ea78d9f237ffbc756abb6ac8df0dbac244bf0a86ed215b8cbf43\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8dc6feb9817f4f356b8e36af43ec3070774f312fbbb4909fa05b8b3e8dd53582\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a0404f5a5ed5a9750d4c479b608e68e8960b448f455d5af9d219aa5edaeb789b\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"message\\\":\\\"\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764578421\\\\\\\\\\\\\\\" (2025-12-01 08:40:21 +0000 UTC to 2025-12-31 08:40:22 +0000 UTC (now=2025-12-01 08:40:38.066068451 +0000 UTC))\\\\\\\"\\\\nI1201 08:40:38.066144 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1201 08:40:38.066166 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1201 08:40:38.066220 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1201 08:40:38.066232 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1201 08:40:38.066338 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764578432\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764578432\\\\\\\\\\\\\\\" (2025-12-01 07:40:32 +0000 UTC to 2026-12-01 07:40:32 +0000 UTC (now=2025-12-01 08:40:38.066301217 +0000 UTC))\\\\\\\"\\\\nI1201 08:40:38.066357 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1201 08:40:38.066385 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1201 08:40:38.066402 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1201 08:40:38.066414 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-63890811/tls.crt::/tmp/serving-cert-63890811/tls.key\\\\\\\"\\\\nI1201 08:40:38.066419 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1201 08:40:38.066572 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1201 08:40:38.067159 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7354f9337d42115bb1d974bf3772f2d03365acba57f2155135c7a0bbab39879e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:19Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05492df5703392cdf0932d6bf66025b18e53bd05fa0857a7cb2ad70979cbdc3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05492df5703392cdf0932d6bf66025b18e53bd05fa0857a7cb2ad70979cbdc3e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:16Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:57Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:57 crc kubenswrapper[4813]: I1201 08:40:57.784744 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3264b3af996af452df2953c59bea6deeb5ea6af1b119ce89f7334595c27e70da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9e879b1f563bd66383d58fd7b435a8a81bd50cb387dc1fabb6fc47d5a3fa468\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:57Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:57 crc kubenswrapper[4813]: I1201 08:40:57.802487 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-b6qqv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9656a03-0096-4c89-b0da-e37e4103611b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4195aebc1b2e388d480920a7f72257f6660afc99ad8c1a09478ef445739b272\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cjvnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:41Z\\\"}}\" for pod \"openshift-multus\"/\"multus-b6qqv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:57Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:57 crc kubenswrapper[4813]: I1201 08:40:57.811834 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:57 crc kubenswrapper[4813]: I1201 08:40:57.812038 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:57 crc kubenswrapper[4813]: I1201 08:40:57.812132 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:57 crc kubenswrapper[4813]: I1201 08:40:57.812220 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:57 crc kubenswrapper[4813]: I1201 08:40:57.812297 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:57Z","lastTransitionTime":"2025-12-01T08:40:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:57 crc kubenswrapper[4813]: I1201 08:40:57.815831 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-96mbr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0789e3f7-5b08-4c04-8596-9ba29f50cdd0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xt4fj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xt4fj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:55Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-96mbr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:57Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:57 crc kubenswrapper[4813]: I1201 08:40:57.914486 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:57 crc kubenswrapper[4813]: I1201 08:40:57.914798 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:57 crc kubenswrapper[4813]: I1201 08:40:57.914879 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:57 crc kubenswrapper[4813]: I1201 08:40:57.914989 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:57 crc kubenswrapper[4813]: I1201 08:40:57.915071 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:57Z","lastTransitionTime":"2025-12-01T08:40:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:58 crc kubenswrapper[4813]: I1201 08:40:58.017401 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:58 crc kubenswrapper[4813]: I1201 08:40:58.017441 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:58 crc kubenswrapper[4813]: I1201 08:40:58.017453 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:58 crc kubenswrapper[4813]: I1201 08:40:58.017470 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:58 crc kubenswrapper[4813]: I1201 08:40:58.017500 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:58Z","lastTransitionTime":"2025-12-01T08:40:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:58 crc kubenswrapper[4813]: I1201 08:40:58.120186 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:58 crc kubenswrapper[4813]: I1201 08:40:58.120243 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:58 crc kubenswrapper[4813]: I1201 08:40:58.120256 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:58 crc kubenswrapper[4813]: I1201 08:40:58.120272 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:58 crc kubenswrapper[4813]: I1201 08:40:58.120282 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:58Z","lastTransitionTime":"2025-12-01T08:40:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:58 crc kubenswrapper[4813]: I1201 08:40:58.223001 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:58 crc kubenswrapper[4813]: I1201 08:40:58.223045 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:58 crc kubenswrapper[4813]: I1201 08:40:58.223054 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:58 crc kubenswrapper[4813]: I1201 08:40:58.223072 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:58 crc kubenswrapper[4813]: I1201 08:40:58.223085 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:58Z","lastTransitionTime":"2025-12-01T08:40:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:58 crc kubenswrapper[4813]: I1201 08:40:58.325814 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:58 crc kubenswrapper[4813]: I1201 08:40:58.326087 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:58 crc kubenswrapper[4813]: I1201 08:40:58.326172 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:58 crc kubenswrapper[4813]: I1201 08:40:58.326316 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:58 crc kubenswrapper[4813]: I1201 08:40:58.329012 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:58Z","lastTransitionTime":"2025-12-01T08:40:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:58 crc kubenswrapper[4813]: I1201 08:40:58.393512 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 08:40:58 crc kubenswrapper[4813]: I1201 08:40:58.393513 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 08:40:58 crc kubenswrapper[4813]: E1201 08:40:58.393854 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 08:40:58 crc kubenswrapper[4813]: E1201 08:40:58.393998 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 08:40:58 crc kubenswrapper[4813]: I1201 08:40:58.409907 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-j4pcc_b7505e14-3560-4d0d-be2c-d6a7a5953865/ovnkube-controller/1.log" Dec 01 08:40:58 crc kubenswrapper[4813]: I1201 08:40:58.433093 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:58 crc kubenswrapper[4813]: I1201 08:40:58.433153 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:58 crc kubenswrapper[4813]: I1201 08:40:58.433168 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:58 crc kubenswrapper[4813]: I1201 08:40:58.433190 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:58 crc kubenswrapper[4813]: I1201 08:40:58.433209 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:58Z","lastTransitionTime":"2025-12-01T08:40:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:58 crc kubenswrapper[4813]: I1201 08:40:58.536060 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:58 crc kubenswrapper[4813]: I1201 08:40:58.536134 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:58 crc kubenswrapper[4813]: I1201 08:40:58.536162 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:58 crc kubenswrapper[4813]: I1201 08:40:58.536194 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:58 crc kubenswrapper[4813]: I1201 08:40:58.536216 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:58Z","lastTransitionTime":"2025-12-01T08:40:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:58 crc kubenswrapper[4813]: I1201 08:40:58.646739 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:58 crc kubenswrapper[4813]: I1201 08:40:58.646779 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:58 crc kubenswrapper[4813]: I1201 08:40:58.646788 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:58 crc kubenswrapper[4813]: I1201 08:40:58.646804 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:58 crc kubenswrapper[4813]: I1201 08:40:58.646814 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:58Z","lastTransitionTime":"2025-12-01T08:40:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:58 crc kubenswrapper[4813]: I1201 08:40:58.754055 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:58 crc kubenswrapper[4813]: I1201 08:40:58.754112 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:58 crc kubenswrapper[4813]: I1201 08:40:58.754133 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:58 crc kubenswrapper[4813]: I1201 08:40:58.754178 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:58 crc kubenswrapper[4813]: I1201 08:40:58.754197 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:58Z","lastTransitionTime":"2025-12-01T08:40:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:58 crc kubenswrapper[4813]: I1201 08:40:58.857421 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:58 crc kubenswrapper[4813]: I1201 08:40:58.857686 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:58 crc kubenswrapper[4813]: I1201 08:40:58.857880 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:58 crc kubenswrapper[4813]: I1201 08:40:58.858053 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:58 crc kubenswrapper[4813]: I1201 08:40:58.858255 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:58Z","lastTransitionTime":"2025-12-01T08:40:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:58 crc kubenswrapper[4813]: I1201 08:40:58.962256 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:58 crc kubenswrapper[4813]: I1201 08:40:58.962477 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:58 crc kubenswrapper[4813]: I1201 08:40:58.962557 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:58 crc kubenswrapper[4813]: I1201 08:40:58.962630 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:58 crc kubenswrapper[4813]: I1201 08:40:58.962696 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:58Z","lastTransitionTime":"2025-12-01T08:40:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:59 crc kubenswrapper[4813]: I1201 08:40:59.065697 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:59 crc kubenswrapper[4813]: I1201 08:40:59.066020 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:59 crc kubenswrapper[4813]: I1201 08:40:59.066120 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:59 crc kubenswrapper[4813]: I1201 08:40:59.066206 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:59 crc kubenswrapper[4813]: I1201 08:40:59.066289 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:59Z","lastTransitionTime":"2025-12-01T08:40:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:59 crc kubenswrapper[4813]: I1201 08:40:59.170536 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:59 crc kubenswrapper[4813]: I1201 08:40:59.170865 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:59 crc kubenswrapper[4813]: I1201 08:40:59.171095 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:59 crc kubenswrapper[4813]: I1201 08:40:59.171249 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:59 crc kubenswrapper[4813]: I1201 08:40:59.171387 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:59Z","lastTransitionTime":"2025-12-01T08:40:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:59 crc kubenswrapper[4813]: I1201 08:40:59.274715 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:59 crc kubenswrapper[4813]: I1201 08:40:59.275340 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:59 crc kubenswrapper[4813]: I1201 08:40:59.275550 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:59 crc kubenswrapper[4813]: I1201 08:40:59.275813 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:59 crc kubenswrapper[4813]: I1201 08:40:59.276059 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:59Z","lastTransitionTime":"2025-12-01T08:40:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:59 crc kubenswrapper[4813]: I1201 08:40:59.379509 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:59 crc kubenswrapper[4813]: I1201 08:40:59.379921 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:59 crc kubenswrapper[4813]: I1201 08:40:59.380188 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:59 crc kubenswrapper[4813]: I1201 08:40:59.380390 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:59 crc kubenswrapper[4813]: I1201 08:40:59.380626 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:59Z","lastTransitionTime":"2025-12-01T08:40:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:59 crc kubenswrapper[4813]: I1201 08:40:59.392953 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 08:40:59 crc kubenswrapper[4813]: E1201 08:40:59.393160 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 08:40:59 crc kubenswrapper[4813]: I1201 08:40:59.392954 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-96mbr" Dec 01 08:40:59 crc kubenswrapper[4813]: E1201 08:40:59.393630 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-96mbr" podUID="0789e3f7-5b08-4c04-8596-9ba29f50cdd0" Dec 01 08:40:59 crc kubenswrapper[4813]: I1201 08:40:59.484104 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:59 crc kubenswrapper[4813]: I1201 08:40:59.484168 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:59 crc kubenswrapper[4813]: I1201 08:40:59.484185 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:59 crc kubenswrapper[4813]: I1201 08:40:59.484209 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:59 crc kubenswrapper[4813]: I1201 08:40:59.484226 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:59Z","lastTransitionTime":"2025-12-01T08:40:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:59 crc kubenswrapper[4813]: I1201 08:40:59.587232 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:59 crc kubenswrapper[4813]: I1201 08:40:59.587274 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:59 crc kubenswrapper[4813]: I1201 08:40:59.587283 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:59 crc kubenswrapper[4813]: I1201 08:40:59.587299 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:59 crc kubenswrapper[4813]: I1201 08:40:59.587309 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:59Z","lastTransitionTime":"2025-12-01T08:40:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:59 crc kubenswrapper[4813]: I1201 08:40:59.650316 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0789e3f7-5b08-4c04-8596-9ba29f50cdd0-metrics-certs\") pod \"network-metrics-daemon-96mbr\" (UID: \"0789e3f7-5b08-4c04-8596-9ba29f50cdd0\") " pod="openshift-multus/network-metrics-daemon-96mbr" Dec 01 08:40:59 crc kubenswrapper[4813]: E1201 08:40:59.651111 4813 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 01 08:40:59 crc kubenswrapper[4813]: E1201 08:40:59.651176 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/0789e3f7-5b08-4c04-8596-9ba29f50cdd0-metrics-certs podName:0789e3f7-5b08-4c04-8596-9ba29f50cdd0 nodeName:}" failed. No retries permitted until 2025-12-01 08:41:03.651161754 +0000 UTC m=+48.274083340 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/0789e3f7-5b08-4c04-8596-9ba29f50cdd0-metrics-certs") pod "network-metrics-daemon-96mbr" (UID: "0789e3f7-5b08-4c04-8596-9ba29f50cdd0") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 01 08:40:59 crc kubenswrapper[4813]: I1201 08:40:59.690939 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:59 crc kubenswrapper[4813]: I1201 08:40:59.691024 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:59 crc kubenswrapper[4813]: I1201 08:40:59.691041 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:59 crc kubenswrapper[4813]: I1201 08:40:59.691067 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:59 crc kubenswrapper[4813]: I1201 08:40:59.691084 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:59Z","lastTransitionTime":"2025-12-01T08:40:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:59 crc kubenswrapper[4813]: I1201 08:40:59.793411 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:59 crc kubenswrapper[4813]: I1201 08:40:59.793460 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:59 crc kubenswrapper[4813]: I1201 08:40:59.793470 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:59 crc kubenswrapper[4813]: I1201 08:40:59.793486 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:59 crc kubenswrapper[4813]: I1201 08:40:59.793495 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:59Z","lastTransitionTime":"2025-12-01T08:40:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:59 crc kubenswrapper[4813]: I1201 08:40:59.896269 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:59 crc kubenswrapper[4813]: I1201 08:40:59.896308 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:59 crc kubenswrapper[4813]: I1201 08:40:59.896317 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:59 crc kubenswrapper[4813]: I1201 08:40:59.896332 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:59 crc kubenswrapper[4813]: I1201 08:40:59.896342 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:59Z","lastTransitionTime":"2025-12-01T08:40:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:59 crc kubenswrapper[4813]: I1201 08:40:59.952290 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:59 crc kubenswrapper[4813]: I1201 08:40:59.952340 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:59 crc kubenswrapper[4813]: I1201 08:40:59.952349 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:59 crc kubenswrapper[4813]: I1201 08:40:59.952363 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:59 crc kubenswrapper[4813]: I1201 08:40:59.952373 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:59Z","lastTransitionTime":"2025-12-01T08:40:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:40:59 crc kubenswrapper[4813]: E1201 08:40:59.969758 4813 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:40:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:40:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:59Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:40:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:40:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:59Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cd271ba1-a5c5-44e0-b03e-cfd1f5e8cb76\\\",\\\"systemUUID\\\":\\\"906e3783-eb6d-4f21-889c-7989aeaa25ba\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:40:59Z is after 2025-08-24T17:21:41Z" Dec 01 08:40:59 crc kubenswrapper[4813]: I1201 08:40:59.975713 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:40:59 crc kubenswrapper[4813]: I1201 08:40:59.976029 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:40:59 crc kubenswrapper[4813]: I1201 08:40:59.976215 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:40:59 crc kubenswrapper[4813]: I1201 08:40:59.976368 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:40:59 crc kubenswrapper[4813]: I1201 08:40:59.976505 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:40:59Z","lastTransitionTime":"2025-12-01T08:40:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:00 crc kubenswrapper[4813]: E1201 08:41:00.010189 4813 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:40:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:40:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:59Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:40:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:40:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:59Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cd271ba1-a5c5-44e0-b03e-cfd1f5e8cb76\\\",\\\"systemUUID\\\":\\\"906e3783-eb6d-4f21-889c-7989aeaa25ba\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:00Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:00 crc kubenswrapper[4813]: I1201 08:41:00.016689 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:00 crc kubenswrapper[4813]: I1201 08:41:00.017025 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:00 crc kubenswrapper[4813]: I1201 08:41:00.017216 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:00 crc kubenswrapper[4813]: I1201 08:41:00.017368 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:00 crc kubenswrapper[4813]: I1201 08:41:00.017528 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:00Z","lastTransitionTime":"2025-12-01T08:41:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:00 crc kubenswrapper[4813]: E1201 08:41:00.038128 4813 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:41:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:41:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:00Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:41:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:41:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:00Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cd271ba1-a5c5-44e0-b03e-cfd1f5e8cb76\\\",\\\"systemUUID\\\":\\\"906e3783-eb6d-4f21-889c-7989aeaa25ba\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:00Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:00 crc kubenswrapper[4813]: I1201 08:41:00.043620 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:00 crc kubenswrapper[4813]: I1201 08:41:00.043670 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:00 crc kubenswrapper[4813]: I1201 08:41:00.043689 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:00 crc kubenswrapper[4813]: I1201 08:41:00.043712 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:00 crc kubenswrapper[4813]: I1201 08:41:00.043727 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:00Z","lastTransitionTime":"2025-12-01T08:41:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:00 crc kubenswrapper[4813]: E1201 08:41:00.058518 4813 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:41:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:41:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:00Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:41:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:41:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:00Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cd271ba1-a5c5-44e0-b03e-cfd1f5e8cb76\\\",\\\"systemUUID\\\":\\\"906e3783-eb6d-4f21-889c-7989aeaa25ba\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:00Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:00 crc kubenswrapper[4813]: I1201 08:41:00.063428 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:00 crc kubenswrapper[4813]: I1201 08:41:00.063490 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:00 crc kubenswrapper[4813]: I1201 08:41:00.063502 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:00 crc kubenswrapper[4813]: I1201 08:41:00.063521 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:00 crc kubenswrapper[4813]: I1201 08:41:00.063537 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:00Z","lastTransitionTime":"2025-12-01T08:41:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:00 crc kubenswrapper[4813]: E1201 08:41:00.083239 4813 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:41:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:41:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:00Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:41:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:41:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:00Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cd271ba1-a5c5-44e0-b03e-cfd1f5e8cb76\\\",\\\"systemUUID\\\":\\\"906e3783-eb6d-4f21-889c-7989aeaa25ba\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:00Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:00 crc kubenswrapper[4813]: E1201 08:41:00.083409 4813 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 01 08:41:00 crc kubenswrapper[4813]: I1201 08:41:00.086387 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:00 crc kubenswrapper[4813]: I1201 08:41:00.086421 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:00 crc kubenswrapper[4813]: I1201 08:41:00.086436 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:00 crc kubenswrapper[4813]: I1201 08:41:00.086456 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:00 crc kubenswrapper[4813]: I1201 08:41:00.086469 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:00Z","lastTransitionTime":"2025-12-01T08:41:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:00 crc kubenswrapper[4813]: I1201 08:41:00.190266 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:00 crc kubenswrapper[4813]: I1201 08:41:00.190704 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:00 crc kubenswrapper[4813]: I1201 08:41:00.190880 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:00 crc kubenswrapper[4813]: I1201 08:41:00.191059 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:00 crc kubenswrapper[4813]: I1201 08:41:00.191198 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:00Z","lastTransitionTime":"2025-12-01T08:41:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:00 crc kubenswrapper[4813]: I1201 08:41:00.294503 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:00 crc kubenswrapper[4813]: I1201 08:41:00.294563 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:00 crc kubenswrapper[4813]: I1201 08:41:00.294577 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:00 crc kubenswrapper[4813]: I1201 08:41:00.294601 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:00 crc kubenswrapper[4813]: I1201 08:41:00.294615 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:00Z","lastTransitionTime":"2025-12-01T08:41:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:00 crc kubenswrapper[4813]: I1201 08:41:00.393208 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 08:41:00 crc kubenswrapper[4813]: E1201 08:41:00.393418 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 08:41:00 crc kubenswrapper[4813]: I1201 08:41:00.394100 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 08:41:00 crc kubenswrapper[4813]: E1201 08:41:00.394261 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 08:41:00 crc kubenswrapper[4813]: I1201 08:41:00.399833 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:00 crc kubenswrapper[4813]: I1201 08:41:00.399914 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:00 crc kubenswrapper[4813]: I1201 08:41:00.399944 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:00 crc kubenswrapper[4813]: I1201 08:41:00.400010 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:00 crc kubenswrapper[4813]: I1201 08:41:00.400037 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:00Z","lastTransitionTime":"2025-12-01T08:41:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:00 crc kubenswrapper[4813]: I1201 08:41:00.502911 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:00 crc kubenswrapper[4813]: I1201 08:41:00.502961 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:00 crc kubenswrapper[4813]: I1201 08:41:00.502998 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:00 crc kubenswrapper[4813]: I1201 08:41:00.503020 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:00 crc kubenswrapper[4813]: I1201 08:41:00.503033 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:00Z","lastTransitionTime":"2025-12-01T08:41:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:00 crc kubenswrapper[4813]: I1201 08:41:00.606252 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:00 crc kubenswrapper[4813]: I1201 08:41:00.606299 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:00 crc kubenswrapper[4813]: I1201 08:41:00.606312 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:00 crc kubenswrapper[4813]: I1201 08:41:00.606330 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:00 crc kubenswrapper[4813]: I1201 08:41:00.606342 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:00Z","lastTransitionTime":"2025-12-01T08:41:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:00 crc kubenswrapper[4813]: I1201 08:41:00.709481 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:00 crc kubenswrapper[4813]: I1201 08:41:00.709544 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:00 crc kubenswrapper[4813]: I1201 08:41:00.709561 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:00 crc kubenswrapper[4813]: I1201 08:41:00.709593 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:00 crc kubenswrapper[4813]: I1201 08:41:00.709611 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:00Z","lastTransitionTime":"2025-12-01T08:41:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:00 crc kubenswrapper[4813]: I1201 08:41:00.814092 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:00 crc kubenswrapper[4813]: I1201 08:41:00.814155 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:00 crc kubenswrapper[4813]: I1201 08:41:00.814179 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:00 crc kubenswrapper[4813]: I1201 08:41:00.814209 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:00 crc kubenswrapper[4813]: I1201 08:41:00.814227 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:00Z","lastTransitionTime":"2025-12-01T08:41:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:00 crc kubenswrapper[4813]: I1201 08:41:00.924676 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:00 crc kubenswrapper[4813]: I1201 08:41:00.924744 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:00 crc kubenswrapper[4813]: I1201 08:41:00.924768 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:00 crc kubenswrapper[4813]: I1201 08:41:00.924798 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:00 crc kubenswrapper[4813]: I1201 08:41:00.924821 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:00Z","lastTransitionTime":"2025-12-01T08:41:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:01 crc kubenswrapper[4813]: I1201 08:41:01.027694 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:01 crc kubenswrapper[4813]: I1201 08:41:01.027756 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:01 crc kubenswrapper[4813]: I1201 08:41:01.027775 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:01 crc kubenswrapper[4813]: I1201 08:41:01.027799 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:01 crc kubenswrapper[4813]: I1201 08:41:01.027845 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:01Z","lastTransitionTime":"2025-12-01T08:41:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:01 crc kubenswrapper[4813]: I1201 08:41:01.131421 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:01 crc kubenswrapper[4813]: I1201 08:41:01.131497 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:01 crc kubenswrapper[4813]: I1201 08:41:01.131519 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:01 crc kubenswrapper[4813]: I1201 08:41:01.131554 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:01 crc kubenswrapper[4813]: I1201 08:41:01.131580 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:01Z","lastTransitionTime":"2025-12-01T08:41:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:01 crc kubenswrapper[4813]: I1201 08:41:01.235796 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:01 crc kubenswrapper[4813]: I1201 08:41:01.235949 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:01 crc kubenswrapper[4813]: I1201 08:41:01.236017 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:01 crc kubenswrapper[4813]: I1201 08:41:01.236047 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:01 crc kubenswrapper[4813]: I1201 08:41:01.236097 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:01Z","lastTransitionTime":"2025-12-01T08:41:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:01 crc kubenswrapper[4813]: I1201 08:41:01.339561 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:01 crc kubenswrapper[4813]: I1201 08:41:01.339636 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:01 crc kubenswrapper[4813]: I1201 08:41:01.339666 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:01 crc kubenswrapper[4813]: I1201 08:41:01.339697 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:01 crc kubenswrapper[4813]: I1201 08:41:01.339721 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:01Z","lastTransitionTime":"2025-12-01T08:41:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:01 crc kubenswrapper[4813]: I1201 08:41:01.392776 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 08:41:01 crc kubenswrapper[4813]: I1201 08:41:01.392776 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-96mbr" Dec 01 08:41:01 crc kubenswrapper[4813]: E1201 08:41:01.393017 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 08:41:01 crc kubenswrapper[4813]: E1201 08:41:01.393216 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-96mbr" podUID="0789e3f7-5b08-4c04-8596-9ba29f50cdd0" Dec 01 08:41:01 crc kubenswrapper[4813]: I1201 08:41:01.442103 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:01 crc kubenswrapper[4813]: I1201 08:41:01.442157 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:01 crc kubenswrapper[4813]: I1201 08:41:01.442173 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:01 crc kubenswrapper[4813]: I1201 08:41:01.442197 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:01 crc kubenswrapper[4813]: I1201 08:41:01.442215 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:01Z","lastTransitionTime":"2025-12-01T08:41:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:01 crc kubenswrapper[4813]: I1201 08:41:01.544993 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:01 crc kubenswrapper[4813]: I1201 08:41:01.545045 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:01 crc kubenswrapper[4813]: I1201 08:41:01.545058 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:01 crc kubenswrapper[4813]: I1201 08:41:01.545080 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:01 crc kubenswrapper[4813]: I1201 08:41:01.545106 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:01Z","lastTransitionTime":"2025-12-01T08:41:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:01 crc kubenswrapper[4813]: I1201 08:41:01.648201 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:01 crc kubenswrapper[4813]: I1201 08:41:01.648299 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:01 crc kubenswrapper[4813]: I1201 08:41:01.648314 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:01 crc kubenswrapper[4813]: I1201 08:41:01.648333 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:01 crc kubenswrapper[4813]: I1201 08:41:01.648347 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:01Z","lastTransitionTime":"2025-12-01T08:41:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:01 crc kubenswrapper[4813]: I1201 08:41:01.750890 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:01 crc kubenswrapper[4813]: I1201 08:41:01.750924 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:01 crc kubenswrapper[4813]: I1201 08:41:01.750932 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:01 crc kubenswrapper[4813]: I1201 08:41:01.750946 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:01 crc kubenswrapper[4813]: I1201 08:41:01.750956 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:01Z","lastTransitionTime":"2025-12-01T08:41:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:01 crc kubenswrapper[4813]: I1201 08:41:01.854446 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:01 crc kubenswrapper[4813]: I1201 08:41:01.854518 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:01 crc kubenswrapper[4813]: I1201 08:41:01.854535 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:01 crc kubenswrapper[4813]: I1201 08:41:01.854564 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:01 crc kubenswrapper[4813]: I1201 08:41:01.854581 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:01Z","lastTransitionTime":"2025-12-01T08:41:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:01 crc kubenswrapper[4813]: I1201 08:41:01.957803 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:01 crc kubenswrapper[4813]: I1201 08:41:01.957935 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:01 crc kubenswrapper[4813]: I1201 08:41:01.958020 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:01 crc kubenswrapper[4813]: I1201 08:41:01.958123 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:01 crc kubenswrapper[4813]: I1201 08:41:01.958206 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:01Z","lastTransitionTime":"2025-12-01T08:41:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:02 crc kubenswrapper[4813]: I1201 08:41:02.061899 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:02 crc kubenswrapper[4813]: I1201 08:41:02.061998 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:02 crc kubenswrapper[4813]: I1201 08:41:02.062019 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:02 crc kubenswrapper[4813]: I1201 08:41:02.062045 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:02 crc kubenswrapper[4813]: I1201 08:41:02.062063 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:02Z","lastTransitionTime":"2025-12-01T08:41:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:02 crc kubenswrapper[4813]: I1201 08:41:02.165532 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:02 crc kubenswrapper[4813]: I1201 08:41:02.165609 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:02 crc kubenswrapper[4813]: I1201 08:41:02.165624 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:02 crc kubenswrapper[4813]: I1201 08:41:02.165643 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:02 crc kubenswrapper[4813]: I1201 08:41:02.165654 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:02Z","lastTransitionTime":"2025-12-01T08:41:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:02 crc kubenswrapper[4813]: I1201 08:41:02.268326 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:02 crc kubenswrapper[4813]: I1201 08:41:02.268358 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:02 crc kubenswrapper[4813]: I1201 08:41:02.268367 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:02 crc kubenswrapper[4813]: I1201 08:41:02.268380 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:02 crc kubenswrapper[4813]: I1201 08:41:02.268389 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:02Z","lastTransitionTime":"2025-12-01T08:41:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:02 crc kubenswrapper[4813]: I1201 08:41:02.370655 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:02 crc kubenswrapper[4813]: I1201 08:41:02.370696 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:02 crc kubenswrapper[4813]: I1201 08:41:02.370705 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:02 crc kubenswrapper[4813]: I1201 08:41:02.370748 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:02 crc kubenswrapper[4813]: I1201 08:41:02.370762 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:02Z","lastTransitionTime":"2025-12-01T08:41:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:02 crc kubenswrapper[4813]: I1201 08:41:02.392728 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 08:41:02 crc kubenswrapper[4813]: I1201 08:41:02.392744 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 08:41:02 crc kubenswrapper[4813]: E1201 08:41:02.393088 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 08:41:02 crc kubenswrapper[4813]: E1201 08:41:02.393344 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 08:41:02 crc kubenswrapper[4813]: I1201 08:41:02.477268 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:02 crc kubenswrapper[4813]: I1201 08:41:02.477599 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:02 crc kubenswrapper[4813]: I1201 08:41:02.477620 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:02 crc kubenswrapper[4813]: I1201 08:41:02.477641 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:02 crc kubenswrapper[4813]: I1201 08:41:02.477661 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:02Z","lastTransitionTime":"2025-12-01T08:41:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:02 crc kubenswrapper[4813]: I1201 08:41:02.581172 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:02 crc kubenswrapper[4813]: I1201 08:41:02.581250 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:02 crc kubenswrapper[4813]: I1201 08:41:02.581267 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:02 crc kubenswrapper[4813]: I1201 08:41:02.581290 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:02 crc kubenswrapper[4813]: I1201 08:41:02.581307 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:02Z","lastTransitionTime":"2025-12-01T08:41:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:02 crc kubenswrapper[4813]: I1201 08:41:02.684575 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:02 crc kubenswrapper[4813]: I1201 08:41:02.684641 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:02 crc kubenswrapper[4813]: I1201 08:41:02.684666 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:02 crc kubenswrapper[4813]: I1201 08:41:02.684716 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:02 crc kubenswrapper[4813]: I1201 08:41:02.684739 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:02Z","lastTransitionTime":"2025-12-01T08:41:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:02 crc kubenswrapper[4813]: I1201 08:41:02.787327 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:02 crc kubenswrapper[4813]: I1201 08:41:02.787381 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:02 crc kubenswrapper[4813]: I1201 08:41:02.787392 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:02 crc kubenswrapper[4813]: I1201 08:41:02.787414 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:02 crc kubenswrapper[4813]: I1201 08:41:02.787425 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:02Z","lastTransitionTime":"2025-12-01T08:41:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:02 crc kubenswrapper[4813]: I1201 08:41:02.889898 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:02 crc kubenswrapper[4813]: I1201 08:41:02.889943 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:02 crc kubenswrapper[4813]: I1201 08:41:02.889956 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:02 crc kubenswrapper[4813]: I1201 08:41:02.889993 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:02 crc kubenswrapper[4813]: I1201 08:41:02.890007 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:02Z","lastTransitionTime":"2025-12-01T08:41:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:02 crc kubenswrapper[4813]: I1201 08:41:02.993509 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:02 crc kubenswrapper[4813]: I1201 08:41:02.993572 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:02 crc kubenswrapper[4813]: I1201 08:41:02.993589 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:02 crc kubenswrapper[4813]: I1201 08:41:02.993616 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:02 crc kubenswrapper[4813]: I1201 08:41:02.993639 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:02Z","lastTransitionTime":"2025-12-01T08:41:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:03 crc kubenswrapper[4813]: I1201 08:41:03.096802 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:03 crc kubenswrapper[4813]: I1201 08:41:03.096854 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:03 crc kubenswrapper[4813]: I1201 08:41:03.096869 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:03 crc kubenswrapper[4813]: I1201 08:41:03.096887 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:03 crc kubenswrapper[4813]: I1201 08:41:03.096898 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:03Z","lastTransitionTime":"2025-12-01T08:41:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:03 crc kubenswrapper[4813]: I1201 08:41:03.202248 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:03 crc kubenswrapper[4813]: I1201 08:41:03.202348 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:03 crc kubenswrapper[4813]: I1201 08:41:03.202373 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:03 crc kubenswrapper[4813]: I1201 08:41:03.202411 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:03 crc kubenswrapper[4813]: I1201 08:41:03.202435 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:03Z","lastTransitionTime":"2025-12-01T08:41:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:03 crc kubenswrapper[4813]: I1201 08:41:03.305747 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:03 crc kubenswrapper[4813]: I1201 08:41:03.305814 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:03 crc kubenswrapper[4813]: I1201 08:41:03.305833 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:03 crc kubenswrapper[4813]: I1201 08:41:03.305859 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:03 crc kubenswrapper[4813]: I1201 08:41:03.305878 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:03Z","lastTransitionTime":"2025-12-01T08:41:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:03 crc kubenswrapper[4813]: I1201 08:41:03.393031 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-96mbr" Dec 01 08:41:03 crc kubenswrapper[4813]: I1201 08:41:03.393122 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 08:41:03 crc kubenswrapper[4813]: E1201 08:41:03.393291 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-96mbr" podUID="0789e3f7-5b08-4c04-8596-9ba29f50cdd0" Dec 01 08:41:03 crc kubenswrapper[4813]: E1201 08:41:03.393445 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 08:41:03 crc kubenswrapper[4813]: I1201 08:41:03.409190 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:03 crc kubenswrapper[4813]: I1201 08:41:03.409276 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:03 crc kubenswrapper[4813]: I1201 08:41:03.409304 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:03 crc kubenswrapper[4813]: I1201 08:41:03.409337 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:03 crc kubenswrapper[4813]: I1201 08:41:03.409366 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:03Z","lastTransitionTime":"2025-12-01T08:41:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:03 crc kubenswrapper[4813]: I1201 08:41:03.513761 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:03 crc kubenswrapper[4813]: I1201 08:41:03.513827 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:03 crc kubenswrapper[4813]: I1201 08:41:03.513844 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:03 crc kubenswrapper[4813]: I1201 08:41:03.513870 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:03 crc kubenswrapper[4813]: I1201 08:41:03.513891 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:03Z","lastTransitionTime":"2025-12-01T08:41:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:03 crc kubenswrapper[4813]: I1201 08:41:03.618283 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:03 crc kubenswrapper[4813]: I1201 08:41:03.618351 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:03 crc kubenswrapper[4813]: I1201 08:41:03.618369 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:03 crc kubenswrapper[4813]: I1201 08:41:03.618394 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:03 crc kubenswrapper[4813]: I1201 08:41:03.618412 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:03Z","lastTransitionTime":"2025-12-01T08:41:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:03 crc kubenswrapper[4813]: I1201 08:41:03.696355 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0789e3f7-5b08-4c04-8596-9ba29f50cdd0-metrics-certs\") pod \"network-metrics-daemon-96mbr\" (UID: \"0789e3f7-5b08-4c04-8596-9ba29f50cdd0\") " pod="openshift-multus/network-metrics-daemon-96mbr" Dec 01 08:41:03 crc kubenswrapper[4813]: E1201 08:41:03.696525 4813 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 01 08:41:03 crc kubenswrapper[4813]: E1201 08:41:03.696592 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/0789e3f7-5b08-4c04-8596-9ba29f50cdd0-metrics-certs podName:0789e3f7-5b08-4c04-8596-9ba29f50cdd0 nodeName:}" failed. No retries permitted until 2025-12-01 08:41:11.69657242 +0000 UTC m=+56.319494026 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/0789e3f7-5b08-4c04-8596-9ba29f50cdd0-metrics-certs") pod "network-metrics-daemon-96mbr" (UID: "0789e3f7-5b08-4c04-8596-9ba29f50cdd0") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 01 08:41:03 crc kubenswrapper[4813]: I1201 08:41:03.721144 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:03 crc kubenswrapper[4813]: I1201 08:41:03.721191 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:03 crc kubenswrapper[4813]: I1201 08:41:03.721202 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:03 crc kubenswrapper[4813]: I1201 08:41:03.721220 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:03 crc kubenswrapper[4813]: I1201 08:41:03.721233 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:03Z","lastTransitionTime":"2025-12-01T08:41:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:03 crc kubenswrapper[4813]: I1201 08:41:03.824406 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:03 crc kubenswrapper[4813]: I1201 08:41:03.824531 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:03 crc kubenswrapper[4813]: I1201 08:41:03.824556 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:03 crc kubenswrapper[4813]: I1201 08:41:03.824581 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:03 crc kubenswrapper[4813]: I1201 08:41:03.824602 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:03Z","lastTransitionTime":"2025-12-01T08:41:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:03 crc kubenswrapper[4813]: I1201 08:41:03.928149 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:03 crc kubenswrapper[4813]: I1201 08:41:03.928490 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:03 crc kubenswrapper[4813]: I1201 08:41:03.928624 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:03 crc kubenswrapper[4813]: I1201 08:41:03.928749 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:03 crc kubenswrapper[4813]: I1201 08:41:03.928936 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:03Z","lastTransitionTime":"2025-12-01T08:41:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:04 crc kubenswrapper[4813]: I1201 08:41:04.032306 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:04 crc kubenswrapper[4813]: I1201 08:41:04.032360 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:04 crc kubenswrapper[4813]: I1201 08:41:04.032374 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:04 crc kubenswrapper[4813]: I1201 08:41:04.032397 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:04 crc kubenswrapper[4813]: I1201 08:41:04.032414 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:04Z","lastTransitionTime":"2025-12-01T08:41:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:04 crc kubenswrapper[4813]: I1201 08:41:04.136085 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:04 crc kubenswrapper[4813]: I1201 08:41:04.136154 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:04 crc kubenswrapper[4813]: I1201 08:41:04.136179 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:04 crc kubenswrapper[4813]: I1201 08:41:04.136212 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:04 crc kubenswrapper[4813]: I1201 08:41:04.136235 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:04Z","lastTransitionTime":"2025-12-01T08:41:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:04 crc kubenswrapper[4813]: I1201 08:41:04.239673 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:04 crc kubenswrapper[4813]: I1201 08:41:04.239748 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:04 crc kubenswrapper[4813]: I1201 08:41:04.239766 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:04 crc kubenswrapper[4813]: I1201 08:41:04.239787 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:04 crc kubenswrapper[4813]: I1201 08:41:04.239800 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:04Z","lastTransitionTime":"2025-12-01T08:41:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:04 crc kubenswrapper[4813]: I1201 08:41:04.342899 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:04 crc kubenswrapper[4813]: I1201 08:41:04.342952 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:04 crc kubenswrapper[4813]: I1201 08:41:04.343003 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:04 crc kubenswrapper[4813]: I1201 08:41:04.343031 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:04 crc kubenswrapper[4813]: I1201 08:41:04.343047 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:04Z","lastTransitionTime":"2025-12-01T08:41:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:04 crc kubenswrapper[4813]: I1201 08:41:04.393494 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 08:41:04 crc kubenswrapper[4813]: E1201 08:41:04.394029 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 08:41:04 crc kubenswrapper[4813]: I1201 08:41:04.394572 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 08:41:04 crc kubenswrapper[4813]: E1201 08:41:04.394822 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 08:41:04 crc kubenswrapper[4813]: I1201 08:41:04.446347 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:04 crc kubenswrapper[4813]: I1201 08:41:04.446375 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:04 crc kubenswrapper[4813]: I1201 08:41:04.446383 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:04 crc kubenswrapper[4813]: I1201 08:41:04.446396 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:04 crc kubenswrapper[4813]: I1201 08:41:04.446405 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:04Z","lastTransitionTime":"2025-12-01T08:41:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:04 crc kubenswrapper[4813]: I1201 08:41:04.549787 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:04 crc kubenswrapper[4813]: I1201 08:41:04.549856 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:04 crc kubenswrapper[4813]: I1201 08:41:04.549876 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:04 crc kubenswrapper[4813]: I1201 08:41:04.549902 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:04 crc kubenswrapper[4813]: I1201 08:41:04.549920 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:04Z","lastTransitionTime":"2025-12-01T08:41:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:04 crc kubenswrapper[4813]: I1201 08:41:04.606267 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 01 08:41:04 crc kubenswrapper[4813]: I1201 08:41:04.620096 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Dec 01 08:41:04 crc kubenswrapper[4813]: I1201 08:41:04.631492 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:04Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:04 crc kubenswrapper[4813]: I1201 08:41:04.651905 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c7574f892d5e9142296b26a0b484eb6759ab53f571139fb6e23b93dc3da05780\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:04Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:04 crc kubenswrapper[4813]: I1201 08:41:04.653624 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:04 crc kubenswrapper[4813]: I1201 08:41:04.653719 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:04 crc kubenswrapper[4813]: I1201 08:41:04.653738 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:04 crc kubenswrapper[4813]: I1201 08:41:04.653783 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:04 crc kubenswrapper[4813]: I1201 08:41:04.653801 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:04Z","lastTransitionTime":"2025-12-01T08:41:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:04 crc kubenswrapper[4813]: I1201 08:41:04.673899 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:04Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:04 crc kubenswrapper[4813]: I1201 08:41:04.688650 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-xpnlc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"22da1fcf-0c57-4708-90d8-77692af17344\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4a36d213bdad2da91d0d33f400d5a8248befb68d5b94d19347873a93c6712a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xc44d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-xpnlc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:04Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:04 crc kubenswrapper[4813]: I1201 08:41:04.712708 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6fb5a648-ae4f-4e46-a530-007e9bf84abb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e24ebfc52dcb28368a2d436e7d9ab78f07f2aec2a4c71405a6c7d3301eee96a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33898118bbf639c8e6eae5aafa10386b95795512444ed5f535db3c663f169df2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9902c00bf7738686c93f62e1772a484d05715d74b75ee98c84c8f23ed383824b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3657b47ec9f1563c83bfbe3d728c8fd123f7078e219d251b1874269e42c31eea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:16Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:04Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:04 crc kubenswrapper[4813]: I1201 08:41:04.734474 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0023b0e358d75aacc889ce364846dda89ea513832c87fa4b066aea7dac57e8a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:04Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:04 crc kubenswrapper[4813]: I1201 08:41:04.754019 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3264b3af996af452df2953c59bea6deeb5ea6af1b119ce89f7334595c27e70da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9e879b1f563bd66383d58fd7b435a8a81bd50cb387dc1fabb6fc47d5a3fa468\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:04Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:04 crc kubenswrapper[4813]: I1201 08:41:04.761959 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:04 crc kubenswrapper[4813]: I1201 08:41:04.762043 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:04 crc kubenswrapper[4813]: I1201 08:41:04.762169 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:04 crc kubenswrapper[4813]: I1201 08:41:04.762193 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:04 crc kubenswrapper[4813]: I1201 08:41:04.762207 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:04Z","lastTransitionTime":"2025-12-01T08:41:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:04 crc kubenswrapper[4813]: I1201 08:41:04.772511 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-b6qqv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9656a03-0096-4c89-b0da-e37e4103611b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4195aebc1b2e388d480920a7f72257f6660afc99ad8c1a09478ef445739b272\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cjvnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:41Z\\\"}}\" for pod \"openshift-multus\"/\"multus-b6qqv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:04Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:04 crc kubenswrapper[4813]: I1201 08:41:04.785802 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-96mbr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0789e3f7-5b08-4c04-8596-9ba29f50cdd0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xt4fj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xt4fj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:55Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-96mbr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:04Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:04 crc kubenswrapper[4813]: I1201 08:41:04.802032 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"349332bd-6b4d-4306-84ae-07019ca3860f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d441a348087ae3cadda1ad8a7d1c4dff02cf3f6e1d346c294b278fdd9131edcc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://da67635daf6d9e32937f4d5d4ceeee3f8f09ae4fd2dd26c3e85be1286c5949de\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e97356ee274ea78d9f237ffbc756abb6ac8df0dbac244bf0a86ed215b8cbf43\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8dc6feb9817f4f356b8e36af43ec3070774f312fbbb4909fa05b8b3e8dd53582\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a0404f5a5ed5a9750d4c479b608e68e8960b448f455d5af9d219aa5edaeb789b\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"message\\\":\\\"\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764578421\\\\\\\\\\\\\\\" (2025-12-01 08:40:21 +0000 UTC to 2025-12-31 08:40:22 +0000 UTC (now=2025-12-01 08:40:38.066068451 +0000 UTC))\\\\\\\"\\\\nI1201 08:40:38.066144 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1201 08:40:38.066166 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1201 08:40:38.066220 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1201 08:40:38.066232 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1201 08:40:38.066338 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764578432\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764578432\\\\\\\\\\\\\\\" (2025-12-01 07:40:32 +0000 UTC to 2026-12-01 07:40:32 +0000 UTC (now=2025-12-01 08:40:38.066301217 +0000 UTC))\\\\\\\"\\\\nI1201 08:40:38.066357 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1201 08:40:38.066385 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1201 08:40:38.066402 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1201 08:40:38.066414 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-63890811/tls.crt::/tmp/serving-cert-63890811/tls.key\\\\\\\"\\\\nI1201 08:40:38.066419 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1201 08:40:38.066572 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1201 08:40:38.067159 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7354f9337d42115bb1d974bf3772f2d03365acba57f2155135c7a0bbab39879e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:19Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05492df5703392cdf0932d6bf66025b18e53bd05fa0857a7cb2ad70979cbdc3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05492df5703392cdf0932d6bf66025b18e53bd05fa0857a7cb2ad70979cbdc3e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:16Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:04Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:04 crc kubenswrapper[4813]: I1201 08:41:04.818753 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-8pg2p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dd53b38d-d9dd-4f98-9d06-dd95f1a4e3a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a214d71293e4e09573c665524d880fc319dfb1c0c6d95d11d78f253865660fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnjhq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ddf9d5989e4195b3074de08f99d53b48066f53fe93ce204145b3c714996fc57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnjhq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:53Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-8pg2p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:04Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:04 crc kubenswrapper[4813]: I1201 08:41:04.832305 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-9wn4q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3d423402-e08f-42e9-b60c-6eb9bf2f6d48\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3049b842978c6580e1033a442a37766a2e76e278aface9a89fc0b2c57faf0df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rzxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c915733072b32a63237e2401cbf84b059aca396c2c0db884f0f439d599a35d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rzxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:41Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-9wn4q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:04Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:04 crc kubenswrapper[4813]: I1201 08:41:04.866125 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-c576v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"17292cd0-1eac-40b5-a60e-67167c57409c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a42206c914d0493ca74788ea1a941de80920ab7b810fb95227cd9b8711e8f4b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7zwfx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:41Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-c576v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:04Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:04 crc kubenswrapper[4813]: I1201 08:41:04.866187 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:04 crc kubenswrapper[4813]: I1201 08:41:04.866738 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:04 crc kubenswrapper[4813]: I1201 08:41:04.866876 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:04 crc kubenswrapper[4813]: I1201 08:41:04.867062 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:04 crc kubenswrapper[4813]: I1201 08:41:04.867214 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:04Z","lastTransitionTime":"2025-12-01T08:41:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:04 crc kubenswrapper[4813]: I1201 08:41:04.897784 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-j4pcc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b7505e14-3560-4d0d-be2c-d6a7a5953865\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ccc30a63dc1c23ff9880fb04af1f31f0656b2696d359988e23deed23989f480a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://944977e9722025088720107530d0c91fc693593f380e4b49678f28a56da180cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc32d930ab23f30103d8fa42291909cbec89ebb75c5c53cd783cd27fe1a3f09e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f19c67b1c6c80c7f3223512102810d46ede1445464d21ae3bbea83424539b2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d984c90302e57f3e5a6f3ca3613b415db04ac74a5a228582f1b52ef1d59f93e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce8963bf1e9ee66926907aaf3a03d18977f4d0a1983d7121b62fbc8e02a37e4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8790f4730881303e7743498b66fc17b4b0df5b6620c8ca640ad7b9c90e6898d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4b11f49feda579bcb62f9680732033819ea11e7ae9937a06c0623a54c0ec106c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T08:40:54Z\\\",\\\"message\\\":\\\" *v1.Pod event handler 3\\\\nI1201 08:40:53.558214 6038 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1201 08:40:53.559209 6038 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1201 08:40:53.559251 6038 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1201 08:40:53.559264 6038 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1201 08:40:53.559268 6038 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1201 08:40:53.559286 6038 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1201 08:40:53.559299 6038 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1201 08:40:53.559315 6038 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1201 08:40:53.559320 6038 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1201 08:40:53.559327 6038 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1201 08:40:53.559329 6038 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1201 08:40:53.559336 6038 handler.go:208] Removed *v1.Node event handler 7\\\\nI1201 08:40:53.559354 6038 factory.go:656] Stopping watch factory\\\\nI1201 08:40:53.559365 6038 handler.go:208] Removed *v1.Node ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:48Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8790f4730881303e7743498b66fc17b4b0df5b6620c8ca640ad7b9c90e6898d2\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T08:40:56Z\\\",\\\"message\\\":\\\"I1201 08:40:56.782175 6307 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1201 08:40:56.782219 6307 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1201 08:40:56.782251 6307 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1201 08:40:56.782302 6307 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1201 08:40:56.782342 6307 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1201 08:40:56.782379 6307 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1201 08:40:56.782414 6307 factory.go:656] Stopping watch factory\\\\nI1201 08:40:56.782444 6307 ovnkube.go:599] Stopped ovnkube\\\\nI1201 08:40:56.782487 6307 handler.go:208] Removed *v1.Node event handler 7\\\\nI1201 08:40:56.782514 6307 handler.go:208] Removed *v1.Node event handler 2\\\\nI1201 08:40:56.782539 6307 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1201 08:40:56.782564 6307 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1201 08:40:56.782591 6307 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1201 08:40:56.782650 6307 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1201 08:40:56.782670 6307 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1201 08:40:56.782663 6307 handler.go:208] Removed *v1.EgressFirewa\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b45454b440c5108fa150a6b856afa1778ba6e9f89633433cb3117b68f00f72c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd67634e0c806323334495401a4e20b48fea4bc5013acd973d2abd3d70870f99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd67634e0c806323334495401a4e20b48fea4bc5013acd973d2abd3d70870f99\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:41Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-j4pcc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:04Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:04 crc kubenswrapper[4813]: I1201 08:41:04.921904 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-h7lvd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d431b613-a91b-4fff-88b9-447804766289\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a32c51c943fcf6ada7e3f614667eb023e29d94c11128d26c1fecca37b03963b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://495593f99e92a9c28100a95e27d06ff5847008c5d996366cb474bedd9b69d5ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://495593f99e92a9c28100a95e27d06ff5847008c5d996366cb474bedd9b69d5ef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://19358f83a6c52957f959fe935a6843fdeef59905cc395e6e52a1fb4a670ac193\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://19358f83a6c52957f959fe935a6843fdeef59905cc395e6e52a1fb4a670ac193\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f089c115d916e339994879d224e3db27642b99aaabf59bcafb2979549052945\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9f089c115d916e339994879d224e3db27642b99aaabf59bcafb2979549052945\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb66f224462305c2005b31ef7bb9ed01efde8101e18518dd42d1dfb0df765bb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb66f224462305c2005b31ef7bb9ed01efde8101e18518dd42d1dfb0df765bb9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c268236c9a155c9b1f47c48d30dbd23948ceb2545592b40f356c60ea1d12e8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c268236c9a155c9b1f47c48d30dbd23948ceb2545592b40f356c60ea1d12e8a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://458aedca5f186e9aaed7e4ee29fd5247221491888b130f55c1fe48d88d565f8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://458aedca5f186e9aaed7e4ee29fd5247221491888b130f55c1fe48d88d565f8f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:41Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-h7lvd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:04Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:04 crc kubenswrapper[4813]: I1201 08:41:04.942478 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:04Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:04 crc kubenswrapper[4813]: I1201 08:41:04.971003 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:04 crc kubenswrapper[4813]: I1201 08:41:04.971249 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:04 crc kubenswrapper[4813]: I1201 08:41:04.971486 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:04 crc kubenswrapper[4813]: I1201 08:41:04.971705 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:04 crc kubenswrapper[4813]: I1201 08:41:04.971876 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:04Z","lastTransitionTime":"2025-12-01T08:41:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:05 crc kubenswrapper[4813]: I1201 08:41:05.075374 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:05 crc kubenswrapper[4813]: I1201 08:41:05.075739 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:05 crc kubenswrapper[4813]: I1201 08:41:05.075956 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:05 crc kubenswrapper[4813]: I1201 08:41:05.076238 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:05 crc kubenswrapper[4813]: I1201 08:41:05.076462 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:05Z","lastTransitionTime":"2025-12-01T08:41:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:05 crc kubenswrapper[4813]: I1201 08:41:05.180282 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:05 crc kubenswrapper[4813]: I1201 08:41:05.180378 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:05 crc kubenswrapper[4813]: I1201 08:41:05.180402 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:05 crc kubenswrapper[4813]: I1201 08:41:05.180435 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:05 crc kubenswrapper[4813]: I1201 08:41:05.180458 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:05Z","lastTransitionTime":"2025-12-01T08:41:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:05 crc kubenswrapper[4813]: I1201 08:41:05.284107 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:05 crc kubenswrapper[4813]: I1201 08:41:05.284174 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:05 crc kubenswrapper[4813]: I1201 08:41:05.284195 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:05 crc kubenswrapper[4813]: I1201 08:41:05.284225 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:05 crc kubenswrapper[4813]: I1201 08:41:05.284247 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:05Z","lastTransitionTime":"2025-12-01T08:41:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:05 crc kubenswrapper[4813]: I1201 08:41:05.387492 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:05 crc kubenswrapper[4813]: I1201 08:41:05.387558 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:05 crc kubenswrapper[4813]: I1201 08:41:05.387575 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:05 crc kubenswrapper[4813]: I1201 08:41:05.387604 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:05 crc kubenswrapper[4813]: I1201 08:41:05.387623 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:05Z","lastTransitionTime":"2025-12-01T08:41:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:05 crc kubenswrapper[4813]: I1201 08:41:05.392916 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-96mbr" Dec 01 08:41:05 crc kubenswrapper[4813]: I1201 08:41:05.393001 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 08:41:05 crc kubenswrapper[4813]: E1201 08:41:05.393193 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-96mbr" podUID="0789e3f7-5b08-4c04-8596-9ba29f50cdd0" Dec 01 08:41:05 crc kubenswrapper[4813]: E1201 08:41:05.393536 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 08:41:05 crc kubenswrapper[4813]: I1201 08:41:05.491249 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:05 crc kubenswrapper[4813]: I1201 08:41:05.491299 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:05 crc kubenswrapper[4813]: I1201 08:41:05.491307 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:05 crc kubenswrapper[4813]: I1201 08:41:05.491323 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:05 crc kubenswrapper[4813]: I1201 08:41:05.491332 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:05Z","lastTransitionTime":"2025-12-01T08:41:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:05 crc kubenswrapper[4813]: I1201 08:41:05.595053 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:05 crc kubenswrapper[4813]: I1201 08:41:05.595121 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:05 crc kubenswrapper[4813]: I1201 08:41:05.595139 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:05 crc kubenswrapper[4813]: I1201 08:41:05.595166 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:05 crc kubenswrapper[4813]: I1201 08:41:05.595183 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:05Z","lastTransitionTime":"2025-12-01T08:41:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:05 crc kubenswrapper[4813]: I1201 08:41:05.697380 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:05 crc kubenswrapper[4813]: I1201 08:41:05.697424 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:05 crc kubenswrapper[4813]: I1201 08:41:05.697434 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:05 crc kubenswrapper[4813]: I1201 08:41:05.697450 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:05 crc kubenswrapper[4813]: I1201 08:41:05.697460 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:05Z","lastTransitionTime":"2025-12-01T08:41:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:05 crc kubenswrapper[4813]: I1201 08:41:05.800805 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:05 crc kubenswrapper[4813]: I1201 08:41:05.800881 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:05 crc kubenswrapper[4813]: I1201 08:41:05.800894 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:05 crc kubenswrapper[4813]: I1201 08:41:05.800912 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:05 crc kubenswrapper[4813]: I1201 08:41:05.800923 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:05Z","lastTransitionTime":"2025-12-01T08:41:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:05 crc kubenswrapper[4813]: I1201 08:41:05.904280 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:05 crc kubenswrapper[4813]: I1201 08:41:05.904332 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:05 crc kubenswrapper[4813]: I1201 08:41:05.904381 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:05 crc kubenswrapper[4813]: I1201 08:41:05.904406 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:05 crc kubenswrapper[4813]: I1201 08:41:05.904423 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:05Z","lastTransitionTime":"2025-12-01T08:41:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:06 crc kubenswrapper[4813]: I1201 08:41:06.007116 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:06 crc kubenswrapper[4813]: I1201 08:41:06.007161 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:06 crc kubenswrapper[4813]: I1201 08:41:06.007171 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:06 crc kubenswrapper[4813]: I1201 08:41:06.007187 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:06 crc kubenswrapper[4813]: I1201 08:41:06.007198 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:06Z","lastTransitionTime":"2025-12-01T08:41:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:06 crc kubenswrapper[4813]: I1201 08:41:06.057726 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-j4pcc" Dec 01 08:41:06 crc kubenswrapper[4813]: I1201 08:41:06.058693 4813 scope.go:117] "RemoveContainer" containerID="8790f4730881303e7743498b66fc17b4b0df5b6620c8ca640ad7b9c90e6898d2" Dec 01 08:41:06 crc kubenswrapper[4813]: E1201 08:41:06.058957 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-j4pcc_openshift-ovn-kubernetes(b7505e14-3560-4d0d-be2c-d6a7a5953865)\"" pod="openshift-ovn-kubernetes/ovnkube-node-j4pcc" podUID="b7505e14-3560-4d0d-be2c-d6a7a5953865" Dec 01 08:41:06 crc kubenswrapper[4813]: I1201 08:41:06.073163 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-9wn4q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3d423402-e08f-42e9-b60c-6eb9bf2f6d48\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3049b842978c6580e1033a442a37766a2e76e278aface9a89fc0b2c57faf0df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rzxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c915733072b32a63237e2401cbf84b059aca396c2c0db884f0f439d599a35d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rzxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:41Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-9wn4q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:06Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:06 crc kubenswrapper[4813]: I1201 08:41:06.087041 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-8pg2p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dd53b38d-d9dd-4f98-9d06-dd95f1a4e3a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a214d71293e4e09573c665524d880fc319dfb1c0c6d95d11d78f253865660fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnjhq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ddf9d5989e4195b3074de08f99d53b48066f53fe93ce204145b3c714996fc57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnjhq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:53Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-8pg2p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:06Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:06 crc kubenswrapper[4813]: I1201 08:41:06.106679 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-h7lvd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d431b613-a91b-4fff-88b9-447804766289\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a32c51c943fcf6ada7e3f614667eb023e29d94c11128d26c1fecca37b03963b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://495593f99e92a9c28100a95e27d06ff5847008c5d996366cb474bedd9b69d5ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://495593f99e92a9c28100a95e27d06ff5847008c5d996366cb474bedd9b69d5ef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://19358f83a6c52957f959fe935a6843fdeef59905cc395e6e52a1fb4a670ac193\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://19358f83a6c52957f959fe935a6843fdeef59905cc395e6e52a1fb4a670ac193\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f089c115d916e339994879d224e3db27642b99aaabf59bcafb2979549052945\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9f089c115d916e339994879d224e3db27642b99aaabf59bcafb2979549052945\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb66f224462305c2005b31ef7bb9ed01efde8101e18518dd42d1dfb0df765bb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb66f224462305c2005b31ef7bb9ed01efde8101e18518dd42d1dfb0df765bb9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c268236c9a155c9b1f47c48d30dbd23948ceb2545592b40f356c60ea1d12e8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c268236c9a155c9b1f47c48d30dbd23948ceb2545592b40f356c60ea1d12e8a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://458aedca5f186e9aaed7e4ee29fd5247221491888b130f55c1fe48d88d565f8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://458aedca5f186e9aaed7e4ee29fd5247221491888b130f55c1fe48d88d565f8f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:41Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-h7lvd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:06Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:06 crc kubenswrapper[4813]: I1201 08:41:06.109277 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:06 crc kubenswrapper[4813]: I1201 08:41:06.109311 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:06 crc kubenswrapper[4813]: I1201 08:41:06.109323 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:06 crc kubenswrapper[4813]: I1201 08:41:06.109342 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:06 crc kubenswrapper[4813]: I1201 08:41:06.109356 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:06Z","lastTransitionTime":"2025-12-01T08:41:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:06 crc kubenswrapper[4813]: I1201 08:41:06.125102 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"490e526a-902b-4aff-b547-0cefea562764\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb48bd28fd3d29781695a5ae630f4913ef2745e97fbe7d463cca39a31e43ba0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74f8cf73f07b4cde260af429d26413e4e6226e7b177b493dd9c0e5c043ab6e1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3aaf696cefbe398bb6b26acfc5be43bda7e2be18218c8060840350c4106e66a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53de1f0682f36aa83ad8e5c226b95daf9f97df4ae154ebfc2fc287f9f238a4a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53de1f0682f36aa83ad8e5c226b95daf9f97df4ae154ebfc2fc287f9f238a4a8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:17Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:16Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:06Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:06 crc kubenswrapper[4813]: I1201 08:41:06.142646 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:06Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:06 crc kubenswrapper[4813]: I1201 08:41:06.155067 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-c576v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"17292cd0-1eac-40b5-a60e-67167c57409c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a42206c914d0493ca74788ea1a941de80920ab7b810fb95227cd9b8711e8f4b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7zwfx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:41Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-c576v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:06Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:06 crc kubenswrapper[4813]: I1201 08:41:06.189226 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-j4pcc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b7505e14-3560-4d0d-be2c-d6a7a5953865\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ccc30a63dc1c23ff9880fb04af1f31f0656b2696d359988e23deed23989f480a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://944977e9722025088720107530d0c91fc693593f380e4b49678f28a56da180cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc32d930ab23f30103d8fa42291909cbec89ebb75c5c53cd783cd27fe1a3f09e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f19c67b1c6c80c7f3223512102810d46ede1445464d21ae3bbea83424539b2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d984c90302e57f3e5a6f3ca3613b415db04ac74a5a228582f1b52ef1d59f93e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce8963bf1e9ee66926907aaf3a03d18977f4d0a1983d7121b62fbc8e02a37e4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8790f4730881303e7743498b66fc17b4b0df5b6620c8ca640ad7b9c90e6898d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8790f4730881303e7743498b66fc17b4b0df5b6620c8ca640ad7b9c90e6898d2\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T08:40:56Z\\\",\\\"message\\\":\\\"I1201 08:40:56.782175 6307 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1201 08:40:56.782219 6307 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1201 08:40:56.782251 6307 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1201 08:40:56.782302 6307 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1201 08:40:56.782342 6307 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1201 08:40:56.782379 6307 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1201 08:40:56.782414 6307 factory.go:656] Stopping watch factory\\\\nI1201 08:40:56.782444 6307 ovnkube.go:599] Stopped ovnkube\\\\nI1201 08:40:56.782487 6307 handler.go:208] Removed *v1.Node event handler 7\\\\nI1201 08:40:56.782514 6307 handler.go:208] Removed *v1.Node event handler 2\\\\nI1201 08:40:56.782539 6307 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1201 08:40:56.782564 6307 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1201 08:40:56.782591 6307 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1201 08:40:56.782650 6307 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1201 08:40:56.782670 6307 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1201 08:40:56.782663 6307 handler.go:208] Removed *v1.EgressFirewa\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:55Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-j4pcc_openshift-ovn-kubernetes(b7505e14-3560-4d0d-be2c-d6a7a5953865)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b45454b440c5108fa150a6b856afa1778ba6e9f89633433cb3117b68f00f72c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd67634e0c806323334495401a4e20b48fea4bc5013acd973d2abd3d70870f99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd67634e0c806323334495401a4e20b48fea4bc5013acd973d2abd3d70870f99\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:41Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-j4pcc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:06Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:06 crc kubenswrapper[4813]: I1201 08:41:06.208290 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:06Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:06 crc kubenswrapper[4813]: I1201 08:41:06.213234 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:06 crc kubenswrapper[4813]: I1201 08:41:06.213274 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:06 crc kubenswrapper[4813]: I1201 08:41:06.213287 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:06 crc kubenswrapper[4813]: I1201 08:41:06.213303 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:06 crc kubenswrapper[4813]: I1201 08:41:06.213511 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:06Z","lastTransitionTime":"2025-12-01T08:41:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:06 crc kubenswrapper[4813]: I1201 08:41:06.218874 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-xpnlc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"22da1fcf-0c57-4708-90d8-77692af17344\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4a36d213bdad2da91d0d33f400d5a8248befb68d5b94d19347873a93c6712a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xc44d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-xpnlc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:06Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:06 crc kubenswrapper[4813]: I1201 08:41:06.234617 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6fb5a648-ae4f-4e46-a530-007e9bf84abb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e24ebfc52dcb28368a2d436e7d9ab78f07f2aec2a4c71405a6c7d3301eee96a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33898118bbf639c8e6eae5aafa10386b95795512444ed5f535db3c663f169df2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9902c00bf7738686c93f62e1772a484d05715d74b75ee98c84c8f23ed383824b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3657b47ec9f1563c83bfbe3d728c8fd123f7078e219d251b1874269e42c31eea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:16Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:06Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:06 crc kubenswrapper[4813]: I1201 08:41:06.251681 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0023b0e358d75aacc889ce364846dda89ea513832c87fa4b066aea7dac57e8a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:06Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:06 crc kubenswrapper[4813]: I1201 08:41:06.271455 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:06Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:06 crc kubenswrapper[4813]: I1201 08:41:06.283759 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c7574f892d5e9142296b26a0b484eb6759ab53f571139fb6e23b93dc3da05780\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:06Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:06 crc kubenswrapper[4813]: I1201 08:41:06.297493 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-96mbr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0789e3f7-5b08-4c04-8596-9ba29f50cdd0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xt4fj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xt4fj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:55Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-96mbr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:06Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:06 crc kubenswrapper[4813]: I1201 08:41:06.312536 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"349332bd-6b4d-4306-84ae-07019ca3860f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d441a348087ae3cadda1ad8a7d1c4dff02cf3f6e1d346c294b278fdd9131edcc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://da67635daf6d9e32937f4d5d4ceeee3f8f09ae4fd2dd26c3e85be1286c5949de\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e97356ee274ea78d9f237ffbc756abb6ac8df0dbac244bf0a86ed215b8cbf43\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8dc6feb9817f4f356b8e36af43ec3070774f312fbbb4909fa05b8b3e8dd53582\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a0404f5a5ed5a9750d4c479b608e68e8960b448f455d5af9d219aa5edaeb789b\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"message\\\":\\\"\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764578421\\\\\\\\\\\\\\\" (2025-12-01 08:40:21 +0000 UTC to 2025-12-31 08:40:22 +0000 UTC (now=2025-12-01 08:40:38.066068451 +0000 UTC))\\\\\\\"\\\\nI1201 08:40:38.066144 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1201 08:40:38.066166 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1201 08:40:38.066220 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1201 08:40:38.066232 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1201 08:40:38.066338 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764578432\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764578432\\\\\\\\\\\\\\\" (2025-12-01 07:40:32 +0000 UTC to 2026-12-01 07:40:32 +0000 UTC (now=2025-12-01 08:40:38.066301217 +0000 UTC))\\\\\\\"\\\\nI1201 08:40:38.066357 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1201 08:40:38.066385 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1201 08:40:38.066402 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1201 08:40:38.066414 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-63890811/tls.crt::/tmp/serving-cert-63890811/tls.key\\\\\\\"\\\\nI1201 08:40:38.066419 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1201 08:40:38.066572 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1201 08:40:38.067159 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7354f9337d42115bb1d974bf3772f2d03365acba57f2155135c7a0bbab39879e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:19Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05492df5703392cdf0932d6bf66025b18e53bd05fa0857a7cb2ad70979cbdc3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05492df5703392cdf0932d6bf66025b18e53bd05fa0857a7cb2ad70979cbdc3e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:16Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:06Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:06 crc kubenswrapper[4813]: I1201 08:41:06.316253 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:06 crc kubenswrapper[4813]: I1201 08:41:06.316284 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:06 crc kubenswrapper[4813]: I1201 08:41:06.316294 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:06 crc kubenswrapper[4813]: I1201 08:41:06.316309 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:06 crc kubenswrapper[4813]: I1201 08:41:06.316318 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:06Z","lastTransitionTime":"2025-12-01T08:41:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:06 crc kubenswrapper[4813]: I1201 08:41:06.326541 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3264b3af996af452df2953c59bea6deeb5ea6af1b119ce89f7334595c27e70da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9e879b1f563bd66383d58fd7b435a8a81bd50cb387dc1fabb6fc47d5a3fa468\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:06Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:06 crc kubenswrapper[4813]: I1201 08:41:06.343015 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-b6qqv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9656a03-0096-4c89-b0da-e37e4103611b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4195aebc1b2e388d480920a7f72257f6660afc99ad8c1a09478ef445739b272\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cjvnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:41Z\\\"}}\" for pod \"openshift-multus\"/\"multus-b6qqv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:06Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:06 crc kubenswrapper[4813]: I1201 08:41:06.393680 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 08:41:06 crc kubenswrapper[4813]: E1201 08:41:06.393944 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 08:41:06 crc kubenswrapper[4813]: I1201 08:41:06.393613 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 08:41:06 crc kubenswrapper[4813]: E1201 08:41:06.395481 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 08:41:06 crc kubenswrapper[4813]: I1201 08:41:06.418747 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:06 crc kubenswrapper[4813]: I1201 08:41:06.418827 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:06 crc kubenswrapper[4813]: I1201 08:41:06.418844 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:06 crc kubenswrapper[4813]: I1201 08:41:06.418866 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:06 crc kubenswrapper[4813]: I1201 08:41:06.418882 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:06Z","lastTransitionTime":"2025-12-01T08:41:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:06 crc kubenswrapper[4813]: I1201 08:41:06.419414 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"349332bd-6b4d-4306-84ae-07019ca3860f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d441a348087ae3cadda1ad8a7d1c4dff02cf3f6e1d346c294b278fdd9131edcc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://da67635daf6d9e32937f4d5d4ceeee3f8f09ae4fd2dd26c3e85be1286c5949de\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e97356ee274ea78d9f237ffbc756abb6ac8df0dbac244bf0a86ed215b8cbf43\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8dc6feb9817f4f356b8e36af43ec3070774f312fbbb4909fa05b8b3e8dd53582\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a0404f5a5ed5a9750d4c479b608e68e8960b448f455d5af9d219aa5edaeb789b\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"message\\\":\\\"\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764578421\\\\\\\\\\\\\\\" (2025-12-01 08:40:21 +0000 UTC to 2025-12-31 08:40:22 +0000 UTC (now=2025-12-01 08:40:38.066068451 +0000 UTC))\\\\\\\"\\\\nI1201 08:40:38.066144 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1201 08:40:38.066166 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1201 08:40:38.066220 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1201 08:40:38.066232 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1201 08:40:38.066338 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764578432\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764578432\\\\\\\\\\\\\\\" (2025-12-01 07:40:32 +0000 UTC to 2026-12-01 07:40:32 +0000 UTC (now=2025-12-01 08:40:38.066301217 +0000 UTC))\\\\\\\"\\\\nI1201 08:40:38.066357 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1201 08:40:38.066385 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1201 08:40:38.066402 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1201 08:40:38.066414 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-63890811/tls.crt::/tmp/serving-cert-63890811/tls.key\\\\\\\"\\\\nI1201 08:40:38.066419 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1201 08:40:38.066572 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1201 08:40:38.067159 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7354f9337d42115bb1d974bf3772f2d03365acba57f2155135c7a0bbab39879e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:19Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05492df5703392cdf0932d6bf66025b18e53bd05fa0857a7cb2ad70979cbdc3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05492df5703392cdf0932d6bf66025b18e53bd05fa0857a7cb2ad70979cbdc3e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:16Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:06Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:06 crc kubenswrapper[4813]: I1201 08:41:06.433129 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3264b3af996af452df2953c59bea6deeb5ea6af1b119ce89f7334595c27e70da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9e879b1f563bd66383d58fd7b435a8a81bd50cb387dc1fabb6fc47d5a3fa468\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:06Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:06 crc kubenswrapper[4813]: I1201 08:41:06.442718 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 08:41:06 crc kubenswrapper[4813]: I1201 08:41:06.449451 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-b6qqv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9656a03-0096-4c89-b0da-e37e4103611b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4195aebc1b2e388d480920a7f72257f6660afc99ad8c1a09478ef445739b272\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cjvnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:41Z\\\"}}\" for pod \"openshift-multus\"/\"multus-b6qqv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:06Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:06 crc kubenswrapper[4813]: I1201 08:41:06.461093 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-96mbr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0789e3f7-5b08-4c04-8596-9ba29f50cdd0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xt4fj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xt4fj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:55Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-96mbr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:06Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:06 crc kubenswrapper[4813]: I1201 08:41:06.471593 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-9wn4q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3d423402-e08f-42e9-b60c-6eb9bf2f6d48\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3049b842978c6580e1033a442a37766a2e76e278aface9a89fc0b2c57faf0df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rzxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c915733072b32a63237e2401cbf84b059aca396c2c0db884f0f439d599a35d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rzxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:41Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-9wn4q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:06Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:06 crc kubenswrapper[4813]: I1201 08:41:06.483503 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-8pg2p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dd53b38d-d9dd-4f98-9d06-dd95f1a4e3a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a214d71293e4e09573c665524d880fc319dfb1c0c6d95d11d78f253865660fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnjhq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ddf9d5989e4195b3074de08f99d53b48066f53fe93ce204145b3c714996fc57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnjhq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:53Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-8pg2p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:06Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:06 crc kubenswrapper[4813]: I1201 08:41:06.499676 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"490e526a-902b-4aff-b547-0cefea562764\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb48bd28fd3d29781695a5ae630f4913ef2745e97fbe7d463cca39a31e43ba0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74f8cf73f07b4cde260af429d26413e4e6226e7b177b493dd9c0e5c043ab6e1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3aaf696cefbe398bb6b26acfc5be43bda7e2be18218c8060840350c4106e66a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53de1f0682f36aa83ad8e5c226b95daf9f97df4ae154ebfc2fc287f9f238a4a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53de1f0682f36aa83ad8e5c226b95daf9f97df4ae154ebfc2fc287f9f238a4a8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:17Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:16Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:06Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:06 crc kubenswrapper[4813]: I1201 08:41:06.514324 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:06Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:06 crc kubenswrapper[4813]: I1201 08:41:06.521616 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:06 crc kubenswrapper[4813]: I1201 08:41:06.521669 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:06 crc kubenswrapper[4813]: I1201 08:41:06.521690 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:06 crc kubenswrapper[4813]: I1201 08:41:06.521708 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:06 crc kubenswrapper[4813]: I1201 08:41:06.521719 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:06Z","lastTransitionTime":"2025-12-01T08:41:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:06 crc kubenswrapper[4813]: I1201 08:41:06.526111 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-c576v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"17292cd0-1eac-40b5-a60e-67167c57409c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a42206c914d0493ca74788ea1a941de80920ab7b810fb95227cd9b8711e8f4b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7zwfx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:41Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-c576v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:06Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:06 crc kubenswrapper[4813]: I1201 08:41:06.548274 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-j4pcc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b7505e14-3560-4d0d-be2c-d6a7a5953865\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ccc30a63dc1c23ff9880fb04af1f31f0656b2696d359988e23deed23989f480a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://944977e9722025088720107530d0c91fc693593f380e4b49678f28a56da180cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc32d930ab23f30103d8fa42291909cbec89ebb75c5c53cd783cd27fe1a3f09e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f19c67b1c6c80c7f3223512102810d46ede1445464d21ae3bbea83424539b2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d984c90302e57f3e5a6f3ca3613b415db04ac74a5a228582f1b52ef1d59f93e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce8963bf1e9ee66926907aaf3a03d18977f4d0a1983d7121b62fbc8e02a37e4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8790f4730881303e7743498b66fc17b4b0df5b6620c8ca640ad7b9c90e6898d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8790f4730881303e7743498b66fc17b4b0df5b6620c8ca640ad7b9c90e6898d2\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T08:40:56Z\\\",\\\"message\\\":\\\"I1201 08:40:56.782175 6307 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1201 08:40:56.782219 6307 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1201 08:40:56.782251 6307 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1201 08:40:56.782302 6307 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1201 08:40:56.782342 6307 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1201 08:40:56.782379 6307 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1201 08:40:56.782414 6307 factory.go:656] Stopping watch factory\\\\nI1201 08:40:56.782444 6307 ovnkube.go:599] Stopped ovnkube\\\\nI1201 08:40:56.782487 6307 handler.go:208] Removed *v1.Node event handler 7\\\\nI1201 08:40:56.782514 6307 handler.go:208] Removed *v1.Node event handler 2\\\\nI1201 08:40:56.782539 6307 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1201 08:40:56.782564 6307 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1201 08:40:56.782591 6307 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1201 08:40:56.782650 6307 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1201 08:40:56.782670 6307 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1201 08:40:56.782663 6307 handler.go:208] Removed *v1.EgressFirewa\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:55Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-j4pcc_openshift-ovn-kubernetes(b7505e14-3560-4d0d-be2c-d6a7a5953865)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b45454b440c5108fa150a6b856afa1778ba6e9f89633433cb3117b68f00f72c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd67634e0c806323334495401a4e20b48fea4bc5013acd973d2abd3d70870f99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd67634e0c806323334495401a4e20b48fea4bc5013acd973d2abd3d70870f99\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:41Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-j4pcc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:06Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:06 crc kubenswrapper[4813]: I1201 08:41:06.564029 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-h7lvd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d431b613-a91b-4fff-88b9-447804766289\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a32c51c943fcf6ada7e3f614667eb023e29d94c11128d26c1fecca37b03963b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://495593f99e92a9c28100a95e27d06ff5847008c5d996366cb474bedd9b69d5ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://495593f99e92a9c28100a95e27d06ff5847008c5d996366cb474bedd9b69d5ef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://19358f83a6c52957f959fe935a6843fdeef59905cc395e6e52a1fb4a670ac193\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://19358f83a6c52957f959fe935a6843fdeef59905cc395e6e52a1fb4a670ac193\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f089c115d916e339994879d224e3db27642b99aaabf59bcafb2979549052945\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9f089c115d916e339994879d224e3db27642b99aaabf59bcafb2979549052945\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb66f224462305c2005b31ef7bb9ed01efde8101e18518dd42d1dfb0df765bb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb66f224462305c2005b31ef7bb9ed01efde8101e18518dd42d1dfb0df765bb9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c268236c9a155c9b1f47c48d30dbd23948ceb2545592b40f356c60ea1d12e8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c268236c9a155c9b1f47c48d30dbd23948ceb2545592b40f356c60ea1d12e8a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://458aedca5f186e9aaed7e4ee29fd5247221491888b130f55c1fe48d88d565f8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://458aedca5f186e9aaed7e4ee29fd5247221491888b130f55c1fe48d88d565f8f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:41Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-h7lvd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:06Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:06 crc kubenswrapper[4813]: I1201 08:41:06.576912 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6fb5a648-ae4f-4e46-a530-007e9bf84abb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e24ebfc52dcb28368a2d436e7d9ab78f07f2aec2a4c71405a6c7d3301eee96a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33898118bbf639c8e6eae5aafa10386b95795512444ed5f535db3c663f169df2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9902c00bf7738686c93f62e1772a484d05715d74b75ee98c84c8f23ed383824b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3657b47ec9f1563c83bfbe3d728c8fd123f7078e219d251b1874269e42c31eea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:16Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:06Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:06 crc kubenswrapper[4813]: I1201 08:41:06.591508 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0023b0e358d75aacc889ce364846dda89ea513832c87fa4b066aea7dac57e8a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:06Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:06 crc kubenswrapper[4813]: I1201 08:41:06.604428 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:06Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:06 crc kubenswrapper[4813]: I1201 08:41:06.617061 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c7574f892d5e9142296b26a0b484eb6759ab53f571139fb6e23b93dc3da05780\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:06Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:06 crc kubenswrapper[4813]: I1201 08:41:06.624953 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:06 crc kubenswrapper[4813]: I1201 08:41:06.625016 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:06 crc kubenswrapper[4813]: I1201 08:41:06.625028 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:06 crc kubenswrapper[4813]: I1201 08:41:06.625044 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:06 crc kubenswrapper[4813]: I1201 08:41:06.625054 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:06Z","lastTransitionTime":"2025-12-01T08:41:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:06 crc kubenswrapper[4813]: I1201 08:41:06.630612 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:06Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:06 crc kubenswrapper[4813]: I1201 08:41:06.640170 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-xpnlc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"22da1fcf-0c57-4708-90d8-77692af17344\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4a36d213bdad2da91d0d33f400d5a8248befb68d5b94d19347873a93c6712a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xc44d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-xpnlc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:06Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:06 crc kubenswrapper[4813]: I1201 08:41:06.650733 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-9wn4q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3d423402-e08f-42e9-b60c-6eb9bf2f6d48\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3049b842978c6580e1033a442a37766a2e76e278aface9a89fc0b2c57faf0df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rzxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c915733072b32a63237e2401cbf84b059aca396c2c0db884f0f439d599a35d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rzxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:41Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-9wn4q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:06Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:06 crc kubenswrapper[4813]: I1201 08:41:06.661667 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-8pg2p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dd53b38d-d9dd-4f98-9d06-dd95f1a4e3a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a214d71293e4e09573c665524d880fc319dfb1c0c6d95d11d78f253865660fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnjhq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ddf9d5989e4195b3074de08f99d53b48066f53fe93ce204145b3c714996fc57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnjhq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:53Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-8pg2p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:06Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:06 crc kubenswrapper[4813]: I1201 08:41:06.674196 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"490e526a-902b-4aff-b547-0cefea562764\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb48bd28fd3d29781695a5ae630f4913ef2745e97fbe7d463cca39a31e43ba0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74f8cf73f07b4cde260af429d26413e4e6226e7b177b493dd9c0e5c043ab6e1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3aaf696cefbe398bb6b26acfc5be43bda7e2be18218c8060840350c4106e66a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53de1f0682f36aa83ad8e5c226b95daf9f97df4ae154ebfc2fc287f9f238a4a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53de1f0682f36aa83ad8e5c226b95daf9f97df4ae154ebfc2fc287f9f238a4a8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:17Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:16Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:06Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:06 crc kubenswrapper[4813]: I1201 08:41:06.686614 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:06Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:06 crc kubenswrapper[4813]: I1201 08:41:06.696474 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-c576v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"17292cd0-1eac-40b5-a60e-67167c57409c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a42206c914d0493ca74788ea1a941de80920ab7b810fb95227cd9b8711e8f4b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7zwfx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:41Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-c576v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:06Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:06 crc kubenswrapper[4813]: I1201 08:41:06.715682 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-j4pcc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b7505e14-3560-4d0d-be2c-d6a7a5953865\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ccc30a63dc1c23ff9880fb04af1f31f0656b2696d359988e23deed23989f480a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://944977e9722025088720107530d0c91fc693593f380e4b49678f28a56da180cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc32d930ab23f30103d8fa42291909cbec89ebb75c5c53cd783cd27fe1a3f09e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f19c67b1c6c80c7f3223512102810d46ede1445464d21ae3bbea83424539b2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d984c90302e57f3e5a6f3ca3613b415db04ac74a5a228582f1b52ef1d59f93e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce8963bf1e9ee66926907aaf3a03d18977f4d0a1983d7121b62fbc8e02a37e4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8790f4730881303e7743498b66fc17b4b0df5b6620c8ca640ad7b9c90e6898d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8790f4730881303e7743498b66fc17b4b0df5b6620c8ca640ad7b9c90e6898d2\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T08:40:56Z\\\",\\\"message\\\":\\\"I1201 08:40:56.782175 6307 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1201 08:40:56.782219 6307 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1201 08:40:56.782251 6307 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1201 08:40:56.782302 6307 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1201 08:40:56.782342 6307 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1201 08:40:56.782379 6307 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1201 08:40:56.782414 6307 factory.go:656] Stopping watch factory\\\\nI1201 08:40:56.782444 6307 ovnkube.go:599] Stopped ovnkube\\\\nI1201 08:40:56.782487 6307 handler.go:208] Removed *v1.Node event handler 7\\\\nI1201 08:40:56.782514 6307 handler.go:208] Removed *v1.Node event handler 2\\\\nI1201 08:40:56.782539 6307 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1201 08:40:56.782564 6307 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1201 08:40:56.782591 6307 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1201 08:40:56.782650 6307 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1201 08:40:56.782670 6307 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1201 08:40:56.782663 6307 handler.go:208] Removed *v1.EgressFirewa\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:55Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-j4pcc_openshift-ovn-kubernetes(b7505e14-3560-4d0d-be2c-d6a7a5953865)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b45454b440c5108fa150a6b856afa1778ba6e9f89633433cb3117b68f00f72c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd67634e0c806323334495401a4e20b48fea4bc5013acd973d2abd3d70870f99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd67634e0c806323334495401a4e20b48fea4bc5013acd973d2abd3d70870f99\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:41Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-j4pcc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:06Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:06 crc kubenswrapper[4813]: I1201 08:41:06.727447 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:06 crc kubenswrapper[4813]: I1201 08:41:06.727494 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:06 crc kubenswrapper[4813]: I1201 08:41:06.727503 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:06 crc kubenswrapper[4813]: I1201 08:41:06.727518 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:06 crc kubenswrapper[4813]: I1201 08:41:06.727527 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:06Z","lastTransitionTime":"2025-12-01T08:41:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:06 crc kubenswrapper[4813]: I1201 08:41:06.729049 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-h7lvd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d431b613-a91b-4fff-88b9-447804766289\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a32c51c943fcf6ada7e3f614667eb023e29d94c11128d26c1fecca37b03963b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://495593f99e92a9c28100a95e27d06ff5847008c5d996366cb474bedd9b69d5ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://495593f99e92a9c28100a95e27d06ff5847008c5d996366cb474bedd9b69d5ef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://19358f83a6c52957f959fe935a6843fdeef59905cc395e6e52a1fb4a670ac193\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://19358f83a6c52957f959fe935a6843fdeef59905cc395e6e52a1fb4a670ac193\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f089c115d916e339994879d224e3db27642b99aaabf59bcafb2979549052945\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9f089c115d916e339994879d224e3db27642b99aaabf59bcafb2979549052945\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb66f224462305c2005b31ef7bb9ed01efde8101e18518dd42d1dfb0df765bb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb66f224462305c2005b31ef7bb9ed01efde8101e18518dd42d1dfb0df765bb9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c268236c9a155c9b1f47c48d30dbd23948ceb2545592b40f356c60ea1d12e8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c268236c9a155c9b1f47c48d30dbd23948ceb2545592b40f356c60ea1d12e8a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://458aedca5f186e9aaed7e4ee29fd5247221491888b130f55c1fe48d88d565f8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://458aedca5f186e9aaed7e4ee29fd5247221491888b130f55c1fe48d88d565f8f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:41Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-h7lvd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:06Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:06 crc kubenswrapper[4813]: I1201 08:41:06.739383 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-xpnlc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"22da1fcf-0c57-4708-90d8-77692af17344\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4a36d213bdad2da91d0d33f400d5a8248befb68d5b94d19347873a93c6712a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xc44d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-xpnlc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:06Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:06 crc kubenswrapper[4813]: I1201 08:41:06.751772 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6fb5a648-ae4f-4e46-a530-007e9bf84abb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e24ebfc52dcb28368a2d436e7d9ab78f07f2aec2a4c71405a6c7d3301eee96a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33898118bbf639c8e6eae5aafa10386b95795512444ed5f535db3c663f169df2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9902c00bf7738686c93f62e1772a484d05715d74b75ee98c84c8f23ed383824b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3657b47ec9f1563c83bfbe3d728c8fd123f7078e219d251b1874269e42c31eea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:16Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:06Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:06 crc kubenswrapper[4813]: I1201 08:41:06.764697 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0023b0e358d75aacc889ce364846dda89ea513832c87fa4b066aea7dac57e8a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:06Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:06 crc kubenswrapper[4813]: I1201 08:41:06.777404 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:06Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:06 crc kubenswrapper[4813]: I1201 08:41:06.788480 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c7574f892d5e9142296b26a0b484eb6759ab53f571139fb6e23b93dc3da05780\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:06Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:06 crc kubenswrapper[4813]: I1201 08:41:06.802296 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:06Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:06 crc kubenswrapper[4813]: I1201 08:41:06.816031 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"349332bd-6b4d-4306-84ae-07019ca3860f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d441a348087ae3cadda1ad8a7d1c4dff02cf3f6e1d346c294b278fdd9131edcc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://da67635daf6d9e32937f4d5d4ceeee3f8f09ae4fd2dd26c3e85be1286c5949de\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e97356ee274ea78d9f237ffbc756abb6ac8df0dbac244bf0a86ed215b8cbf43\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8dc6feb9817f4f356b8e36af43ec3070774f312fbbb4909fa05b8b3e8dd53582\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a0404f5a5ed5a9750d4c479b608e68e8960b448f455d5af9d219aa5edaeb789b\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"message\\\":\\\"\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764578421\\\\\\\\\\\\\\\" (2025-12-01 08:40:21 +0000 UTC to 2025-12-31 08:40:22 +0000 UTC (now=2025-12-01 08:40:38.066068451 +0000 UTC))\\\\\\\"\\\\nI1201 08:40:38.066144 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1201 08:40:38.066166 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1201 08:40:38.066220 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1201 08:40:38.066232 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1201 08:40:38.066338 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764578432\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764578432\\\\\\\\\\\\\\\" (2025-12-01 07:40:32 +0000 UTC to 2026-12-01 07:40:32 +0000 UTC (now=2025-12-01 08:40:38.066301217 +0000 UTC))\\\\\\\"\\\\nI1201 08:40:38.066357 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1201 08:40:38.066385 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1201 08:40:38.066402 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1201 08:40:38.066414 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-63890811/tls.crt::/tmp/serving-cert-63890811/tls.key\\\\\\\"\\\\nI1201 08:40:38.066419 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1201 08:40:38.066572 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1201 08:40:38.067159 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7354f9337d42115bb1d974bf3772f2d03365acba57f2155135c7a0bbab39879e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:19Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05492df5703392cdf0932d6bf66025b18e53bd05fa0857a7cb2ad70979cbdc3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05492df5703392cdf0932d6bf66025b18e53bd05fa0857a7cb2ad70979cbdc3e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:16Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:06Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:06 crc kubenswrapper[4813]: I1201 08:41:06.829020 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3264b3af996af452df2953c59bea6deeb5ea6af1b119ce89f7334595c27e70da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9e879b1f563bd66383d58fd7b435a8a81bd50cb387dc1fabb6fc47d5a3fa468\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:06Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:06 crc kubenswrapper[4813]: I1201 08:41:06.830201 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:06 crc kubenswrapper[4813]: I1201 08:41:06.830242 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:06 crc kubenswrapper[4813]: I1201 08:41:06.830256 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:06 crc kubenswrapper[4813]: I1201 08:41:06.830274 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:06 crc kubenswrapper[4813]: I1201 08:41:06.830285 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:06Z","lastTransitionTime":"2025-12-01T08:41:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:06 crc kubenswrapper[4813]: I1201 08:41:06.841974 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-b6qqv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9656a03-0096-4c89-b0da-e37e4103611b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4195aebc1b2e388d480920a7f72257f6660afc99ad8c1a09478ef445739b272\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cjvnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:41Z\\\"}}\" for pod \"openshift-multus\"/\"multus-b6qqv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:06Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:06 crc kubenswrapper[4813]: I1201 08:41:06.854002 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-96mbr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0789e3f7-5b08-4c04-8596-9ba29f50cdd0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xt4fj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xt4fj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:55Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-96mbr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:06Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:06 crc kubenswrapper[4813]: I1201 08:41:06.932848 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:06 crc kubenswrapper[4813]: I1201 08:41:06.933366 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:06 crc kubenswrapper[4813]: I1201 08:41:06.933476 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:06 crc kubenswrapper[4813]: I1201 08:41:06.933559 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:06 crc kubenswrapper[4813]: I1201 08:41:06.933634 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:06Z","lastTransitionTime":"2025-12-01T08:41:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:07 crc kubenswrapper[4813]: I1201 08:41:07.036563 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:07 crc kubenswrapper[4813]: I1201 08:41:07.036840 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:07 crc kubenswrapper[4813]: I1201 08:41:07.036911 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:07 crc kubenswrapper[4813]: I1201 08:41:07.037002 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:07 crc kubenswrapper[4813]: I1201 08:41:07.037081 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:07Z","lastTransitionTime":"2025-12-01T08:41:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:07 crc kubenswrapper[4813]: I1201 08:41:07.179781 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:07 crc kubenswrapper[4813]: I1201 08:41:07.179819 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:07 crc kubenswrapper[4813]: I1201 08:41:07.179835 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:07 crc kubenswrapper[4813]: I1201 08:41:07.179852 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:07 crc kubenswrapper[4813]: I1201 08:41:07.179864 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:07Z","lastTransitionTime":"2025-12-01T08:41:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:07 crc kubenswrapper[4813]: I1201 08:41:07.283441 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:07 crc kubenswrapper[4813]: I1201 08:41:07.283508 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:07 crc kubenswrapper[4813]: I1201 08:41:07.283520 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:07 crc kubenswrapper[4813]: I1201 08:41:07.283541 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:07 crc kubenswrapper[4813]: I1201 08:41:07.283553 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:07Z","lastTransitionTime":"2025-12-01T08:41:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:07 crc kubenswrapper[4813]: I1201 08:41:07.386675 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:07 crc kubenswrapper[4813]: I1201 08:41:07.386735 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:07 crc kubenswrapper[4813]: I1201 08:41:07.386753 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:07 crc kubenswrapper[4813]: I1201 08:41:07.386779 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:07 crc kubenswrapper[4813]: I1201 08:41:07.386794 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:07Z","lastTransitionTime":"2025-12-01T08:41:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:07 crc kubenswrapper[4813]: I1201 08:41:07.393041 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 08:41:07 crc kubenswrapper[4813]: I1201 08:41:07.393048 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-96mbr" Dec 01 08:41:07 crc kubenswrapper[4813]: E1201 08:41:07.393219 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 08:41:07 crc kubenswrapper[4813]: E1201 08:41:07.393320 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-96mbr" podUID="0789e3f7-5b08-4c04-8596-9ba29f50cdd0" Dec 01 08:41:07 crc kubenswrapper[4813]: I1201 08:41:07.489224 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:07 crc kubenswrapper[4813]: I1201 08:41:07.489280 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:07 crc kubenswrapper[4813]: I1201 08:41:07.489296 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:07 crc kubenswrapper[4813]: I1201 08:41:07.489320 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:07 crc kubenswrapper[4813]: I1201 08:41:07.489337 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:07Z","lastTransitionTime":"2025-12-01T08:41:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:07 crc kubenswrapper[4813]: I1201 08:41:07.591352 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:07 crc kubenswrapper[4813]: I1201 08:41:07.591404 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:07 crc kubenswrapper[4813]: I1201 08:41:07.591414 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:07 crc kubenswrapper[4813]: I1201 08:41:07.591429 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:07 crc kubenswrapper[4813]: I1201 08:41:07.591441 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:07Z","lastTransitionTime":"2025-12-01T08:41:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:07 crc kubenswrapper[4813]: I1201 08:41:07.694757 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:07 crc kubenswrapper[4813]: I1201 08:41:07.694855 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:07 crc kubenswrapper[4813]: I1201 08:41:07.694895 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:07 crc kubenswrapper[4813]: I1201 08:41:07.694927 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:07 crc kubenswrapper[4813]: I1201 08:41:07.694952 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:07Z","lastTransitionTime":"2025-12-01T08:41:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:07 crc kubenswrapper[4813]: I1201 08:41:07.799472 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:07 crc kubenswrapper[4813]: I1201 08:41:07.799553 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:07 crc kubenswrapper[4813]: I1201 08:41:07.799579 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:07 crc kubenswrapper[4813]: I1201 08:41:07.799611 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:07 crc kubenswrapper[4813]: I1201 08:41:07.799648 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:07Z","lastTransitionTime":"2025-12-01T08:41:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:07 crc kubenswrapper[4813]: I1201 08:41:07.902618 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:07 crc kubenswrapper[4813]: I1201 08:41:07.902667 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:07 crc kubenswrapper[4813]: I1201 08:41:07.902676 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:07 crc kubenswrapper[4813]: I1201 08:41:07.902691 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:07 crc kubenswrapper[4813]: I1201 08:41:07.902701 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:07Z","lastTransitionTime":"2025-12-01T08:41:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:08 crc kubenswrapper[4813]: I1201 08:41:08.005261 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:08 crc kubenswrapper[4813]: I1201 08:41:08.005328 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:08 crc kubenswrapper[4813]: I1201 08:41:08.005346 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:08 crc kubenswrapper[4813]: I1201 08:41:08.005371 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:08 crc kubenswrapper[4813]: I1201 08:41:08.005387 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:08Z","lastTransitionTime":"2025-12-01T08:41:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:08 crc kubenswrapper[4813]: I1201 08:41:08.107697 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:08 crc kubenswrapper[4813]: I1201 08:41:08.107744 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:08 crc kubenswrapper[4813]: I1201 08:41:08.107757 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:08 crc kubenswrapper[4813]: I1201 08:41:08.107775 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:08 crc kubenswrapper[4813]: I1201 08:41:08.107787 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:08Z","lastTransitionTime":"2025-12-01T08:41:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:08 crc kubenswrapper[4813]: I1201 08:41:08.211258 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:08 crc kubenswrapper[4813]: I1201 08:41:08.211334 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:08 crc kubenswrapper[4813]: I1201 08:41:08.211357 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:08 crc kubenswrapper[4813]: I1201 08:41:08.211433 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:08 crc kubenswrapper[4813]: I1201 08:41:08.211461 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:08Z","lastTransitionTime":"2025-12-01T08:41:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:08 crc kubenswrapper[4813]: I1201 08:41:08.314901 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:08 crc kubenswrapper[4813]: I1201 08:41:08.314946 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:08 crc kubenswrapper[4813]: I1201 08:41:08.314959 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:08 crc kubenswrapper[4813]: I1201 08:41:08.315029 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:08 crc kubenswrapper[4813]: I1201 08:41:08.315041 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:08Z","lastTransitionTime":"2025-12-01T08:41:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:08 crc kubenswrapper[4813]: I1201 08:41:08.393806 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 08:41:08 crc kubenswrapper[4813]: I1201 08:41:08.393909 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 08:41:08 crc kubenswrapper[4813]: E1201 08:41:08.393995 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 08:41:08 crc kubenswrapper[4813]: E1201 08:41:08.394068 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 08:41:08 crc kubenswrapper[4813]: I1201 08:41:08.417508 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:08 crc kubenswrapper[4813]: I1201 08:41:08.417583 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:08 crc kubenswrapper[4813]: I1201 08:41:08.417608 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:08 crc kubenswrapper[4813]: I1201 08:41:08.417637 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:08 crc kubenswrapper[4813]: I1201 08:41:08.417658 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:08Z","lastTransitionTime":"2025-12-01T08:41:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:08 crc kubenswrapper[4813]: I1201 08:41:08.520073 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:08 crc kubenswrapper[4813]: I1201 08:41:08.520164 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:08 crc kubenswrapper[4813]: I1201 08:41:08.520187 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:08 crc kubenswrapper[4813]: I1201 08:41:08.520212 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:08 crc kubenswrapper[4813]: I1201 08:41:08.520230 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:08Z","lastTransitionTime":"2025-12-01T08:41:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:08 crc kubenswrapper[4813]: I1201 08:41:08.624308 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:08 crc kubenswrapper[4813]: I1201 08:41:08.624410 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:08 crc kubenswrapper[4813]: I1201 08:41:08.624474 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:08 crc kubenswrapper[4813]: I1201 08:41:08.624498 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:08 crc kubenswrapper[4813]: I1201 08:41:08.624513 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:08Z","lastTransitionTime":"2025-12-01T08:41:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:08 crc kubenswrapper[4813]: I1201 08:41:08.727416 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:08 crc kubenswrapper[4813]: I1201 08:41:08.727459 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:08 crc kubenswrapper[4813]: I1201 08:41:08.727470 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:08 crc kubenswrapper[4813]: I1201 08:41:08.727485 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:08 crc kubenswrapper[4813]: I1201 08:41:08.727496 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:08Z","lastTransitionTime":"2025-12-01T08:41:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:08 crc kubenswrapper[4813]: I1201 08:41:08.830358 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:08 crc kubenswrapper[4813]: I1201 08:41:08.830425 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:08 crc kubenswrapper[4813]: I1201 08:41:08.830441 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:08 crc kubenswrapper[4813]: I1201 08:41:08.830467 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:08 crc kubenswrapper[4813]: I1201 08:41:08.830483 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:08Z","lastTransitionTime":"2025-12-01T08:41:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:08 crc kubenswrapper[4813]: I1201 08:41:08.943937 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:08 crc kubenswrapper[4813]: I1201 08:41:08.944059 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:08 crc kubenswrapper[4813]: I1201 08:41:08.944072 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:08 crc kubenswrapper[4813]: I1201 08:41:08.944088 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:08 crc kubenswrapper[4813]: I1201 08:41:08.944099 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:08Z","lastTransitionTime":"2025-12-01T08:41:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:09 crc kubenswrapper[4813]: I1201 08:41:09.046933 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:09 crc kubenswrapper[4813]: I1201 08:41:09.046990 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:09 crc kubenswrapper[4813]: I1201 08:41:09.047001 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:09 crc kubenswrapper[4813]: I1201 08:41:09.047016 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:09 crc kubenswrapper[4813]: I1201 08:41:09.047027 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:09Z","lastTransitionTime":"2025-12-01T08:41:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:09 crc kubenswrapper[4813]: I1201 08:41:09.150867 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:09 crc kubenswrapper[4813]: I1201 08:41:09.150923 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:09 crc kubenswrapper[4813]: I1201 08:41:09.150945 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:09 crc kubenswrapper[4813]: I1201 08:41:09.151012 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:09 crc kubenswrapper[4813]: I1201 08:41:09.151035 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:09Z","lastTransitionTime":"2025-12-01T08:41:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:09 crc kubenswrapper[4813]: I1201 08:41:09.253809 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:09 crc kubenswrapper[4813]: I1201 08:41:09.253845 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:09 crc kubenswrapper[4813]: I1201 08:41:09.253854 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:09 crc kubenswrapper[4813]: I1201 08:41:09.253867 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:09 crc kubenswrapper[4813]: I1201 08:41:09.253878 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:09Z","lastTransitionTime":"2025-12-01T08:41:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:09 crc kubenswrapper[4813]: I1201 08:41:09.356389 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:09 crc kubenswrapper[4813]: I1201 08:41:09.356442 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:09 crc kubenswrapper[4813]: I1201 08:41:09.356456 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:09 crc kubenswrapper[4813]: I1201 08:41:09.356485 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:09 crc kubenswrapper[4813]: I1201 08:41:09.356509 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:09Z","lastTransitionTime":"2025-12-01T08:41:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:09 crc kubenswrapper[4813]: I1201 08:41:09.392910 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-96mbr" Dec 01 08:41:09 crc kubenswrapper[4813]: I1201 08:41:09.393016 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 08:41:09 crc kubenswrapper[4813]: E1201 08:41:09.393119 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-96mbr" podUID="0789e3f7-5b08-4c04-8596-9ba29f50cdd0" Dec 01 08:41:09 crc kubenswrapper[4813]: E1201 08:41:09.393246 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 08:41:09 crc kubenswrapper[4813]: I1201 08:41:09.459762 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:09 crc kubenswrapper[4813]: I1201 08:41:09.459813 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:09 crc kubenswrapper[4813]: I1201 08:41:09.459824 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:09 crc kubenswrapper[4813]: I1201 08:41:09.459841 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:09 crc kubenswrapper[4813]: I1201 08:41:09.459855 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:09Z","lastTransitionTime":"2025-12-01T08:41:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:09 crc kubenswrapper[4813]: I1201 08:41:09.563787 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:09 crc kubenswrapper[4813]: I1201 08:41:09.563855 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:09 crc kubenswrapper[4813]: I1201 08:41:09.563877 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:09 crc kubenswrapper[4813]: I1201 08:41:09.563902 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:09 crc kubenswrapper[4813]: I1201 08:41:09.563917 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:09Z","lastTransitionTime":"2025-12-01T08:41:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:09 crc kubenswrapper[4813]: I1201 08:41:09.667036 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:09 crc kubenswrapper[4813]: I1201 08:41:09.667372 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:09 crc kubenswrapper[4813]: I1201 08:41:09.667480 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:09 crc kubenswrapper[4813]: I1201 08:41:09.667623 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:09 crc kubenswrapper[4813]: I1201 08:41:09.667726 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:09Z","lastTransitionTime":"2025-12-01T08:41:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:09 crc kubenswrapper[4813]: I1201 08:41:09.771283 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:09 crc kubenswrapper[4813]: I1201 08:41:09.771619 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:09 crc kubenswrapper[4813]: I1201 08:41:09.771760 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:09 crc kubenswrapper[4813]: I1201 08:41:09.771907 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:09 crc kubenswrapper[4813]: I1201 08:41:09.772114 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:09Z","lastTransitionTime":"2025-12-01T08:41:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:09 crc kubenswrapper[4813]: I1201 08:41:09.875925 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:09 crc kubenswrapper[4813]: I1201 08:41:09.876015 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:09 crc kubenswrapper[4813]: I1201 08:41:09.876035 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:09 crc kubenswrapper[4813]: I1201 08:41:09.876061 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:09 crc kubenswrapper[4813]: I1201 08:41:09.876078 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:09Z","lastTransitionTime":"2025-12-01T08:41:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:09 crc kubenswrapper[4813]: I1201 08:41:09.979835 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:09 crc kubenswrapper[4813]: I1201 08:41:09.979941 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:09 crc kubenswrapper[4813]: I1201 08:41:09.979997 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:09 crc kubenswrapper[4813]: I1201 08:41:09.980043 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:09 crc kubenswrapper[4813]: I1201 08:41:09.980077 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:09Z","lastTransitionTime":"2025-12-01T08:41:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:10 crc kubenswrapper[4813]: I1201 08:41:10.083578 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:10 crc kubenswrapper[4813]: I1201 08:41:10.083642 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:10 crc kubenswrapper[4813]: I1201 08:41:10.083666 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:10 crc kubenswrapper[4813]: I1201 08:41:10.083702 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:10 crc kubenswrapper[4813]: I1201 08:41:10.083726 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:10Z","lastTransitionTime":"2025-12-01T08:41:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:10 crc kubenswrapper[4813]: I1201 08:41:10.187852 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:10 crc kubenswrapper[4813]: I1201 08:41:10.187930 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:10 crc kubenswrapper[4813]: I1201 08:41:10.187959 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:10 crc kubenswrapper[4813]: I1201 08:41:10.188028 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:10 crc kubenswrapper[4813]: I1201 08:41:10.188048 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:10Z","lastTransitionTime":"2025-12-01T08:41:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:10 crc kubenswrapper[4813]: I1201 08:41:10.207069 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 08:41:10 crc kubenswrapper[4813]: E1201 08:41:10.207506 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 08:41:42.207463005 +0000 UTC m=+86.830384631 (durationBeforeRetry 32s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:41:10 crc kubenswrapper[4813]: I1201 08:41:10.207822 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 08:41:10 crc kubenswrapper[4813]: E1201 08:41:10.208137 4813 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 01 08:41:10 crc kubenswrapper[4813]: E1201 08:41:10.208256 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-01 08:41:42.208234807 +0000 UTC m=+86.831156453 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 01 08:41:10 crc kubenswrapper[4813]: I1201 08:41:10.208147 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 08:41:10 crc kubenswrapper[4813]: E1201 08:41:10.208739 4813 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 01 08:41:10 crc kubenswrapper[4813]: E1201 08:41:10.209091 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-01 08:41:42.209053781 +0000 UTC m=+86.831975427 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 01 08:41:10 crc kubenswrapper[4813]: I1201 08:41:10.229741 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:10 crc kubenswrapper[4813]: I1201 08:41:10.230186 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:10 crc kubenswrapper[4813]: I1201 08:41:10.230229 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:10 crc kubenswrapper[4813]: I1201 08:41:10.230271 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:10 crc kubenswrapper[4813]: I1201 08:41:10.230306 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:10Z","lastTransitionTime":"2025-12-01T08:41:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:10 crc kubenswrapper[4813]: E1201 08:41:10.255469 4813 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:41:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:41:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:10Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:41:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:41:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:10Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cd271ba1-a5c5-44e0-b03e-cfd1f5e8cb76\\\",\\\"systemUUID\\\":\\\"906e3783-eb6d-4f21-889c-7989aeaa25ba\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:10Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:10 crc kubenswrapper[4813]: I1201 08:41:10.261573 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:10 crc kubenswrapper[4813]: I1201 08:41:10.261635 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:10 crc kubenswrapper[4813]: I1201 08:41:10.261653 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:10 crc kubenswrapper[4813]: I1201 08:41:10.261679 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:10 crc kubenswrapper[4813]: I1201 08:41:10.261696 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:10Z","lastTransitionTime":"2025-12-01T08:41:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:10 crc kubenswrapper[4813]: E1201 08:41:10.287085 4813 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:41:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:41:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:10Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:41:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:41:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:10Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cd271ba1-a5c5-44e0-b03e-cfd1f5e8cb76\\\",\\\"systemUUID\\\":\\\"906e3783-eb6d-4f21-889c-7989aeaa25ba\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:10Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:10 crc kubenswrapper[4813]: I1201 08:41:10.292801 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:10 crc kubenswrapper[4813]: I1201 08:41:10.293146 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:10 crc kubenswrapper[4813]: I1201 08:41:10.293382 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:10 crc kubenswrapper[4813]: I1201 08:41:10.293606 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:10 crc kubenswrapper[4813]: I1201 08:41:10.293792 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:10Z","lastTransitionTime":"2025-12-01T08:41:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:10 crc kubenswrapper[4813]: I1201 08:41:10.310242 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 08:41:10 crc kubenswrapper[4813]: I1201 08:41:10.310524 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 08:41:10 crc kubenswrapper[4813]: E1201 08:41:10.310542 4813 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 01 08:41:10 crc kubenswrapper[4813]: E1201 08:41:10.310873 4813 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 01 08:41:10 crc kubenswrapper[4813]: E1201 08:41:10.311117 4813 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 08:41:10 crc kubenswrapper[4813]: E1201 08:41:10.311398 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-01 08:41:42.311370664 +0000 UTC m=+86.934292290 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 08:41:10 crc kubenswrapper[4813]: E1201 08:41:10.310663 4813 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 01 08:41:10 crc kubenswrapper[4813]: E1201 08:41:10.311780 4813 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 01 08:41:10 crc kubenswrapper[4813]: E1201 08:41:10.311919 4813 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 08:41:10 crc kubenswrapper[4813]: E1201 08:41:10.312547 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-01 08:41:42.312523327 +0000 UTC m=+86.935444953 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 08:41:10 crc kubenswrapper[4813]: E1201 08:41:10.319631 4813 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:41:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:41:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:10Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:41:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:41:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:10Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cd271ba1-a5c5-44e0-b03e-cfd1f5e8cb76\\\",\\\"systemUUID\\\":\\\"906e3783-eb6d-4f21-889c-7989aeaa25ba\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:10Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:10 crc kubenswrapper[4813]: I1201 08:41:10.325534 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:10 crc kubenswrapper[4813]: I1201 08:41:10.325698 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:10 crc kubenswrapper[4813]: I1201 08:41:10.325817 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:10 crc kubenswrapper[4813]: I1201 08:41:10.326078 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:10 crc kubenswrapper[4813]: I1201 08:41:10.326213 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:10Z","lastTransitionTime":"2025-12-01T08:41:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:10 crc kubenswrapper[4813]: E1201 08:41:10.374418 4813 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:41:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:41:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:10Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:41:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:41:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:10Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cd271ba1-a5c5-44e0-b03e-cfd1f5e8cb76\\\",\\\"systemUUID\\\":\\\"906e3783-eb6d-4f21-889c-7989aeaa25ba\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:10Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:10 crc kubenswrapper[4813]: I1201 08:41:10.380688 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:10 crc kubenswrapper[4813]: I1201 08:41:10.380732 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:10 crc kubenswrapper[4813]: I1201 08:41:10.380747 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:10 crc kubenswrapper[4813]: I1201 08:41:10.380767 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:10 crc kubenswrapper[4813]: I1201 08:41:10.380782 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:10Z","lastTransitionTime":"2025-12-01T08:41:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:10 crc kubenswrapper[4813]: I1201 08:41:10.394468 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 08:41:10 crc kubenswrapper[4813]: I1201 08:41:10.394554 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 08:41:10 crc kubenswrapper[4813]: E1201 08:41:10.394686 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 08:41:10 crc kubenswrapper[4813]: E1201 08:41:10.394911 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 08:41:10 crc kubenswrapper[4813]: E1201 08:41:10.407212 4813 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:41:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:41:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:10Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:41:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:41:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:10Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cd271ba1-a5c5-44e0-b03e-cfd1f5e8cb76\\\",\\\"systemUUID\\\":\\\"906e3783-eb6d-4f21-889c-7989aeaa25ba\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:10Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:10 crc kubenswrapper[4813]: E1201 08:41:10.407389 4813 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 01 08:41:10 crc kubenswrapper[4813]: I1201 08:41:10.409210 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:10 crc kubenswrapper[4813]: I1201 08:41:10.409239 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:10 crc kubenswrapper[4813]: I1201 08:41:10.409250 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:10 crc kubenswrapper[4813]: I1201 08:41:10.409267 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:10 crc kubenswrapper[4813]: I1201 08:41:10.409279 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:10Z","lastTransitionTime":"2025-12-01T08:41:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:10 crc kubenswrapper[4813]: I1201 08:41:10.511866 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:10 crc kubenswrapper[4813]: I1201 08:41:10.512195 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:10 crc kubenswrapper[4813]: I1201 08:41:10.512309 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:10 crc kubenswrapper[4813]: I1201 08:41:10.512423 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:10 crc kubenswrapper[4813]: I1201 08:41:10.512522 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:10Z","lastTransitionTime":"2025-12-01T08:41:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:10 crc kubenswrapper[4813]: I1201 08:41:10.615475 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:10 crc kubenswrapper[4813]: I1201 08:41:10.615539 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:10 crc kubenswrapper[4813]: I1201 08:41:10.615577 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:10 crc kubenswrapper[4813]: I1201 08:41:10.615612 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:10 crc kubenswrapper[4813]: I1201 08:41:10.615634 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:10Z","lastTransitionTime":"2025-12-01T08:41:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:10 crc kubenswrapper[4813]: I1201 08:41:10.718921 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:10 crc kubenswrapper[4813]: I1201 08:41:10.718983 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:10 crc kubenswrapper[4813]: I1201 08:41:10.719004 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:10 crc kubenswrapper[4813]: I1201 08:41:10.719025 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:10 crc kubenswrapper[4813]: I1201 08:41:10.719038 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:10Z","lastTransitionTime":"2025-12-01T08:41:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:10 crc kubenswrapper[4813]: I1201 08:41:10.821690 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:10 crc kubenswrapper[4813]: I1201 08:41:10.822160 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:10 crc kubenswrapper[4813]: I1201 08:41:10.822399 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:10 crc kubenswrapper[4813]: I1201 08:41:10.822842 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:10 crc kubenswrapper[4813]: I1201 08:41:10.823066 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:10Z","lastTransitionTime":"2025-12-01T08:41:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:10 crc kubenswrapper[4813]: I1201 08:41:10.926188 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:10 crc kubenswrapper[4813]: I1201 08:41:10.926255 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:10 crc kubenswrapper[4813]: I1201 08:41:10.926273 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:10 crc kubenswrapper[4813]: I1201 08:41:10.926297 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:10 crc kubenswrapper[4813]: I1201 08:41:10.926316 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:10Z","lastTransitionTime":"2025-12-01T08:41:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:11 crc kubenswrapper[4813]: I1201 08:41:11.028908 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:11 crc kubenswrapper[4813]: I1201 08:41:11.028985 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:11 crc kubenswrapper[4813]: I1201 08:41:11.028995 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:11 crc kubenswrapper[4813]: I1201 08:41:11.029013 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:11 crc kubenswrapper[4813]: I1201 08:41:11.029023 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:11Z","lastTransitionTime":"2025-12-01T08:41:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:11 crc kubenswrapper[4813]: I1201 08:41:11.132263 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:11 crc kubenswrapper[4813]: I1201 08:41:11.133001 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:11 crc kubenswrapper[4813]: I1201 08:41:11.133244 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:11 crc kubenswrapper[4813]: I1201 08:41:11.133342 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:11 crc kubenswrapper[4813]: I1201 08:41:11.133437 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:11Z","lastTransitionTime":"2025-12-01T08:41:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:11 crc kubenswrapper[4813]: I1201 08:41:11.236722 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:11 crc kubenswrapper[4813]: I1201 08:41:11.236771 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:11 crc kubenswrapper[4813]: I1201 08:41:11.236784 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:11 crc kubenswrapper[4813]: I1201 08:41:11.236803 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:11 crc kubenswrapper[4813]: I1201 08:41:11.236816 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:11Z","lastTransitionTime":"2025-12-01T08:41:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:11 crc kubenswrapper[4813]: I1201 08:41:11.340413 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:11 crc kubenswrapper[4813]: I1201 08:41:11.340487 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:11 crc kubenswrapper[4813]: I1201 08:41:11.340505 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:11 crc kubenswrapper[4813]: I1201 08:41:11.340534 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:11 crc kubenswrapper[4813]: I1201 08:41:11.340552 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:11Z","lastTransitionTime":"2025-12-01T08:41:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:11 crc kubenswrapper[4813]: I1201 08:41:11.392931 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 08:41:11 crc kubenswrapper[4813]: I1201 08:41:11.392931 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-96mbr" Dec 01 08:41:11 crc kubenswrapper[4813]: E1201 08:41:11.393193 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 08:41:11 crc kubenswrapper[4813]: E1201 08:41:11.393360 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-96mbr" podUID="0789e3f7-5b08-4c04-8596-9ba29f50cdd0" Dec 01 08:41:11 crc kubenswrapper[4813]: I1201 08:41:11.443843 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:11 crc kubenswrapper[4813]: I1201 08:41:11.443891 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:11 crc kubenswrapper[4813]: I1201 08:41:11.443904 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:11 crc kubenswrapper[4813]: I1201 08:41:11.443924 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:11 crc kubenswrapper[4813]: I1201 08:41:11.443936 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:11Z","lastTransitionTime":"2025-12-01T08:41:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:11 crc kubenswrapper[4813]: I1201 08:41:11.546944 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:11 crc kubenswrapper[4813]: I1201 08:41:11.547334 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:11 crc kubenswrapper[4813]: I1201 08:41:11.547536 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:11 crc kubenswrapper[4813]: I1201 08:41:11.547686 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:11 crc kubenswrapper[4813]: I1201 08:41:11.547833 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:11Z","lastTransitionTime":"2025-12-01T08:41:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:11 crc kubenswrapper[4813]: I1201 08:41:11.651353 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:11 crc kubenswrapper[4813]: I1201 08:41:11.651398 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:11 crc kubenswrapper[4813]: I1201 08:41:11.651409 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:11 crc kubenswrapper[4813]: I1201 08:41:11.651428 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:11 crc kubenswrapper[4813]: I1201 08:41:11.651440 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:11Z","lastTransitionTime":"2025-12-01T08:41:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:11 crc kubenswrapper[4813]: I1201 08:41:11.727068 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0789e3f7-5b08-4c04-8596-9ba29f50cdd0-metrics-certs\") pod \"network-metrics-daemon-96mbr\" (UID: \"0789e3f7-5b08-4c04-8596-9ba29f50cdd0\") " pod="openshift-multus/network-metrics-daemon-96mbr" Dec 01 08:41:11 crc kubenswrapper[4813]: E1201 08:41:11.727283 4813 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 01 08:41:11 crc kubenswrapper[4813]: E1201 08:41:11.727705 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/0789e3f7-5b08-4c04-8596-9ba29f50cdd0-metrics-certs podName:0789e3f7-5b08-4c04-8596-9ba29f50cdd0 nodeName:}" failed. No retries permitted until 2025-12-01 08:41:27.727680336 +0000 UTC m=+72.350601932 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/0789e3f7-5b08-4c04-8596-9ba29f50cdd0-metrics-certs") pod "network-metrics-daemon-96mbr" (UID: "0789e3f7-5b08-4c04-8596-9ba29f50cdd0") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 01 08:41:11 crc kubenswrapper[4813]: I1201 08:41:11.757493 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:11 crc kubenswrapper[4813]: I1201 08:41:11.757846 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:11 crc kubenswrapper[4813]: I1201 08:41:11.758050 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:11 crc kubenswrapper[4813]: I1201 08:41:11.758230 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:11 crc kubenswrapper[4813]: I1201 08:41:11.758425 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:11Z","lastTransitionTime":"2025-12-01T08:41:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:11 crc kubenswrapper[4813]: I1201 08:41:11.862228 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:11 crc kubenswrapper[4813]: I1201 08:41:11.862321 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:11 crc kubenswrapper[4813]: I1201 08:41:11.862343 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:11 crc kubenswrapper[4813]: I1201 08:41:11.862372 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:11 crc kubenswrapper[4813]: I1201 08:41:11.862391 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:11Z","lastTransitionTime":"2025-12-01T08:41:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:11 crc kubenswrapper[4813]: I1201 08:41:11.966142 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:11 crc kubenswrapper[4813]: I1201 08:41:11.966208 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:11 crc kubenswrapper[4813]: I1201 08:41:11.966222 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:11 crc kubenswrapper[4813]: I1201 08:41:11.966240 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:11 crc kubenswrapper[4813]: I1201 08:41:11.966254 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:11Z","lastTransitionTime":"2025-12-01T08:41:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:12 crc kubenswrapper[4813]: I1201 08:41:12.068940 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:12 crc kubenswrapper[4813]: I1201 08:41:12.069250 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:12 crc kubenswrapper[4813]: I1201 08:41:12.069346 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:12 crc kubenswrapper[4813]: I1201 08:41:12.069446 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:12 crc kubenswrapper[4813]: I1201 08:41:12.069552 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:12Z","lastTransitionTime":"2025-12-01T08:41:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:12 crc kubenswrapper[4813]: I1201 08:41:12.173490 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:12 crc kubenswrapper[4813]: I1201 08:41:12.173550 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:12 crc kubenswrapper[4813]: I1201 08:41:12.173569 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:12 crc kubenswrapper[4813]: I1201 08:41:12.173595 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:12 crc kubenswrapper[4813]: I1201 08:41:12.173616 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:12Z","lastTransitionTime":"2025-12-01T08:41:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:12 crc kubenswrapper[4813]: I1201 08:41:12.276796 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:12 crc kubenswrapper[4813]: I1201 08:41:12.277167 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:12 crc kubenswrapper[4813]: I1201 08:41:12.277291 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:12 crc kubenswrapper[4813]: I1201 08:41:12.277416 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:12 crc kubenswrapper[4813]: I1201 08:41:12.277533 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:12Z","lastTransitionTime":"2025-12-01T08:41:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:12 crc kubenswrapper[4813]: I1201 08:41:12.381422 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:12 crc kubenswrapper[4813]: I1201 08:41:12.381793 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:12 crc kubenswrapper[4813]: I1201 08:41:12.381936 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:12 crc kubenswrapper[4813]: I1201 08:41:12.382145 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:12 crc kubenswrapper[4813]: I1201 08:41:12.382290 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:12Z","lastTransitionTime":"2025-12-01T08:41:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:12 crc kubenswrapper[4813]: I1201 08:41:12.393586 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 08:41:12 crc kubenswrapper[4813]: I1201 08:41:12.393592 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 08:41:12 crc kubenswrapper[4813]: E1201 08:41:12.393731 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 08:41:12 crc kubenswrapper[4813]: E1201 08:41:12.393819 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 08:41:12 crc kubenswrapper[4813]: I1201 08:41:12.486097 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:12 crc kubenswrapper[4813]: I1201 08:41:12.486148 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:12 crc kubenswrapper[4813]: I1201 08:41:12.486182 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:12 crc kubenswrapper[4813]: I1201 08:41:12.486205 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:12 crc kubenswrapper[4813]: I1201 08:41:12.486220 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:12Z","lastTransitionTime":"2025-12-01T08:41:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:12 crc kubenswrapper[4813]: I1201 08:41:12.588741 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:12 crc kubenswrapper[4813]: I1201 08:41:12.588793 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:12 crc kubenswrapper[4813]: I1201 08:41:12.588808 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:12 crc kubenswrapper[4813]: I1201 08:41:12.588827 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:12 crc kubenswrapper[4813]: I1201 08:41:12.588838 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:12Z","lastTransitionTime":"2025-12-01T08:41:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:12 crc kubenswrapper[4813]: I1201 08:41:12.691568 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:12 crc kubenswrapper[4813]: I1201 08:41:12.691614 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:12 crc kubenswrapper[4813]: I1201 08:41:12.691623 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:12 crc kubenswrapper[4813]: I1201 08:41:12.691639 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:12 crc kubenswrapper[4813]: I1201 08:41:12.691648 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:12Z","lastTransitionTime":"2025-12-01T08:41:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:12 crc kubenswrapper[4813]: I1201 08:41:12.794568 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:12 crc kubenswrapper[4813]: I1201 08:41:12.794606 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:12 crc kubenswrapper[4813]: I1201 08:41:12.794614 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:12 crc kubenswrapper[4813]: I1201 08:41:12.794627 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:12 crc kubenswrapper[4813]: I1201 08:41:12.794637 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:12Z","lastTransitionTime":"2025-12-01T08:41:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:12 crc kubenswrapper[4813]: I1201 08:41:12.897678 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:12 crc kubenswrapper[4813]: I1201 08:41:12.897772 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:12 crc kubenswrapper[4813]: I1201 08:41:12.897796 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:12 crc kubenswrapper[4813]: I1201 08:41:12.897824 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:12 crc kubenswrapper[4813]: I1201 08:41:12.897842 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:12Z","lastTransitionTime":"2025-12-01T08:41:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:13 crc kubenswrapper[4813]: I1201 08:41:13.000663 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:13 crc kubenswrapper[4813]: I1201 08:41:13.000717 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:13 crc kubenswrapper[4813]: I1201 08:41:13.000730 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:13 crc kubenswrapper[4813]: I1201 08:41:13.000748 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:13 crc kubenswrapper[4813]: I1201 08:41:13.000759 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:13Z","lastTransitionTime":"2025-12-01T08:41:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:13 crc kubenswrapper[4813]: I1201 08:41:13.104340 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:13 crc kubenswrapper[4813]: I1201 08:41:13.104397 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:13 crc kubenswrapper[4813]: I1201 08:41:13.104415 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:13 crc kubenswrapper[4813]: I1201 08:41:13.104440 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:13 crc kubenswrapper[4813]: I1201 08:41:13.104459 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:13Z","lastTransitionTime":"2025-12-01T08:41:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:13 crc kubenswrapper[4813]: I1201 08:41:13.208306 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:13 crc kubenswrapper[4813]: I1201 08:41:13.208360 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:13 crc kubenswrapper[4813]: I1201 08:41:13.208373 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:13 crc kubenswrapper[4813]: I1201 08:41:13.208393 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:13 crc kubenswrapper[4813]: I1201 08:41:13.208407 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:13Z","lastTransitionTime":"2025-12-01T08:41:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:13 crc kubenswrapper[4813]: I1201 08:41:13.311983 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:13 crc kubenswrapper[4813]: I1201 08:41:13.312034 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:13 crc kubenswrapper[4813]: I1201 08:41:13.312047 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:13 crc kubenswrapper[4813]: I1201 08:41:13.312065 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:13 crc kubenswrapper[4813]: I1201 08:41:13.312077 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:13Z","lastTransitionTime":"2025-12-01T08:41:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:13 crc kubenswrapper[4813]: I1201 08:41:13.393284 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 08:41:13 crc kubenswrapper[4813]: I1201 08:41:13.393294 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-96mbr" Dec 01 08:41:13 crc kubenswrapper[4813]: E1201 08:41:13.393479 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 08:41:13 crc kubenswrapper[4813]: E1201 08:41:13.393690 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-96mbr" podUID="0789e3f7-5b08-4c04-8596-9ba29f50cdd0" Dec 01 08:41:13 crc kubenswrapper[4813]: I1201 08:41:13.416075 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:13 crc kubenswrapper[4813]: I1201 08:41:13.416170 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:13 crc kubenswrapper[4813]: I1201 08:41:13.416195 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:13 crc kubenswrapper[4813]: I1201 08:41:13.416230 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:13 crc kubenswrapper[4813]: I1201 08:41:13.416252 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:13Z","lastTransitionTime":"2025-12-01T08:41:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:13 crc kubenswrapper[4813]: I1201 08:41:13.519637 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:13 crc kubenswrapper[4813]: I1201 08:41:13.519687 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:13 crc kubenswrapper[4813]: I1201 08:41:13.519698 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:13 crc kubenswrapper[4813]: I1201 08:41:13.519715 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:13 crc kubenswrapper[4813]: I1201 08:41:13.519726 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:13Z","lastTransitionTime":"2025-12-01T08:41:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:13 crc kubenswrapper[4813]: I1201 08:41:13.623004 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:13 crc kubenswrapper[4813]: I1201 08:41:13.623051 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:13 crc kubenswrapper[4813]: I1201 08:41:13.623060 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:13 crc kubenswrapper[4813]: I1201 08:41:13.623076 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:13 crc kubenswrapper[4813]: I1201 08:41:13.623085 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:13Z","lastTransitionTime":"2025-12-01T08:41:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:13 crc kubenswrapper[4813]: I1201 08:41:13.725731 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:13 crc kubenswrapper[4813]: I1201 08:41:13.725838 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:13 crc kubenswrapper[4813]: I1201 08:41:13.725881 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:13 crc kubenswrapper[4813]: I1201 08:41:13.725915 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:13 crc kubenswrapper[4813]: I1201 08:41:13.725937 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:13Z","lastTransitionTime":"2025-12-01T08:41:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:13 crc kubenswrapper[4813]: I1201 08:41:13.828662 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:13 crc kubenswrapper[4813]: I1201 08:41:13.828764 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:13 crc kubenswrapper[4813]: I1201 08:41:13.828799 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:13 crc kubenswrapper[4813]: I1201 08:41:13.828831 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:13 crc kubenswrapper[4813]: I1201 08:41:13.828869 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:13Z","lastTransitionTime":"2025-12-01T08:41:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:13 crc kubenswrapper[4813]: I1201 08:41:13.931916 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:13 crc kubenswrapper[4813]: I1201 08:41:13.931986 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:13 crc kubenswrapper[4813]: I1201 08:41:13.932004 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:13 crc kubenswrapper[4813]: I1201 08:41:13.932022 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:13 crc kubenswrapper[4813]: I1201 08:41:13.932034 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:13Z","lastTransitionTime":"2025-12-01T08:41:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:14 crc kubenswrapper[4813]: I1201 08:41:14.034926 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:14 crc kubenswrapper[4813]: I1201 08:41:14.034975 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:14 crc kubenswrapper[4813]: I1201 08:41:14.034985 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:14 crc kubenswrapper[4813]: I1201 08:41:14.034999 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:14 crc kubenswrapper[4813]: I1201 08:41:14.035011 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:14Z","lastTransitionTime":"2025-12-01T08:41:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:14 crc kubenswrapper[4813]: I1201 08:41:14.136995 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:14 crc kubenswrapper[4813]: I1201 08:41:14.137064 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:14 crc kubenswrapper[4813]: I1201 08:41:14.137076 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:14 crc kubenswrapper[4813]: I1201 08:41:14.137094 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:14 crc kubenswrapper[4813]: I1201 08:41:14.137107 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:14Z","lastTransitionTime":"2025-12-01T08:41:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:14 crc kubenswrapper[4813]: I1201 08:41:14.239814 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:14 crc kubenswrapper[4813]: I1201 08:41:14.239858 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:14 crc kubenswrapper[4813]: I1201 08:41:14.239869 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:14 crc kubenswrapper[4813]: I1201 08:41:14.239887 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:14 crc kubenswrapper[4813]: I1201 08:41:14.239898 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:14Z","lastTransitionTime":"2025-12-01T08:41:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:14 crc kubenswrapper[4813]: I1201 08:41:14.342110 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:14 crc kubenswrapper[4813]: I1201 08:41:14.342151 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:14 crc kubenswrapper[4813]: I1201 08:41:14.342163 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:14 crc kubenswrapper[4813]: I1201 08:41:14.342180 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:14 crc kubenswrapper[4813]: I1201 08:41:14.342201 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:14Z","lastTransitionTime":"2025-12-01T08:41:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:14 crc kubenswrapper[4813]: I1201 08:41:14.393168 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 08:41:14 crc kubenswrapper[4813]: I1201 08:41:14.393177 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 08:41:14 crc kubenswrapper[4813]: E1201 08:41:14.393349 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 08:41:14 crc kubenswrapper[4813]: E1201 08:41:14.393460 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 08:41:14 crc kubenswrapper[4813]: I1201 08:41:14.445884 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:14 crc kubenswrapper[4813]: I1201 08:41:14.445956 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:14 crc kubenswrapper[4813]: I1201 08:41:14.446001 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:14 crc kubenswrapper[4813]: I1201 08:41:14.446030 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:14 crc kubenswrapper[4813]: I1201 08:41:14.446048 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:14Z","lastTransitionTime":"2025-12-01T08:41:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:14 crc kubenswrapper[4813]: I1201 08:41:14.549144 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:14 crc kubenswrapper[4813]: I1201 08:41:14.549187 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:14 crc kubenswrapper[4813]: I1201 08:41:14.549197 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:14 crc kubenswrapper[4813]: I1201 08:41:14.549215 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:14 crc kubenswrapper[4813]: I1201 08:41:14.549225 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:14Z","lastTransitionTime":"2025-12-01T08:41:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:14 crc kubenswrapper[4813]: I1201 08:41:14.651743 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:14 crc kubenswrapper[4813]: I1201 08:41:14.651811 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:14 crc kubenswrapper[4813]: I1201 08:41:14.651838 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:14 crc kubenswrapper[4813]: I1201 08:41:14.651871 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:14 crc kubenswrapper[4813]: I1201 08:41:14.651894 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:14Z","lastTransitionTime":"2025-12-01T08:41:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:14 crc kubenswrapper[4813]: I1201 08:41:14.755051 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:14 crc kubenswrapper[4813]: I1201 08:41:14.755093 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:14 crc kubenswrapper[4813]: I1201 08:41:14.755107 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:14 crc kubenswrapper[4813]: I1201 08:41:14.755125 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:14 crc kubenswrapper[4813]: I1201 08:41:14.755137 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:14Z","lastTransitionTime":"2025-12-01T08:41:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:14 crc kubenswrapper[4813]: I1201 08:41:14.857370 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:14 crc kubenswrapper[4813]: I1201 08:41:14.857424 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:14 crc kubenswrapper[4813]: I1201 08:41:14.857432 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:14 crc kubenswrapper[4813]: I1201 08:41:14.857447 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:14 crc kubenswrapper[4813]: I1201 08:41:14.857474 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:14Z","lastTransitionTime":"2025-12-01T08:41:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:14 crc kubenswrapper[4813]: I1201 08:41:14.960420 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:14 crc kubenswrapper[4813]: I1201 08:41:14.961082 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:14 crc kubenswrapper[4813]: I1201 08:41:14.961116 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:14 crc kubenswrapper[4813]: I1201 08:41:14.961142 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:14 crc kubenswrapper[4813]: I1201 08:41:14.961156 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:14Z","lastTransitionTime":"2025-12-01T08:41:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:15 crc kubenswrapper[4813]: I1201 08:41:15.063918 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:15 crc kubenswrapper[4813]: I1201 08:41:15.063983 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:15 crc kubenswrapper[4813]: I1201 08:41:15.063996 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:15 crc kubenswrapper[4813]: I1201 08:41:15.064013 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:15 crc kubenswrapper[4813]: I1201 08:41:15.064026 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:15Z","lastTransitionTime":"2025-12-01T08:41:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:15 crc kubenswrapper[4813]: I1201 08:41:15.166466 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:15 crc kubenswrapper[4813]: I1201 08:41:15.166506 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:15 crc kubenswrapper[4813]: I1201 08:41:15.166519 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:15 crc kubenswrapper[4813]: I1201 08:41:15.166534 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:15 crc kubenswrapper[4813]: I1201 08:41:15.166545 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:15Z","lastTransitionTime":"2025-12-01T08:41:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:15 crc kubenswrapper[4813]: I1201 08:41:15.268469 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:15 crc kubenswrapper[4813]: I1201 08:41:15.268540 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:15 crc kubenswrapper[4813]: I1201 08:41:15.268564 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:15 crc kubenswrapper[4813]: I1201 08:41:15.268595 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:15 crc kubenswrapper[4813]: I1201 08:41:15.268618 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:15Z","lastTransitionTime":"2025-12-01T08:41:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:15 crc kubenswrapper[4813]: I1201 08:41:15.372107 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:15 crc kubenswrapper[4813]: I1201 08:41:15.372173 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:15 crc kubenswrapper[4813]: I1201 08:41:15.372191 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:15 crc kubenswrapper[4813]: I1201 08:41:15.372216 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:15 crc kubenswrapper[4813]: I1201 08:41:15.372234 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:15Z","lastTransitionTime":"2025-12-01T08:41:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:15 crc kubenswrapper[4813]: I1201 08:41:15.393533 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 08:41:15 crc kubenswrapper[4813]: I1201 08:41:15.393558 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-96mbr" Dec 01 08:41:15 crc kubenswrapper[4813]: E1201 08:41:15.393788 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 08:41:15 crc kubenswrapper[4813]: E1201 08:41:15.393835 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-96mbr" podUID="0789e3f7-5b08-4c04-8596-9ba29f50cdd0" Dec 01 08:41:15 crc kubenswrapper[4813]: I1201 08:41:15.475313 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:15 crc kubenswrapper[4813]: I1201 08:41:15.475394 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:15 crc kubenswrapper[4813]: I1201 08:41:15.475429 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:15 crc kubenswrapper[4813]: I1201 08:41:15.475462 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:15 crc kubenswrapper[4813]: I1201 08:41:15.475486 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:15Z","lastTransitionTime":"2025-12-01T08:41:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:15 crc kubenswrapper[4813]: I1201 08:41:15.578642 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:15 crc kubenswrapper[4813]: I1201 08:41:15.578718 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:15 crc kubenswrapper[4813]: I1201 08:41:15.578741 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:15 crc kubenswrapper[4813]: I1201 08:41:15.578773 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:15 crc kubenswrapper[4813]: I1201 08:41:15.578795 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:15Z","lastTransitionTime":"2025-12-01T08:41:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:15 crc kubenswrapper[4813]: I1201 08:41:15.681253 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:15 crc kubenswrapper[4813]: I1201 08:41:15.681327 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:15 crc kubenswrapper[4813]: I1201 08:41:15.681350 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:15 crc kubenswrapper[4813]: I1201 08:41:15.681381 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:15 crc kubenswrapper[4813]: I1201 08:41:15.681405 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:15Z","lastTransitionTime":"2025-12-01T08:41:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:15 crc kubenswrapper[4813]: I1201 08:41:15.785155 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:15 crc kubenswrapper[4813]: I1201 08:41:15.785235 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:15 crc kubenswrapper[4813]: I1201 08:41:15.785256 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:15 crc kubenswrapper[4813]: I1201 08:41:15.785286 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:15 crc kubenswrapper[4813]: I1201 08:41:15.785309 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:15Z","lastTransitionTime":"2025-12-01T08:41:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:15 crc kubenswrapper[4813]: I1201 08:41:15.888913 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:15 crc kubenswrapper[4813]: I1201 08:41:15.889035 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:15 crc kubenswrapper[4813]: I1201 08:41:15.889073 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:15 crc kubenswrapper[4813]: I1201 08:41:15.889113 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:15 crc kubenswrapper[4813]: I1201 08:41:15.889138 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:15Z","lastTransitionTime":"2025-12-01T08:41:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:15 crc kubenswrapper[4813]: I1201 08:41:15.992767 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:15 crc kubenswrapper[4813]: I1201 08:41:15.992847 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:15 crc kubenswrapper[4813]: I1201 08:41:15.992864 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:15 crc kubenswrapper[4813]: I1201 08:41:15.992889 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:15 crc kubenswrapper[4813]: I1201 08:41:15.992907 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:15Z","lastTransitionTime":"2025-12-01T08:41:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:16 crc kubenswrapper[4813]: I1201 08:41:16.095855 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:16 crc kubenswrapper[4813]: I1201 08:41:16.095923 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:16 crc kubenswrapper[4813]: I1201 08:41:16.095958 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:16 crc kubenswrapper[4813]: I1201 08:41:16.096047 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:16 crc kubenswrapper[4813]: I1201 08:41:16.096071 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:16Z","lastTransitionTime":"2025-12-01T08:41:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:16 crc kubenswrapper[4813]: I1201 08:41:16.199412 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:16 crc kubenswrapper[4813]: I1201 08:41:16.199464 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:16 crc kubenswrapper[4813]: I1201 08:41:16.199481 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:16 crc kubenswrapper[4813]: I1201 08:41:16.199501 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:16 crc kubenswrapper[4813]: I1201 08:41:16.199513 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:16Z","lastTransitionTime":"2025-12-01T08:41:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:16 crc kubenswrapper[4813]: I1201 08:41:16.302126 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:16 crc kubenswrapper[4813]: I1201 08:41:16.302177 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:16 crc kubenswrapper[4813]: I1201 08:41:16.302197 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:16 crc kubenswrapper[4813]: I1201 08:41:16.302212 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:16 crc kubenswrapper[4813]: I1201 08:41:16.302221 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:16Z","lastTransitionTime":"2025-12-01T08:41:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:16 crc kubenswrapper[4813]: I1201 08:41:16.392764 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 08:41:16 crc kubenswrapper[4813]: I1201 08:41:16.392867 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 08:41:16 crc kubenswrapper[4813]: E1201 08:41:16.393045 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 08:41:16 crc kubenswrapper[4813]: E1201 08:41:16.393446 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 08:41:16 crc kubenswrapper[4813]: I1201 08:41:16.409270 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:16 crc kubenswrapper[4813]: I1201 08:41:16.409529 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:16 crc kubenswrapper[4813]: I1201 08:41:16.409680 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:16 crc kubenswrapper[4813]: I1201 08:41:16.410004 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:16 crc kubenswrapper[4813]: I1201 08:41:16.410142 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:16Z","lastTransitionTime":"2025-12-01T08:41:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:16 crc kubenswrapper[4813]: I1201 08:41:16.414171 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-9wn4q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3d423402-e08f-42e9-b60c-6eb9bf2f6d48\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3049b842978c6580e1033a442a37766a2e76e278aface9a89fc0b2c57faf0df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rzxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c915733072b32a63237e2401cbf84b059aca396c2c0db884f0f439d599a35d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rzxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:41Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-9wn4q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:16Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:16 crc kubenswrapper[4813]: I1201 08:41:16.432860 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-8pg2p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dd53b38d-d9dd-4f98-9d06-dd95f1a4e3a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a214d71293e4e09573c665524d880fc319dfb1c0c6d95d11d78f253865660fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnjhq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ddf9d5989e4195b3074de08f99d53b48066f53fe93ce204145b3c714996fc57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnjhq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:53Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-8pg2p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:16Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:16 crc kubenswrapper[4813]: I1201 08:41:16.451051 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"490e526a-902b-4aff-b547-0cefea562764\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb48bd28fd3d29781695a5ae630f4913ef2745e97fbe7d463cca39a31e43ba0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74f8cf73f07b4cde260af429d26413e4e6226e7b177b493dd9c0e5c043ab6e1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3aaf696cefbe398bb6b26acfc5be43bda7e2be18218c8060840350c4106e66a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53de1f0682f36aa83ad8e5c226b95daf9f97df4ae154ebfc2fc287f9f238a4a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53de1f0682f36aa83ad8e5c226b95daf9f97df4ae154ebfc2fc287f9f238a4a8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:17Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:16Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:16Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:16 crc kubenswrapper[4813]: I1201 08:41:16.467340 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:16Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:16 crc kubenswrapper[4813]: I1201 08:41:16.483744 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-c576v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"17292cd0-1eac-40b5-a60e-67167c57409c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a42206c914d0493ca74788ea1a941de80920ab7b810fb95227cd9b8711e8f4b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7zwfx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:41Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-c576v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:16Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:16 crc kubenswrapper[4813]: I1201 08:41:16.509513 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-j4pcc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b7505e14-3560-4d0d-be2c-d6a7a5953865\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ccc30a63dc1c23ff9880fb04af1f31f0656b2696d359988e23deed23989f480a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://944977e9722025088720107530d0c91fc693593f380e4b49678f28a56da180cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc32d930ab23f30103d8fa42291909cbec89ebb75c5c53cd783cd27fe1a3f09e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f19c67b1c6c80c7f3223512102810d46ede1445464d21ae3bbea83424539b2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d984c90302e57f3e5a6f3ca3613b415db04ac74a5a228582f1b52ef1d59f93e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce8963bf1e9ee66926907aaf3a03d18977f4d0a1983d7121b62fbc8e02a37e4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8790f4730881303e7743498b66fc17b4b0df5b6620c8ca640ad7b9c90e6898d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8790f4730881303e7743498b66fc17b4b0df5b6620c8ca640ad7b9c90e6898d2\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T08:40:56Z\\\",\\\"message\\\":\\\"I1201 08:40:56.782175 6307 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1201 08:40:56.782219 6307 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1201 08:40:56.782251 6307 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1201 08:40:56.782302 6307 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1201 08:40:56.782342 6307 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1201 08:40:56.782379 6307 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1201 08:40:56.782414 6307 factory.go:656] Stopping watch factory\\\\nI1201 08:40:56.782444 6307 ovnkube.go:599] Stopped ovnkube\\\\nI1201 08:40:56.782487 6307 handler.go:208] Removed *v1.Node event handler 7\\\\nI1201 08:40:56.782514 6307 handler.go:208] Removed *v1.Node event handler 2\\\\nI1201 08:40:56.782539 6307 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1201 08:40:56.782564 6307 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1201 08:40:56.782591 6307 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1201 08:40:56.782650 6307 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1201 08:40:56.782670 6307 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1201 08:40:56.782663 6307 handler.go:208] Removed *v1.EgressFirewa\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:55Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-j4pcc_openshift-ovn-kubernetes(b7505e14-3560-4d0d-be2c-d6a7a5953865)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b45454b440c5108fa150a6b856afa1778ba6e9f89633433cb3117b68f00f72c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd67634e0c806323334495401a4e20b48fea4bc5013acd973d2abd3d70870f99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd67634e0c806323334495401a4e20b48fea4bc5013acd973d2abd3d70870f99\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:41Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-j4pcc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:16Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:16 crc kubenswrapper[4813]: I1201 08:41:16.513574 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:16 crc kubenswrapper[4813]: I1201 08:41:16.513644 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:16 crc kubenswrapper[4813]: I1201 08:41:16.513667 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:16 crc kubenswrapper[4813]: I1201 08:41:16.513696 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:16 crc kubenswrapper[4813]: I1201 08:41:16.513713 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:16Z","lastTransitionTime":"2025-12-01T08:41:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:16 crc kubenswrapper[4813]: I1201 08:41:16.529573 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-h7lvd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d431b613-a91b-4fff-88b9-447804766289\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a32c51c943fcf6ada7e3f614667eb023e29d94c11128d26c1fecca37b03963b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://495593f99e92a9c28100a95e27d06ff5847008c5d996366cb474bedd9b69d5ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://495593f99e92a9c28100a95e27d06ff5847008c5d996366cb474bedd9b69d5ef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://19358f83a6c52957f959fe935a6843fdeef59905cc395e6e52a1fb4a670ac193\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://19358f83a6c52957f959fe935a6843fdeef59905cc395e6e52a1fb4a670ac193\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f089c115d916e339994879d224e3db27642b99aaabf59bcafb2979549052945\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9f089c115d916e339994879d224e3db27642b99aaabf59bcafb2979549052945\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb66f224462305c2005b31ef7bb9ed01efde8101e18518dd42d1dfb0df765bb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb66f224462305c2005b31ef7bb9ed01efde8101e18518dd42d1dfb0df765bb9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c268236c9a155c9b1f47c48d30dbd23948ceb2545592b40f356c60ea1d12e8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c268236c9a155c9b1f47c48d30dbd23948ceb2545592b40f356c60ea1d12e8a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://458aedca5f186e9aaed7e4ee29fd5247221491888b130f55c1fe48d88d565f8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://458aedca5f186e9aaed7e4ee29fd5247221491888b130f55c1fe48d88d565f8f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:41Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-h7lvd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:16Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:16 crc kubenswrapper[4813]: I1201 08:41:16.543939 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6fb5a648-ae4f-4e46-a530-007e9bf84abb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e24ebfc52dcb28368a2d436e7d9ab78f07f2aec2a4c71405a6c7d3301eee96a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33898118bbf639c8e6eae5aafa10386b95795512444ed5f535db3c663f169df2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9902c00bf7738686c93f62e1772a484d05715d74b75ee98c84c8f23ed383824b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3657b47ec9f1563c83bfbe3d728c8fd123f7078e219d251b1874269e42c31eea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:16Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:16Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:16 crc kubenswrapper[4813]: I1201 08:41:16.559695 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0023b0e358d75aacc889ce364846dda89ea513832c87fa4b066aea7dac57e8a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:16Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:16 crc kubenswrapper[4813]: I1201 08:41:16.572271 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:16Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:16 crc kubenswrapper[4813]: I1201 08:41:16.588473 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c7574f892d5e9142296b26a0b484eb6759ab53f571139fb6e23b93dc3da05780\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:16Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:16 crc kubenswrapper[4813]: I1201 08:41:16.601123 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:16Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:16 crc kubenswrapper[4813]: I1201 08:41:16.612854 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-xpnlc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"22da1fcf-0c57-4708-90d8-77692af17344\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4a36d213bdad2da91d0d33f400d5a8248befb68d5b94d19347873a93c6712a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xc44d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-xpnlc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:16Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:16 crc kubenswrapper[4813]: I1201 08:41:16.616516 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:16 crc kubenswrapper[4813]: I1201 08:41:16.616558 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:16 crc kubenswrapper[4813]: I1201 08:41:16.616572 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:16 crc kubenswrapper[4813]: I1201 08:41:16.616590 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:16 crc kubenswrapper[4813]: I1201 08:41:16.616605 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:16Z","lastTransitionTime":"2025-12-01T08:41:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:16 crc kubenswrapper[4813]: I1201 08:41:16.628996 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"349332bd-6b4d-4306-84ae-07019ca3860f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d441a348087ae3cadda1ad8a7d1c4dff02cf3f6e1d346c294b278fdd9131edcc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://da67635daf6d9e32937f4d5d4ceeee3f8f09ae4fd2dd26c3e85be1286c5949de\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e97356ee274ea78d9f237ffbc756abb6ac8df0dbac244bf0a86ed215b8cbf43\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8dc6feb9817f4f356b8e36af43ec3070774f312fbbb4909fa05b8b3e8dd53582\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a0404f5a5ed5a9750d4c479b608e68e8960b448f455d5af9d219aa5edaeb789b\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"message\\\":\\\"\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764578421\\\\\\\\\\\\\\\" (2025-12-01 08:40:21 +0000 UTC to 2025-12-31 08:40:22 +0000 UTC (now=2025-12-01 08:40:38.066068451 +0000 UTC))\\\\\\\"\\\\nI1201 08:40:38.066144 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1201 08:40:38.066166 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1201 08:40:38.066220 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1201 08:40:38.066232 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1201 08:40:38.066338 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764578432\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764578432\\\\\\\\\\\\\\\" (2025-12-01 07:40:32 +0000 UTC to 2026-12-01 07:40:32 +0000 UTC (now=2025-12-01 08:40:38.066301217 +0000 UTC))\\\\\\\"\\\\nI1201 08:40:38.066357 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1201 08:40:38.066385 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1201 08:40:38.066402 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1201 08:40:38.066414 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-63890811/tls.crt::/tmp/serving-cert-63890811/tls.key\\\\\\\"\\\\nI1201 08:40:38.066419 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1201 08:40:38.066572 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1201 08:40:38.067159 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7354f9337d42115bb1d974bf3772f2d03365acba57f2155135c7a0bbab39879e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:19Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05492df5703392cdf0932d6bf66025b18e53bd05fa0857a7cb2ad70979cbdc3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05492df5703392cdf0932d6bf66025b18e53bd05fa0857a7cb2ad70979cbdc3e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:16Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:16Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:16 crc kubenswrapper[4813]: I1201 08:41:16.644484 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3264b3af996af452df2953c59bea6deeb5ea6af1b119ce89f7334595c27e70da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9e879b1f563bd66383d58fd7b435a8a81bd50cb387dc1fabb6fc47d5a3fa468\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:16Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:16 crc kubenswrapper[4813]: I1201 08:41:16.657931 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-b6qqv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9656a03-0096-4c89-b0da-e37e4103611b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4195aebc1b2e388d480920a7f72257f6660afc99ad8c1a09478ef445739b272\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cjvnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:41Z\\\"}}\" for pod \"openshift-multus\"/\"multus-b6qqv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:16Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:16 crc kubenswrapper[4813]: I1201 08:41:16.681624 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-96mbr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0789e3f7-5b08-4c04-8596-9ba29f50cdd0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xt4fj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xt4fj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:55Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-96mbr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:16Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:16 crc kubenswrapper[4813]: I1201 08:41:16.719231 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:16 crc kubenswrapper[4813]: I1201 08:41:16.719285 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:16 crc kubenswrapper[4813]: I1201 08:41:16.719295 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:16 crc kubenswrapper[4813]: I1201 08:41:16.719316 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:16 crc kubenswrapper[4813]: I1201 08:41:16.719327 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:16Z","lastTransitionTime":"2025-12-01T08:41:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:16 crc kubenswrapper[4813]: I1201 08:41:16.822783 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:16 crc kubenswrapper[4813]: I1201 08:41:16.822842 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:16 crc kubenswrapper[4813]: I1201 08:41:16.822866 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:16 crc kubenswrapper[4813]: I1201 08:41:16.823302 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:16 crc kubenswrapper[4813]: I1201 08:41:16.823332 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:16Z","lastTransitionTime":"2025-12-01T08:41:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:16 crc kubenswrapper[4813]: I1201 08:41:16.927148 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:16 crc kubenswrapper[4813]: I1201 08:41:16.927228 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:16 crc kubenswrapper[4813]: I1201 08:41:16.927250 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:16 crc kubenswrapper[4813]: I1201 08:41:16.927284 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:16 crc kubenswrapper[4813]: I1201 08:41:16.927309 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:16Z","lastTransitionTime":"2025-12-01T08:41:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:17 crc kubenswrapper[4813]: I1201 08:41:17.031038 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:17 crc kubenswrapper[4813]: I1201 08:41:17.031120 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:17 crc kubenswrapper[4813]: I1201 08:41:17.031143 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:17 crc kubenswrapper[4813]: I1201 08:41:17.031166 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:17 crc kubenswrapper[4813]: I1201 08:41:17.031185 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:17Z","lastTransitionTime":"2025-12-01T08:41:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:17 crc kubenswrapper[4813]: I1201 08:41:17.134695 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:17 crc kubenswrapper[4813]: I1201 08:41:17.134841 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:17 crc kubenswrapper[4813]: I1201 08:41:17.134865 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:17 crc kubenswrapper[4813]: I1201 08:41:17.134894 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:17 crc kubenswrapper[4813]: I1201 08:41:17.134914 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:17Z","lastTransitionTime":"2025-12-01T08:41:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:17 crc kubenswrapper[4813]: I1201 08:41:17.238811 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:17 crc kubenswrapper[4813]: I1201 08:41:17.238901 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:17 crc kubenswrapper[4813]: I1201 08:41:17.238921 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:17 crc kubenswrapper[4813]: I1201 08:41:17.238948 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:17 crc kubenswrapper[4813]: I1201 08:41:17.239001 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:17Z","lastTransitionTime":"2025-12-01T08:41:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:17 crc kubenswrapper[4813]: I1201 08:41:17.341476 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:17 crc kubenswrapper[4813]: I1201 08:41:17.341514 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:17 crc kubenswrapper[4813]: I1201 08:41:17.341526 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:17 crc kubenswrapper[4813]: I1201 08:41:17.341542 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:17 crc kubenswrapper[4813]: I1201 08:41:17.341554 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:17Z","lastTransitionTime":"2025-12-01T08:41:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:17 crc kubenswrapper[4813]: I1201 08:41:17.392887 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-96mbr" Dec 01 08:41:17 crc kubenswrapper[4813]: I1201 08:41:17.392887 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 08:41:17 crc kubenswrapper[4813]: E1201 08:41:17.393210 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-96mbr" podUID="0789e3f7-5b08-4c04-8596-9ba29f50cdd0" Dec 01 08:41:17 crc kubenswrapper[4813]: E1201 08:41:17.393348 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 08:41:17 crc kubenswrapper[4813]: I1201 08:41:17.445032 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:17 crc kubenswrapper[4813]: I1201 08:41:17.445094 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:17 crc kubenswrapper[4813]: I1201 08:41:17.445113 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:17 crc kubenswrapper[4813]: I1201 08:41:17.445141 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:17 crc kubenswrapper[4813]: I1201 08:41:17.445165 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:17Z","lastTransitionTime":"2025-12-01T08:41:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:17 crc kubenswrapper[4813]: I1201 08:41:17.548755 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:17 crc kubenswrapper[4813]: I1201 08:41:17.548809 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:17 crc kubenswrapper[4813]: I1201 08:41:17.548863 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:17 crc kubenswrapper[4813]: I1201 08:41:17.548886 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:17 crc kubenswrapper[4813]: I1201 08:41:17.548900 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:17Z","lastTransitionTime":"2025-12-01T08:41:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:17 crc kubenswrapper[4813]: I1201 08:41:17.652729 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:17 crc kubenswrapper[4813]: I1201 08:41:17.652789 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:17 crc kubenswrapper[4813]: I1201 08:41:17.652807 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:17 crc kubenswrapper[4813]: I1201 08:41:17.652831 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:17 crc kubenswrapper[4813]: I1201 08:41:17.652849 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:17Z","lastTransitionTime":"2025-12-01T08:41:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:17 crc kubenswrapper[4813]: I1201 08:41:17.756455 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:17 crc kubenswrapper[4813]: I1201 08:41:17.756545 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:17 crc kubenswrapper[4813]: I1201 08:41:17.756572 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:17 crc kubenswrapper[4813]: I1201 08:41:17.757136 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:17 crc kubenswrapper[4813]: I1201 08:41:17.757401 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:17Z","lastTransitionTime":"2025-12-01T08:41:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:17 crc kubenswrapper[4813]: I1201 08:41:17.861600 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:17 crc kubenswrapper[4813]: I1201 08:41:17.861664 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:17 crc kubenswrapper[4813]: I1201 08:41:17.861681 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:17 crc kubenswrapper[4813]: I1201 08:41:17.861705 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:17 crc kubenswrapper[4813]: I1201 08:41:17.861726 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:17Z","lastTransitionTime":"2025-12-01T08:41:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:17 crc kubenswrapper[4813]: I1201 08:41:17.965146 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:17 crc kubenswrapper[4813]: I1201 08:41:17.965203 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:17 crc kubenswrapper[4813]: I1201 08:41:17.965218 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:17 crc kubenswrapper[4813]: I1201 08:41:17.965248 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:17 crc kubenswrapper[4813]: I1201 08:41:17.965270 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:17Z","lastTransitionTime":"2025-12-01T08:41:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:18 crc kubenswrapper[4813]: I1201 08:41:18.067441 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:18 crc kubenswrapper[4813]: I1201 08:41:18.067515 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:18 crc kubenswrapper[4813]: I1201 08:41:18.067536 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:18 crc kubenswrapper[4813]: I1201 08:41:18.067560 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:18 crc kubenswrapper[4813]: I1201 08:41:18.067577 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:18Z","lastTransitionTime":"2025-12-01T08:41:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:18 crc kubenswrapper[4813]: I1201 08:41:18.170157 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:18 crc kubenswrapper[4813]: I1201 08:41:18.170204 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:18 crc kubenswrapper[4813]: I1201 08:41:18.170213 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:18 crc kubenswrapper[4813]: I1201 08:41:18.170228 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:18 crc kubenswrapper[4813]: I1201 08:41:18.170236 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:18Z","lastTransitionTime":"2025-12-01T08:41:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:18 crc kubenswrapper[4813]: I1201 08:41:18.272730 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:18 crc kubenswrapper[4813]: I1201 08:41:18.272773 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:18 crc kubenswrapper[4813]: I1201 08:41:18.272787 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:18 crc kubenswrapper[4813]: I1201 08:41:18.272803 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:18 crc kubenswrapper[4813]: I1201 08:41:18.272812 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:18Z","lastTransitionTime":"2025-12-01T08:41:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:18 crc kubenswrapper[4813]: I1201 08:41:18.374992 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:18 crc kubenswrapper[4813]: I1201 08:41:18.375037 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:18 crc kubenswrapper[4813]: I1201 08:41:18.375071 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:18 crc kubenswrapper[4813]: I1201 08:41:18.375090 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:18 crc kubenswrapper[4813]: I1201 08:41:18.375099 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:18Z","lastTransitionTime":"2025-12-01T08:41:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:18 crc kubenswrapper[4813]: I1201 08:41:18.393719 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 08:41:18 crc kubenswrapper[4813]: I1201 08:41:18.393780 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 08:41:18 crc kubenswrapper[4813]: E1201 08:41:18.393877 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 08:41:18 crc kubenswrapper[4813]: E1201 08:41:18.394062 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 08:41:18 crc kubenswrapper[4813]: I1201 08:41:18.477463 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:18 crc kubenswrapper[4813]: I1201 08:41:18.477512 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:18 crc kubenswrapper[4813]: I1201 08:41:18.477520 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:18 crc kubenswrapper[4813]: I1201 08:41:18.477534 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:18 crc kubenswrapper[4813]: I1201 08:41:18.477542 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:18Z","lastTransitionTime":"2025-12-01T08:41:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:18 crc kubenswrapper[4813]: I1201 08:41:18.581428 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:18 crc kubenswrapper[4813]: I1201 08:41:18.581529 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:18 crc kubenswrapper[4813]: I1201 08:41:18.581557 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:18 crc kubenswrapper[4813]: I1201 08:41:18.581589 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:18 crc kubenswrapper[4813]: I1201 08:41:18.581613 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:18Z","lastTransitionTime":"2025-12-01T08:41:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:18 crc kubenswrapper[4813]: I1201 08:41:18.685618 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:18 crc kubenswrapper[4813]: I1201 08:41:18.686138 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:18 crc kubenswrapper[4813]: I1201 08:41:18.686164 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:18 crc kubenswrapper[4813]: I1201 08:41:18.686189 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:18 crc kubenswrapper[4813]: I1201 08:41:18.686209 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:18Z","lastTransitionTime":"2025-12-01T08:41:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:18 crc kubenswrapper[4813]: I1201 08:41:18.789348 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:18 crc kubenswrapper[4813]: I1201 08:41:18.789420 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:18 crc kubenswrapper[4813]: I1201 08:41:18.789443 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:18 crc kubenswrapper[4813]: I1201 08:41:18.789472 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:18 crc kubenswrapper[4813]: I1201 08:41:18.789499 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:18Z","lastTransitionTime":"2025-12-01T08:41:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:18 crc kubenswrapper[4813]: I1201 08:41:18.893604 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:18 crc kubenswrapper[4813]: I1201 08:41:18.893668 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:18 crc kubenswrapper[4813]: I1201 08:41:18.893694 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:18 crc kubenswrapper[4813]: I1201 08:41:18.893724 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:18 crc kubenswrapper[4813]: I1201 08:41:18.893745 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:18Z","lastTransitionTime":"2025-12-01T08:41:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:18 crc kubenswrapper[4813]: I1201 08:41:18.996618 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:18 crc kubenswrapper[4813]: I1201 08:41:18.996667 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:18 crc kubenswrapper[4813]: I1201 08:41:18.996677 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:18 crc kubenswrapper[4813]: I1201 08:41:18.996693 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:18 crc kubenswrapper[4813]: I1201 08:41:18.996704 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:18Z","lastTransitionTime":"2025-12-01T08:41:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:19 crc kubenswrapper[4813]: I1201 08:41:19.099270 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:19 crc kubenswrapper[4813]: I1201 08:41:19.099331 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:19 crc kubenswrapper[4813]: I1201 08:41:19.099343 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:19 crc kubenswrapper[4813]: I1201 08:41:19.099362 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:19 crc kubenswrapper[4813]: I1201 08:41:19.099375 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:19Z","lastTransitionTime":"2025-12-01T08:41:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:19 crc kubenswrapper[4813]: I1201 08:41:19.203592 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:19 crc kubenswrapper[4813]: I1201 08:41:19.203633 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:19 crc kubenswrapper[4813]: I1201 08:41:19.203642 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:19 crc kubenswrapper[4813]: I1201 08:41:19.203659 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:19 crc kubenswrapper[4813]: I1201 08:41:19.203670 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:19Z","lastTransitionTime":"2025-12-01T08:41:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:19 crc kubenswrapper[4813]: I1201 08:41:19.306137 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:19 crc kubenswrapper[4813]: I1201 08:41:19.306204 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:19 crc kubenswrapper[4813]: I1201 08:41:19.306219 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:19 crc kubenswrapper[4813]: I1201 08:41:19.306251 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:19 crc kubenswrapper[4813]: I1201 08:41:19.306269 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:19Z","lastTransitionTime":"2025-12-01T08:41:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:19 crc kubenswrapper[4813]: I1201 08:41:19.393369 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-96mbr" Dec 01 08:41:19 crc kubenswrapper[4813]: I1201 08:41:19.393384 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 08:41:19 crc kubenswrapper[4813]: E1201 08:41:19.393625 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-96mbr" podUID="0789e3f7-5b08-4c04-8596-9ba29f50cdd0" Dec 01 08:41:19 crc kubenswrapper[4813]: E1201 08:41:19.393714 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 08:41:19 crc kubenswrapper[4813]: I1201 08:41:19.408555 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:19 crc kubenswrapper[4813]: I1201 08:41:19.408605 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:19 crc kubenswrapper[4813]: I1201 08:41:19.408613 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:19 crc kubenswrapper[4813]: I1201 08:41:19.408630 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:19 crc kubenswrapper[4813]: I1201 08:41:19.408640 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:19Z","lastTransitionTime":"2025-12-01T08:41:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:19 crc kubenswrapper[4813]: I1201 08:41:19.511899 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:19 crc kubenswrapper[4813]: I1201 08:41:19.511950 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:19 crc kubenswrapper[4813]: I1201 08:41:19.512025 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:19 crc kubenswrapper[4813]: I1201 08:41:19.512052 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:19 crc kubenswrapper[4813]: I1201 08:41:19.512063 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:19Z","lastTransitionTime":"2025-12-01T08:41:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:19 crc kubenswrapper[4813]: I1201 08:41:19.614063 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:19 crc kubenswrapper[4813]: I1201 08:41:19.614100 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:19 crc kubenswrapper[4813]: I1201 08:41:19.614109 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:19 crc kubenswrapper[4813]: I1201 08:41:19.614123 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:19 crc kubenswrapper[4813]: I1201 08:41:19.614132 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:19Z","lastTransitionTime":"2025-12-01T08:41:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:19 crc kubenswrapper[4813]: I1201 08:41:19.716615 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:19 crc kubenswrapper[4813]: I1201 08:41:19.716673 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:19 crc kubenswrapper[4813]: I1201 08:41:19.716687 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:19 crc kubenswrapper[4813]: I1201 08:41:19.716706 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:19 crc kubenswrapper[4813]: I1201 08:41:19.716721 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:19Z","lastTransitionTime":"2025-12-01T08:41:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:19 crc kubenswrapper[4813]: I1201 08:41:19.818735 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:19 crc kubenswrapper[4813]: I1201 08:41:19.818772 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:19 crc kubenswrapper[4813]: I1201 08:41:19.818782 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:19 crc kubenswrapper[4813]: I1201 08:41:19.818796 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:19 crc kubenswrapper[4813]: I1201 08:41:19.818806 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:19Z","lastTransitionTime":"2025-12-01T08:41:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:19 crc kubenswrapper[4813]: I1201 08:41:19.921918 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:19 crc kubenswrapper[4813]: I1201 08:41:19.922034 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:19 crc kubenswrapper[4813]: I1201 08:41:19.922059 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:19 crc kubenswrapper[4813]: I1201 08:41:19.922092 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:19 crc kubenswrapper[4813]: I1201 08:41:19.922114 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:19Z","lastTransitionTime":"2025-12-01T08:41:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:20 crc kubenswrapper[4813]: I1201 08:41:20.024572 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:20 crc kubenswrapper[4813]: I1201 08:41:20.024613 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:20 crc kubenswrapper[4813]: I1201 08:41:20.024624 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:20 crc kubenswrapper[4813]: I1201 08:41:20.024653 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:20 crc kubenswrapper[4813]: I1201 08:41:20.024664 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:20Z","lastTransitionTime":"2025-12-01T08:41:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:20 crc kubenswrapper[4813]: I1201 08:41:20.128722 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:20 crc kubenswrapper[4813]: I1201 08:41:20.128805 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:20 crc kubenswrapper[4813]: I1201 08:41:20.128828 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:20 crc kubenswrapper[4813]: I1201 08:41:20.128854 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:20 crc kubenswrapper[4813]: I1201 08:41:20.128871 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:20Z","lastTransitionTime":"2025-12-01T08:41:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:20 crc kubenswrapper[4813]: I1201 08:41:20.232091 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:20 crc kubenswrapper[4813]: I1201 08:41:20.232133 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:20 crc kubenswrapper[4813]: I1201 08:41:20.232144 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:20 crc kubenswrapper[4813]: I1201 08:41:20.232159 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:20 crc kubenswrapper[4813]: I1201 08:41:20.232170 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:20Z","lastTransitionTime":"2025-12-01T08:41:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:20 crc kubenswrapper[4813]: I1201 08:41:20.334516 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:20 crc kubenswrapper[4813]: I1201 08:41:20.334559 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:20 crc kubenswrapper[4813]: I1201 08:41:20.334572 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:20 crc kubenswrapper[4813]: I1201 08:41:20.334587 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:20 crc kubenswrapper[4813]: I1201 08:41:20.334601 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:20Z","lastTransitionTime":"2025-12-01T08:41:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:20 crc kubenswrapper[4813]: I1201 08:41:20.393180 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 08:41:20 crc kubenswrapper[4813]: I1201 08:41:20.393180 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 08:41:20 crc kubenswrapper[4813]: E1201 08:41:20.393340 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 08:41:20 crc kubenswrapper[4813]: E1201 08:41:20.393493 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 08:41:20 crc kubenswrapper[4813]: I1201 08:41:20.394243 4813 scope.go:117] "RemoveContainer" containerID="8790f4730881303e7743498b66fc17b4b0df5b6620c8ca640ad7b9c90e6898d2" Dec 01 08:41:20 crc kubenswrapper[4813]: I1201 08:41:20.438249 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:20 crc kubenswrapper[4813]: I1201 08:41:20.438823 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:20 crc kubenswrapper[4813]: I1201 08:41:20.438843 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:20 crc kubenswrapper[4813]: I1201 08:41:20.438869 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:20 crc kubenswrapper[4813]: I1201 08:41:20.438885 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:20Z","lastTransitionTime":"2025-12-01T08:41:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:20 crc kubenswrapper[4813]: I1201 08:41:20.537686 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-j4pcc_b7505e14-3560-4d0d-be2c-d6a7a5953865/ovnkube-controller/1.log" Dec 01 08:41:20 crc kubenswrapper[4813]: I1201 08:41:20.540361 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:20 crc kubenswrapper[4813]: I1201 08:41:20.540399 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:20 crc kubenswrapper[4813]: I1201 08:41:20.540410 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:20 crc kubenswrapper[4813]: I1201 08:41:20.540427 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:20 crc kubenswrapper[4813]: I1201 08:41:20.540438 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:20Z","lastTransitionTime":"2025-12-01T08:41:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:20 crc kubenswrapper[4813]: I1201 08:41:20.541336 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-j4pcc" event={"ID":"b7505e14-3560-4d0d-be2c-d6a7a5953865","Type":"ContainerStarted","Data":"fb17fb9ae11fa62235328567418fb259f4f8540ebac635842d3afe751dcbde8c"} Dec 01 08:41:20 crc kubenswrapper[4813]: I1201 08:41:20.542476 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-j4pcc" Dec 01 08:41:20 crc kubenswrapper[4813]: I1201 08:41:20.556917 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c7574f892d5e9142296b26a0b484eb6759ab53f571139fb6e23b93dc3da05780\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:20Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:20 crc kubenswrapper[4813]: I1201 08:41:20.570530 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:20Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:20 crc kubenswrapper[4813]: I1201 08:41:20.581242 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-xpnlc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"22da1fcf-0c57-4708-90d8-77692af17344\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4a36d213bdad2da91d0d33f400d5a8248befb68d5b94d19347873a93c6712a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xc44d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-xpnlc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:20Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:20 crc kubenswrapper[4813]: I1201 08:41:20.595714 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6fb5a648-ae4f-4e46-a530-007e9bf84abb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e24ebfc52dcb28368a2d436e7d9ab78f07f2aec2a4c71405a6c7d3301eee96a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33898118bbf639c8e6eae5aafa10386b95795512444ed5f535db3c663f169df2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9902c00bf7738686c93f62e1772a484d05715d74b75ee98c84c8f23ed383824b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3657b47ec9f1563c83bfbe3d728c8fd123f7078e219d251b1874269e42c31eea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:16Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:20Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:20 crc kubenswrapper[4813]: I1201 08:41:20.610020 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0023b0e358d75aacc889ce364846dda89ea513832c87fa4b066aea7dac57e8a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:20Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:20 crc kubenswrapper[4813]: I1201 08:41:20.624146 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:20Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:20 crc kubenswrapper[4813]: I1201 08:41:20.638195 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-b6qqv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9656a03-0096-4c89-b0da-e37e4103611b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4195aebc1b2e388d480920a7f72257f6660afc99ad8c1a09478ef445739b272\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cjvnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:41Z\\\"}}\" for pod \"openshift-multus\"/\"multus-b6qqv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:20Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:20 crc kubenswrapper[4813]: I1201 08:41:20.642525 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:20 crc kubenswrapper[4813]: I1201 08:41:20.642566 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:20 crc kubenswrapper[4813]: I1201 08:41:20.642575 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:20 crc kubenswrapper[4813]: I1201 08:41:20.642590 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:20 crc kubenswrapper[4813]: I1201 08:41:20.642600 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:20Z","lastTransitionTime":"2025-12-01T08:41:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:20 crc kubenswrapper[4813]: I1201 08:41:20.649514 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-96mbr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0789e3f7-5b08-4c04-8596-9ba29f50cdd0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xt4fj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xt4fj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:55Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-96mbr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:20Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:20 crc kubenswrapper[4813]: I1201 08:41:20.664364 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"349332bd-6b4d-4306-84ae-07019ca3860f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d441a348087ae3cadda1ad8a7d1c4dff02cf3f6e1d346c294b278fdd9131edcc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://da67635daf6d9e32937f4d5d4ceeee3f8f09ae4fd2dd26c3e85be1286c5949de\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e97356ee274ea78d9f237ffbc756abb6ac8df0dbac244bf0a86ed215b8cbf43\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8dc6feb9817f4f356b8e36af43ec3070774f312fbbb4909fa05b8b3e8dd53582\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a0404f5a5ed5a9750d4c479b608e68e8960b448f455d5af9d219aa5edaeb789b\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"message\\\":\\\"\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764578421\\\\\\\\\\\\\\\" (2025-12-01 08:40:21 +0000 UTC to 2025-12-31 08:40:22 +0000 UTC (now=2025-12-01 08:40:38.066068451 +0000 UTC))\\\\\\\"\\\\nI1201 08:40:38.066144 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1201 08:40:38.066166 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1201 08:40:38.066220 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1201 08:40:38.066232 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1201 08:40:38.066338 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764578432\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764578432\\\\\\\\\\\\\\\" (2025-12-01 07:40:32 +0000 UTC to 2026-12-01 07:40:32 +0000 UTC (now=2025-12-01 08:40:38.066301217 +0000 UTC))\\\\\\\"\\\\nI1201 08:40:38.066357 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1201 08:40:38.066385 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1201 08:40:38.066402 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1201 08:40:38.066414 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-63890811/tls.crt::/tmp/serving-cert-63890811/tls.key\\\\\\\"\\\\nI1201 08:40:38.066419 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1201 08:40:38.066572 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1201 08:40:38.067159 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7354f9337d42115bb1d974bf3772f2d03365acba57f2155135c7a0bbab39879e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:19Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05492df5703392cdf0932d6bf66025b18e53bd05fa0857a7cb2ad70979cbdc3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05492df5703392cdf0932d6bf66025b18e53bd05fa0857a7cb2ad70979cbdc3e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:16Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:20Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:20 crc kubenswrapper[4813]: I1201 08:41:20.678463 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3264b3af996af452df2953c59bea6deeb5ea6af1b119ce89f7334595c27e70da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9e879b1f563bd66383d58fd7b435a8a81bd50cb387dc1fabb6fc47d5a3fa468\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:20Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:20 crc kubenswrapper[4813]: I1201 08:41:20.689663 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-9wn4q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3d423402-e08f-42e9-b60c-6eb9bf2f6d48\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3049b842978c6580e1033a442a37766a2e76e278aface9a89fc0b2c57faf0df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rzxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c915733072b32a63237e2401cbf84b059aca396c2c0db884f0f439d599a35d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rzxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:41Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-9wn4q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:20Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:20 crc kubenswrapper[4813]: I1201 08:41:20.698986 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-8pg2p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dd53b38d-d9dd-4f98-9d06-dd95f1a4e3a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a214d71293e4e09573c665524d880fc319dfb1c0c6d95d11d78f253865660fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnjhq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ddf9d5989e4195b3074de08f99d53b48066f53fe93ce204145b3c714996fc57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnjhq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:53Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-8pg2p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:20Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:20 crc kubenswrapper[4813]: I1201 08:41:20.720077 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-j4pcc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b7505e14-3560-4d0d-be2c-d6a7a5953865\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ccc30a63dc1c23ff9880fb04af1f31f0656b2696d359988e23deed23989f480a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://944977e9722025088720107530d0c91fc693593f380e4b49678f28a56da180cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc32d930ab23f30103d8fa42291909cbec89ebb75c5c53cd783cd27fe1a3f09e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f19c67b1c6c80c7f3223512102810d46ede1445464d21ae3bbea83424539b2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d984c90302e57f3e5a6f3ca3613b415db04ac74a5a228582f1b52ef1d59f93e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce8963bf1e9ee66926907aaf3a03d18977f4d0a1983d7121b62fbc8e02a37e4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb17fb9ae11fa62235328567418fb259f4f8540ebac635842d3afe751dcbde8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8790f4730881303e7743498b66fc17b4b0df5b6620c8ca640ad7b9c90e6898d2\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T08:40:56Z\\\",\\\"message\\\":\\\"I1201 08:40:56.782175 6307 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1201 08:40:56.782219 6307 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1201 08:40:56.782251 6307 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1201 08:40:56.782302 6307 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1201 08:40:56.782342 6307 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1201 08:40:56.782379 6307 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1201 08:40:56.782414 6307 factory.go:656] Stopping watch factory\\\\nI1201 08:40:56.782444 6307 ovnkube.go:599] Stopped ovnkube\\\\nI1201 08:40:56.782487 6307 handler.go:208] Removed *v1.Node event handler 7\\\\nI1201 08:40:56.782514 6307 handler.go:208] Removed *v1.Node event handler 2\\\\nI1201 08:40:56.782539 6307 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1201 08:40:56.782564 6307 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1201 08:40:56.782591 6307 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1201 08:40:56.782650 6307 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1201 08:40:56.782670 6307 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1201 08:40:56.782663 6307 handler.go:208] Removed *v1.EgressFirewa\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:55Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:41:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b45454b440c5108fa150a6b856afa1778ba6e9f89633433cb3117b68f00f72c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd67634e0c806323334495401a4e20b48fea4bc5013acd973d2abd3d70870f99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd67634e0c806323334495401a4e20b48fea4bc5013acd973d2abd3d70870f99\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:41Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-j4pcc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:20Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:20 crc kubenswrapper[4813]: I1201 08:41:20.733203 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-h7lvd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d431b613-a91b-4fff-88b9-447804766289\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a32c51c943fcf6ada7e3f614667eb023e29d94c11128d26c1fecca37b03963b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://495593f99e92a9c28100a95e27d06ff5847008c5d996366cb474bedd9b69d5ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://495593f99e92a9c28100a95e27d06ff5847008c5d996366cb474bedd9b69d5ef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://19358f83a6c52957f959fe935a6843fdeef59905cc395e6e52a1fb4a670ac193\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://19358f83a6c52957f959fe935a6843fdeef59905cc395e6e52a1fb4a670ac193\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f089c115d916e339994879d224e3db27642b99aaabf59bcafb2979549052945\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9f089c115d916e339994879d224e3db27642b99aaabf59bcafb2979549052945\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb66f224462305c2005b31ef7bb9ed01efde8101e18518dd42d1dfb0df765bb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb66f224462305c2005b31ef7bb9ed01efde8101e18518dd42d1dfb0df765bb9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c268236c9a155c9b1f47c48d30dbd23948ceb2545592b40f356c60ea1d12e8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c268236c9a155c9b1f47c48d30dbd23948ceb2545592b40f356c60ea1d12e8a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://458aedca5f186e9aaed7e4ee29fd5247221491888b130f55c1fe48d88d565f8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://458aedca5f186e9aaed7e4ee29fd5247221491888b130f55c1fe48d88d565f8f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:41Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-h7lvd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:20Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:20 crc kubenswrapper[4813]: I1201 08:41:20.742690 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"490e526a-902b-4aff-b547-0cefea562764\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb48bd28fd3d29781695a5ae630f4913ef2745e97fbe7d463cca39a31e43ba0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74f8cf73f07b4cde260af429d26413e4e6226e7b177b493dd9c0e5c043ab6e1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3aaf696cefbe398bb6b26acfc5be43bda7e2be18218c8060840350c4106e66a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53de1f0682f36aa83ad8e5c226b95daf9f97df4ae154ebfc2fc287f9f238a4a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53de1f0682f36aa83ad8e5c226b95daf9f97df4ae154ebfc2fc287f9f238a4a8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:17Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:16Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:20Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:20 crc kubenswrapper[4813]: I1201 08:41:20.744539 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:20 crc kubenswrapper[4813]: I1201 08:41:20.744567 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:20 crc kubenswrapper[4813]: I1201 08:41:20.744577 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:20 crc kubenswrapper[4813]: I1201 08:41:20.744591 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:20 crc kubenswrapper[4813]: I1201 08:41:20.744599 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:20Z","lastTransitionTime":"2025-12-01T08:41:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:20 crc kubenswrapper[4813]: I1201 08:41:20.755392 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:20Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:20 crc kubenswrapper[4813]: I1201 08:41:20.765484 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-c576v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"17292cd0-1eac-40b5-a60e-67167c57409c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a42206c914d0493ca74788ea1a941de80920ab7b810fb95227cd9b8711e8f4b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7zwfx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:41Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-c576v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:20Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:20 crc kubenswrapper[4813]: I1201 08:41:20.782914 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:20 crc kubenswrapper[4813]: I1201 08:41:20.782985 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:20 crc kubenswrapper[4813]: I1201 08:41:20.782999 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:20 crc kubenswrapper[4813]: I1201 08:41:20.783018 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:20 crc kubenswrapper[4813]: I1201 08:41:20.783030 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:20Z","lastTransitionTime":"2025-12-01T08:41:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:20 crc kubenswrapper[4813]: E1201 08:41:20.795107 4813 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:41:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:41:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:20Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:41:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:41:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:20Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cd271ba1-a5c5-44e0-b03e-cfd1f5e8cb76\\\",\\\"systemUUID\\\":\\\"906e3783-eb6d-4f21-889c-7989aeaa25ba\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:20Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:20 crc kubenswrapper[4813]: I1201 08:41:20.798422 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:20 crc kubenswrapper[4813]: I1201 08:41:20.798469 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:20 crc kubenswrapper[4813]: I1201 08:41:20.798479 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:20 crc kubenswrapper[4813]: I1201 08:41:20.798511 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:20 crc kubenswrapper[4813]: I1201 08:41:20.798520 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:20Z","lastTransitionTime":"2025-12-01T08:41:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:20 crc kubenswrapper[4813]: E1201 08:41:20.811634 4813 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:41:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:41:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:20Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:41:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:41:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:20Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cd271ba1-a5c5-44e0-b03e-cfd1f5e8cb76\\\",\\\"systemUUID\\\":\\\"906e3783-eb6d-4f21-889c-7989aeaa25ba\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:20Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:20 crc kubenswrapper[4813]: I1201 08:41:20.815123 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:20 crc kubenswrapper[4813]: I1201 08:41:20.815165 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:20 crc kubenswrapper[4813]: I1201 08:41:20.815175 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:20 crc kubenswrapper[4813]: I1201 08:41:20.815194 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:20 crc kubenswrapper[4813]: I1201 08:41:20.815206 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:20Z","lastTransitionTime":"2025-12-01T08:41:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:20 crc kubenswrapper[4813]: E1201 08:41:20.828760 4813 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:41:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:41:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:20Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:41:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:41:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:20Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cd271ba1-a5c5-44e0-b03e-cfd1f5e8cb76\\\",\\\"systemUUID\\\":\\\"906e3783-eb6d-4f21-889c-7989aeaa25ba\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:20Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:20 crc kubenswrapper[4813]: I1201 08:41:20.832387 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:20 crc kubenswrapper[4813]: I1201 08:41:20.832427 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:20 crc kubenswrapper[4813]: I1201 08:41:20.832438 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:20 crc kubenswrapper[4813]: I1201 08:41:20.832453 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:20 crc kubenswrapper[4813]: I1201 08:41:20.832469 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:20Z","lastTransitionTime":"2025-12-01T08:41:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:20 crc kubenswrapper[4813]: E1201 08:41:20.844269 4813 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:41:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:41:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:20Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:41:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:41:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:20Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cd271ba1-a5c5-44e0-b03e-cfd1f5e8cb76\\\",\\\"systemUUID\\\":\\\"906e3783-eb6d-4f21-889c-7989aeaa25ba\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:20Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:20 crc kubenswrapper[4813]: I1201 08:41:20.847675 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:20 crc kubenswrapper[4813]: I1201 08:41:20.847724 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:20 crc kubenswrapper[4813]: I1201 08:41:20.847735 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:20 crc kubenswrapper[4813]: I1201 08:41:20.847751 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:20 crc kubenswrapper[4813]: I1201 08:41:20.847760 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:20Z","lastTransitionTime":"2025-12-01T08:41:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:20 crc kubenswrapper[4813]: E1201 08:41:20.859935 4813 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:41:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:41:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:20Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:41:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:41:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:20Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cd271ba1-a5c5-44e0-b03e-cfd1f5e8cb76\\\",\\\"systemUUID\\\":\\\"906e3783-eb6d-4f21-889c-7989aeaa25ba\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:20Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:20 crc kubenswrapper[4813]: E1201 08:41:20.860061 4813 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 01 08:41:20 crc kubenswrapper[4813]: I1201 08:41:20.861661 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:20 crc kubenswrapper[4813]: I1201 08:41:20.861709 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:20 crc kubenswrapper[4813]: I1201 08:41:20.861718 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:20 crc kubenswrapper[4813]: I1201 08:41:20.861734 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:20 crc kubenswrapper[4813]: I1201 08:41:20.861744 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:20Z","lastTransitionTime":"2025-12-01T08:41:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:20 crc kubenswrapper[4813]: I1201 08:41:20.963799 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:20 crc kubenswrapper[4813]: I1201 08:41:20.963845 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:20 crc kubenswrapper[4813]: I1201 08:41:20.963856 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:20 crc kubenswrapper[4813]: I1201 08:41:20.963873 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:20 crc kubenswrapper[4813]: I1201 08:41:20.963882 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:20Z","lastTransitionTime":"2025-12-01T08:41:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:21 crc kubenswrapper[4813]: I1201 08:41:21.066426 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:21 crc kubenswrapper[4813]: I1201 08:41:21.066475 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:21 crc kubenswrapper[4813]: I1201 08:41:21.066485 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:21 crc kubenswrapper[4813]: I1201 08:41:21.066502 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:21 crc kubenswrapper[4813]: I1201 08:41:21.066513 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:21Z","lastTransitionTime":"2025-12-01T08:41:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:21 crc kubenswrapper[4813]: I1201 08:41:21.168650 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:21 crc kubenswrapper[4813]: I1201 08:41:21.168699 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:21 crc kubenswrapper[4813]: I1201 08:41:21.168711 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:21 crc kubenswrapper[4813]: I1201 08:41:21.168731 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:21 crc kubenswrapper[4813]: I1201 08:41:21.168743 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:21Z","lastTransitionTime":"2025-12-01T08:41:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:21 crc kubenswrapper[4813]: I1201 08:41:21.271386 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:21 crc kubenswrapper[4813]: I1201 08:41:21.271427 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:21 crc kubenswrapper[4813]: I1201 08:41:21.271438 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:21 crc kubenswrapper[4813]: I1201 08:41:21.271455 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:21 crc kubenswrapper[4813]: I1201 08:41:21.271468 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:21Z","lastTransitionTime":"2025-12-01T08:41:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:21 crc kubenswrapper[4813]: I1201 08:41:21.374803 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:21 crc kubenswrapper[4813]: I1201 08:41:21.374879 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:21 crc kubenswrapper[4813]: I1201 08:41:21.374902 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:21 crc kubenswrapper[4813]: I1201 08:41:21.374931 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:21 crc kubenswrapper[4813]: I1201 08:41:21.374953 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:21Z","lastTransitionTime":"2025-12-01T08:41:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:21 crc kubenswrapper[4813]: I1201 08:41:21.393110 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-96mbr" Dec 01 08:41:21 crc kubenswrapper[4813]: I1201 08:41:21.393199 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 08:41:21 crc kubenswrapper[4813]: E1201 08:41:21.393244 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-96mbr" podUID="0789e3f7-5b08-4c04-8596-9ba29f50cdd0" Dec 01 08:41:21 crc kubenswrapper[4813]: E1201 08:41:21.393396 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 08:41:21 crc kubenswrapper[4813]: I1201 08:41:21.477409 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:21 crc kubenswrapper[4813]: I1201 08:41:21.477444 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:21 crc kubenswrapper[4813]: I1201 08:41:21.477452 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:21 crc kubenswrapper[4813]: I1201 08:41:21.477469 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:21 crc kubenswrapper[4813]: I1201 08:41:21.477481 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:21Z","lastTransitionTime":"2025-12-01T08:41:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:21 crc kubenswrapper[4813]: I1201 08:41:21.579934 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:21 crc kubenswrapper[4813]: I1201 08:41:21.580007 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:21 crc kubenswrapper[4813]: I1201 08:41:21.580021 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:21 crc kubenswrapper[4813]: I1201 08:41:21.580036 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:21 crc kubenswrapper[4813]: I1201 08:41:21.580047 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:21Z","lastTransitionTime":"2025-12-01T08:41:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:21 crc kubenswrapper[4813]: I1201 08:41:21.682466 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:21 crc kubenswrapper[4813]: I1201 08:41:21.682501 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:21 crc kubenswrapper[4813]: I1201 08:41:21.682510 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:21 crc kubenswrapper[4813]: I1201 08:41:21.682523 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:21 crc kubenswrapper[4813]: I1201 08:41:21.682534 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:21Z","lastTransitionTime":"2025-12-01T08:41:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:21 crc kubenswrapper[4813]: I1201 08:41:21.784624 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:21 crc kubenswrapper[4813]: I1201 08:41:21.784665 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:21 crc kubenswrapper[4813]: I1201 08:41:21.784675 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:21 crc kubenswrapper[4813]: I1201 08:41:21.784689 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:21 crc kubenswrapper[4813]: I1201 08:41:21.784698 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:21Z","lastTransitionTime":"2025-12-01T08:41:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:21 crc kubenswrapper[4813]: I1201 08:41:21.891020 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:21 crc kubenswrapper[4813]: I1201 08:41:21.891496 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:21 crc kubenswrapper[4813]: I1201 08:41:21.891509 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:21 crc kubenswrapper[4813]: I1201 08:41:21.891528 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:21 crc kubenswrapper[4813]: I1201 08:41:21.891588 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:21Z","lastTransitionTime":"2025-12-01T08:41:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:21 crc kubenswrapper[4813]: I1201 08:41:21.994146 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:21 crc kubenswrapper[4813]: I1201 08:41:21.994189 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:21 crc kubenswrapper[4813]: I1201 08:41:21.994200 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:21 crc kubenswrapper[4813]: I1201 08:41:21.994213 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:21 crc kubenswrapper[4813]: I1201 08:41:21.994222 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:21Z","lastTransitionTime":"2025-12-01T08:41:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:22 crc kubenswrapper[4813]: I1201 08:41:22.096796 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:22 crc kubenswrapper[4813]: I1201 08:41:22.096825 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:22 crc kubenswrapper[4813]: I1201 08:41:22.096833 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:22 crc kubenswrapper[4813]: I1201 08:41:22.096846 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:22 crc kubenswrapper[4813]: I1201 08:41:22.096854 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:22Z","lastTransitionTime":"2025-12-01T08:41:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:22 crc kubenswrapper[4813]: I1201 08:41:22.204028 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:22 crc kubenswrapper[4813]: I1201 08:41:22.204089 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:22 crc kubenswrapper[4813]: I1201 08:41:22.204106 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:22 crc kubenswrapper[4813]: I1201 08:41:22.204133 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:22 crc kubenswrapper[4813]: I1201 08:41:22.204146 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:22Z","lastTransitionTime":"2025-12-01T08:41:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:22 crc kubenswrapper[4813]: I1201 08:41:22.307578 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:22 crc kubenswrapper[4813]: I1201 08:41:22.307631 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:22 crc kubenswrapper[4813]: I1201 08:41:22.307652 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:22 crc kubenswrapper[4813]: I1201 08:41:22.307696 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:22 crc kubenswrapper[4813]: I1201 08:41:22.307716 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:22Z","lastTransitionTime":"2025-12-01T08:41:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:22 crc kubenswrapper[4813]: I1201 08:41:22.447604 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 08:41:22 crc kubenswrapper[4813]: I1201 08:41:22.447728 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 08:41:22 crc kubenswrapper[4813]: E1201 08:41:22.448588 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 08:41:22 crc kubenswrapper[4813]: E1201 08:41:22.448736 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 08:41:22 crc kubenswrapper[4813]: I1201 08:41:22.451170 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:22 crc kubenswrapper[4813]: I1201 08:41:22.451291 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:22 crc kubenswrapper[4813]: I1201 08:41:22.451388 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:22 crc kubenswrapper[4813]: I1201 08:41:22.451512 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:22 crc kubenswrapper[4813]: I1201 08:41:22.451622 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:22Z","lastTransitionTime":"2025-12-01T08:41:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:22 crc kubenswrapper[4813]: I1201 08:41:22.555217 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-j4pcc_b7505e14-3560-4d0d-be2c-d6a7a5953865/ovnkube-controller/2.log" Dec 01 08:41:22 crc kubenswrapper[4813]: I1201 08:41:22.555318 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:22 crc kubenswrapper[4813]: I1201 08:41:22.555344 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:22 crc kubenswrapper[4813]: I1201 08:41:22.555353 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:22 crc kubenswrapper[4813]: I1201 08:41:22.555366 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:22 crc kubenswrapper[4813]: I1201 08:41:22.555375 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:22Z","lastTransitionTime":"2025-12-01T08:41:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:22 crc kubenswrapper[4813]: I1201 08:41:22.556174 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-j4pcc_b7505e14-3560-4d0d-be2c-d6a7a5953865/ovnkube-controller/1.log" Dec 01 08:41:22 crc kubenswrapper[4813]: I1201 08:41:22.560188 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-j4pcc" event={"ID":"b7505e14-3560-4d0d-be2c-d6a7a5953865","Type":"ContainerDied","Data":"fb17fb9ae11fa62235328567418fb259f4f8540ebac635842d3afe751dcbde8c"} Dec 01 08:41:22 crc kubenswrapper[4813]: I1201 08:41:22.560159 4813 generic.go:334] "Generic (PLEG): container finished" podID="b7505e14-3560-4d0d-be2c-d6a7a5953865" containerID="fb17fb9ae11fa62235328567418fb259f4f8540ebac635842d3afe751dcbde8c" exitCode=1 Dec 01 08:41:22 crc kubenswrapper[4813]: I1201 08:41:22.560294 4813 scope.go:117] "RemoveContainer" containerID="8790f4730881303e7743498b66fc17b4b0df5b6620c8ca640ad7b9c90e6898d2" Dec 01 08:41:22 crc kubenswrapper[4813]: I1201 08:41:22.561232 4813 scope.go:117] "RemoveContainer" containerID="fb17fb9ae11fa62235328567418fb259f4f8540ebac635842d3afe751dcbde8c" Dec 01 08:41:22 crc kubenswrapper[4813]: E1201 08:41:22.561496 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-j4pcc_openshift-ovn-kubernetes(b7505e14-3560-4d0d-be2c-d6a7a5953865)\"" pod="openshift-ovn-kubernetes/ovnkube-node-j4pcc" podUID="b7505e14-3560-4d0d-be2c-d6a7a5953865" Dec 01 08:41:22 crc kubenswrapper[4813]: I1201 08:41:22.577486 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:22Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:22 crc kubenswrapper[4813]: I1201 08:41:22.588244 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-xpnlc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"22da1fcf-0c57-4708-90d8-77692af17344\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4a36d213bdad2da91d0d33f400d5a8248befb68d5b94d19347873a93c6712a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xc44d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-xpnlc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:22Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:22 crc kubenswrapper[4813]: I1201 08:41:22.599093 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6fb5a648-ae4f-4e46-a530-007e9bf84abb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e24ebfc52dcb28368a2d436e7d9ab78f07f2aec2a4c71405a6c7d3301eee96a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33898118bbf639c8e6eae5aafa10386b95795512444ed5f535db3c663f169df2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9902c00bf7738686c93f62e1772a484d05715d74b75ee98c84c8f23ed383824b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3657b47ec9f1563c83bfbe3d728c8fd123f7078e219d251b1874269e42c31eea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:16Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:22Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:22 crc kubenswrapper[4813]: I1201 08:41:22.610928 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0023b0e358d75aacc889ce364846dda89ea513832c87fa4b066aea7dac57e8a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:22Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:22 crc kubenswrapper[4813]: I1201 08:41:22.632591 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:22Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:22 crc kubenswrapper[4813]: I1201 08:41:22.644803 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c7574f892d5e9142296b26a0b484eb6759ab53f571139fb6e23b93dc3da05780\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:22Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:22 crc kubenswrapper[4813]: I1201 08:41:22.658151 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:22 crc kubenswrapper[4813]: I1201 08:41:22.658388 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:22 crc kubenswrapper[4813]: I1201 08:41:22.658530 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:22 crc kubenswrapper[4813]: I1201 08:41:22.658628 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:22 crc kubenswrapper[4813]: I1201 08:41:22.658704 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:22Z","lastTransitionTime":"2025-12-01T08:41:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:22 crc kubenswrapper[4813]: I1201 08:41:22.659181 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-96mbr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0789e3f7-5b08-4c04-8596-9ba29f50cdd0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xt4fj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xt4fj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:55Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-96mbr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:22Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:22 crc kubenswrapper[4813]: I1201 08:41:22.674665 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"349332bd-6b4d-4306-84ae-07019ca3860f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d441a348087ae3cadda1ad8a7d1c4dff02cf3f6e1d346c294b278fdd9131edcc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://da67635daf6d9e32937f4d5d4ceeee3f8f09ae4fd2dd26c3e85be1286c5949de\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e97356ee274ea78d9f237ffbc756abb6ac8df0dbac244bf0a86ed215b8cbf43\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8dc6feb9817f4f356b8e36af43ec3070774f312fbbb4909fa05b8b3e8dd53582\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a0404f5a5ed5a9750d4c479b608e68e8960b448f455d5af9d219aa5edaeb789b\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"message\\\":\\\"\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764578421\\\\\\\\\\\\\\\" (2025-12-01 08:40:21 +0000 UTC to 2025-12-31 08:40:22 +0000 UTC (now=2025-12-01 08:40:38.066068451 +0000 UTC))\\\\\\\"\\\\nI1201 08:40:38.066144 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1201 08:40:38.066166 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1201 08:40:38.066220 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1201 08:40:38.066232 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1201 08:40:38.066338 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764578432\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764578432\\\\\\\\\\\\\\\" (2025-12-01 07:40:32 +0000 UTC to 2026-12-01 07:40:32 +0000 UTC (now=2025-12-01 08:40:38.066301217 +0000 UTC))\\\\\\\"\\\\nI1201 08:40:38.066357 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1201 08:40:38.066385 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1201 08:40:38.066402 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1201 08:40:38.066414 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-63890811/tls.crt::/tmp/serving-cert-63890811/tls.key\\\\\\\"\\\\nI1201 08:40:38.066419 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1201 08:40:38.066572 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1201 08:40:38.067159 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7354f9337d42115bb1d974bf3772f2d03365acba57f2155135c7a0bbab39879e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:19Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05492df5703392cdf0932d6bf66025b18e53bd05fa0857a7cb2ad70979cbdc3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05492df5703392cdf0932d6bf66025b18e53bd05fa0857a7cb2ad70979cbdc3e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:16Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:22Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:22 crc kubenswrapper[4813]: I1201 08:41:22.691567 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3264b3af996af452df2953c59bea6deeb5ea6af1b119ce89f7334595c27e70da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9e879b1f563bd66383d58fd7b435a8a81bd50cb387dc1fabb6fc47d5a3fa468\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:22Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:22 crc kubenswrapper[4813]: I1201 08:41:22.705307 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-b6qqv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9656a03-0096-4c89-b0da-e37e4103611b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4195aebc1b2e388d480920a7f72257f6660afc99ad8c1a09478ef445739b272\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cjvnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:41Z\\\"}}\" for pod \"openshift-multus\"/\"multus-b6qqv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:22Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:22 crc kubenswrapper[4813]: I1201 08:41:22.717920 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-9wn4q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3d423402-e08f-42e9-b60c-6eb9bf2f6d48\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3049b842978c6580e1033a442a37766a2e76e278aface9a89fc0b2c57faf0df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rzxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c915733072b32a63237e2401cbf84b059aca396c2c0db884f0f439d599a35d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rzxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:41Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-9wn4q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:22Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:22 crc kubenswrapper[4813]: I1201 08:41:22.730721 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-8pg2p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dd53b38d-d9dd-4f98-9d06-dd95f1a4e3a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a214d71293e4e09573c665524d880fc319dfb1c0c6d95d11d78f253865660fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnjhq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ddf9d5989e4195b3074de08f99d53b48066f53fe93ce204145b3c714996fc57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnjhq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:53Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-8pg2p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:22Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:22 crc kubenswrapper[4813]: I1201 08:41:22.746102 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-h7lvd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d431b613-a91b-4fff-88b9-447804766289\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a32c51c943fcf6ada7e3f614667eb023e29d94c11128d26c1fecca37b03963b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://495593f99e92a9c28100a95e27d06ff5847008c5d996366cb474bedd9b69d5ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://495593f99e92a9c28100a95e27d06ff5847008c5d996366cb474bedd9b69d5ef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://19358f83a6c52957f959fe935a6843fdeef59905cc395e6e52a1fb4a670ac193\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://19358f83a6c52957f959fe935a6843fdeef59905cc395e6e52a1fb4a670ac193\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f089c115d916e339994879d224e3db27642b99aaabf59bcafb2979549052945\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9f089c115d916e339994879d224e3db27642b99aaabf59bcafb2979549052945\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb66f224462305c2005b31ef7bb9ed01efde8101e18518dd42d1dfb0df765bb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb66f224462305c2005b31ef7bb9ed01efde8101e18518dd42d1dfb0df765bb9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c268236c9a155c9b1f47c48d30dbd23948ceb2545592b40f356c60ea1d12e8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c268236c9a155c9b1f47c48d30dbd23948ceb2545592b40f356c60ea1d12e8a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://458aedca5f186e9aaed7e4ee29fd5247221491888b130f55c1fe48d88d565f8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://458aedca5f186e9aaed7e4ee29fd5247221491888b130f55c1fe48d88d565f8f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:41Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-h7lvd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:22Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:22 crc kubenswrapper[4813]: I1201 08:41:22.760589 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"490e526a-902b-4aff-b547-0cefea562764\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb48bd28fd3d29781695a5ae630f4913ef2745e97fbe7d463cca39a31e43ba0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74f8cf73f07b4cde260af429d26413e4e6226e7b177b493dd9c0e5c043ab6e1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3aaf696cefbe398bb6b26acfc5be43bda7e2be18218c8060840350c4106e66a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53de1f0682f36aa83ad8e5c226b95daf9f97df4ae154ebfc2fc287f9f238a4a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53de1f0682f36aa83ad8e5c226b95daf9f97df4ae154ebfc2fc287f9f238a4a8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:17Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:16Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:22Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:22 crc kubenswrapper[4813]: I1201 08:41:22.762258 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:22 crc kubenswrapper[4813]: I1201 08:41:22.762299 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:22 crc kubenswrapper[4813]: I1201 08:41:22.762311 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:22 crc kubenswrapper[4813]: I1201 08:41:22.762329 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:22 crc kubenswrapper[4813]: I1201 08:41:22.762341 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:22Z","lastTransitionTime":"2025-12-01T08:41:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:22 crc kubenswrapper[4813]: I1201 08:41:22.772396 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:22Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:22 crc kubenswrapper[4813]: I1201 08:41:22.782502 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-c576v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"17292cd0-1eac-40b5-a60e-67167c57409c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a42206c914d0493ca74788ea1a941de80920ab7b810fb95227cd9b8711e8f4b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7zwfx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:41Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-c576v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:22Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:22 crc kubenswrapper[4813]: I1201 08:41:22.801578 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-j4pcc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b7505e14-3560-4d0d-be2c-d6a7a5953865\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ccc30a63dc1c23ff9880fb04af1f31f0656b2696d359988e23deed23989f480a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://944977e9722025088720107530d0c91fc693593f380e4b49678f28a56da180cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc32d930ab23f30103d8fa42291909cbec89ebb75c5c53cd783cd27fe1a3f09e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f19c67b1c6c80c7f3223512102810d46ede1445464d21ae3bbea83424539b2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d984c90302e57f3e5a6f3ca3613b415db04ac74a5a228582f1b52ef1d59f93e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce8963bf1e9ee66926907aaf3a03d18977f4d0a1983d7121b62fbc8e02a37e4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb17fb9ae11fa62235328567418fb259f4f8540ebac635842d3afe751dcbde8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8790f4730881303e7743498b66fc17b4b0df5b6620c8ca640ad7b9c90e6898d2\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T08:40:56Z\\\",\\\"message\\\":\\\"I1201 08:40:56.782175 6307 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1201 08:40:56.782219 6307 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1201 08:40:56.782251 6307 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1201 08:40:56.782302 6307 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1201 08:40:56.782342 6307 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1201 08:40:56.782379 6307 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1201 08:40:56.782414 6307 factory.go:656] Stopping watch factory\\\\nI1201 08:40:56.782444 6307 ovnkube.go:599] Stopped ovnkube\\\\nI1201 08:40:56.782487 6307 handler.go:208] Removed *v1.Node event handler 7\\\\nI1201 08:40:56.782514 6307 handler.go:208] Removed *v1.Node event handler 2\\\\nI1201 08:40:56.782539 6307 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1201 08:40:56.782564 6307 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1201 08:40:56.782591 6307 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1201 08:40:56.782650 6307 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1201 08:40:56.782670 6307 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1201 08:40:56.782663 6307 handler.go:208] Removed *v1.EgressFirewa\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:55Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fb17fb9ae11fa62235328567418fb259f4f8540ebac635842d3afe751dcbde8c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T08:41:21Z\\\",\\\"message\\\":\\\"guring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-console-operator/metrics]} name:Service_openshift-console-operator/metrics_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.88:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {ebd4748e-0473-49fb-88ad-83dbb221791a}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1201 08:41:21.455187 6576 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1201 08:41:21.455499 6576 obj_retry.go:303] Retry object setup: *v1.Pod openshift-image-registry/node-ca-xpnlc\\\\nF1201 08:41:21.455509 6576 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal erro\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T08:41:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b45454b440c5108fa150a6b856afa1778ba6e9f89633433cb3117b68f00f72c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd67634e0c806323334495401a4e20b48fea4bc5013acd973d2abd3d70870f99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd67634e0c806323334495401a4e20b48fea4bc5013acd973d2abd3d70870f99\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:41Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-j4pcc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:22Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:22 crc kubenswrapper[4813]: I1201 08:41:22.865432 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:22 crc kubenswrapper[4813]: I1201 08:41:22.865727 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:22 crc kubenswrapper[4813]: I1201 08:41:22.865802 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:22 crc kubenswrapper[4813]: I1201 08:41:22.865875 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:22 crc kubenswrapper[4813]: I1201 08:41:22.865957 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:22Z","lastTransitionTime":"2025-12-01T08:41:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:22 crc kubenswrapper[4813]: I1201 08:41:22.969260 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:22 crc kubenswrapper[4813]: I1201 08:41:22.969538 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:22 crc kubenswrapper[4813]: I1201 08:41:22.969642 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:22 crc kubenswrapper[4813]: I1201 08:41:22.969749 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:22 crc kubenswrapper[4813]: I1201 08:41:22.969835 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:22Z","lastTransitionTime":"2025-12-01T08:41:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:23 crc kubenswrapper[4813]: I1201 08:41:23.073187 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:23 crc kubenswrapper[4813]: I1201 08:41:23.073441 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:23 crc kubenswrapper[4813]: I1201 08:41:23.073529 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:23 crc kubenswrapper[4813]: I1201 08:41:23.073620 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:23 crc kubenswrapper[4813]: I1201 08:41:23.073698 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:23Z","lastTransitionTime":"2025-12-01T08:41:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:23 crc kubenswrapper[4813]: I1201 08:41:23.176990 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:23 crc kubenswrapper[4813]: I1201 08:41:23.177337 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:23 crc kubenswrapper[4813]: I1201 08:41:23.177473 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:23 crc kubenswrapper[4813]: I1201 08:41:23.177615 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:23 crc kubenswrapper[4813]: I1201 08:41:23.177731 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:23Z","lastTransitionTime":"2025-12-01T08:41:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:23 crc kubenswrapper[4813]: I1201 08:41:23.279764 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:23 crc kubenswrapper[4813]: I1201 08:41:23.280034 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:23 crc kubenswrapper[4813]: I1201 08:41:23.280161 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:23 crc kubenswrapper[4813]: I1201 08:41:23.280263 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:23 crc kubenswrapper[4813]: I1201 08:41:23.280352 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:23Z","lastTransitionTime":"2025-12-01T08:41:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:23 crc kubenswrapper[4813]: I1201 08:41:23.383516 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:23 crc kubenswrapper[4813]: I1201 08:41:23.383560 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:23 crc kubenswrapper[4813]: I1201 08:41:23.383569 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:23 crc kubenswrapper[4813]: I1201 08:41:23.383587 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:23 crc kubenswrapper[4813]: I1201 08:41:23.383596 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:23Z","lastTransitionTime":"2025-12-01T08:41:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:23 crc kubenswrapper[4813]: I1201 08:41:23.392864 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-96mbr" Dec 01 08:41:23 crc kubenswrapper[4813]: I1201 08:41:23.392886 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 08:41:23 crc kubenswrapper[4813]: E1201 08:41:23.393159 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 08:41:23 crc kubenswrapper[4813]: E1201 08:41:23.393171 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-96mbr" podUID="0789e3f7-5b08-4c04-8596-9ba29f50cdd0" Dec 01 08:41:23 crc kubenswrapper[4813]: I1201 08:41:23.491094 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:23 crc kubenswrapper[4813]: I1201 08:41:23.491346 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:23 crc kubenswrapper[4813]: I1201 08:41:23.491369 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:23 crc kubenswrapper[4813]: I1201 08:41:23.491409 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:23 crc kubenswrapper[4813]: I1201 08:41:23.491430 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:23Z","lastTransitionTime":"2025-12-01T08:41:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:23 crc kubenswrapper[4813]: I1201 08:41:23.568147 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-j4pcc_b7505e14-3560-4d0d-be2c-d6a7a5953865/ovnkube-controller/2.log" Dec 01 08:41:23 crc kubenswrapper[4813]: I1201 08:41:23.593804 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:23 crc kubenswrapper[4813]: I1201 08:41:23.593838 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:23 crc kubenswrapper[4813]: I1201 08:41:23.593847 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:23 crc kubenswrapper[4813]: I1201 08:41:23.593862 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:23 crc kubenswrapper[4813]: I1201 08:41:23.593873 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:23Z","lastTransitionTime":"2025-12-01T08:41:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:23 crc kubenswrapper[4813]: I1201 08:41:23.697230 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:23 crc kubenswrapper[4813]: I1201 08:41:23.697477 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:23 crc kubenswrapper[4813]: I1201 08:41:23.697557 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:23 crc kubenswrapper[4813]: I1201 08:41:23.697627 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:23 crc kubenswrapper[4813]: I1201 08:41:23.697692 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:23Z","lastTransitionTime":"2025-12-01T08:41:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:23 crc kubenswrapper[4813]: I1201 08:41:23.801526 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:23 crc kubenswrapper[4813]: I1201 08:41:23.801568 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:23 crc kubenswrapper[4813]: I1201 08:41:23.801577 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:23 crc kubenswrapper[4813]: I1201 08:41:23.801595 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:23 crc kubenswrapper[4813]: I1201 08:41:23.801605 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:23Z","lastTransitionTime":"2025-12-01T08:41:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:23 crc kubenswrapper[4813]: I1201 08:41:23.904815 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:23 crc kubenswrapper[4813]: I1201 08:41:23.904848 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:23 crc kubenswrapper[4813]: I1201 08:41:23.904857 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:23 crc kubenswrapper[4813]: I1201 08:41:23.904870 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:23 crc kubenswrapper[4813]: I1201 08:41:23.904879 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:23Z","lastTransitionTime":"2025-12-01T08:41:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:24 crc kubenswrapper[4813]: I1201 08:41:24.018607 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:24 crc kubenswrapper[4813]: I1201 08:41:24.018688 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:24 crc kubenswrapper[4813]: I1201 08:41:24.018712 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:24 crc kubenswrapper[4813]: I1201 08:41:24.018742 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:24 crc kubenswrapper[4813]: I1201 08:41:24.018767 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:24Z","lastTransitionTime":"2025-12-01T08:41:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:24 crc kubenswrapper[4813]: I1201 08:41:24.120959 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:24 crc kubenswrapper[4813]: I1201 08:41:24.121019 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:24 crc kubenswrapper[4813]: I1201 08:41:24.121030 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:24 crc kubenswrapper[4813]: I1201 08:41:24.121050 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:24 crc kubenswrapper[4813]: I1201 08:41:24.121062 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:24Z","lastTransitionTime":"2025-12-01T08:41:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:24 crc kubenswrapper[4813]: I1201 08:41:24.223652 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:24 crc kubenswrapper[4813]: I1201 08:41:24.223981 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:24 crc kubenswrapper[4813]: I1201 08:41:24.224074 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:24 crc kubenswrapper[4813]: I1201 08:41:24.224180 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:24 crc kubenswrapper[4813]: I1201 08:41:24.224260 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:24Z","lastTransitionTime":"2025-12-01T08:41:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:24 crc kubenswrapper[4813]: I1201 08:41:24.328468 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:24 crc kubenswrapper[4813]: I1201 08:41:24.328527 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:24 crc kubenswrapper[4813]: I1201 08:41:24.328545 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:24 crc kubenswrapper[4813]: I1201 08:41:24.328569 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:24 crc kubenswrapper[4813]: I1201 08:41:24.328596 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:24Z","lastTransitionTime":"2025-12-01T08:41:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:24 crc kubenswrapper[4813]: I1201 08:41:24.393369 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 08:41:24 crc kubenswrapper[4813]: E1201 08:41:24.393584 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 08:41:24 crc kubenswrapper[4813]: I1201 08:41:24.393914 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 08:41:24 crc kubenswrapper[4813]: E1201 08:41:24.394068 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 08:41:24 crc kubenswrapper[4813]: I1201 08:41:24.434624 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:24 crc kubenswrapper[4813]: I1201 08:41:24.434737 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:24 crc kubenswrapper[4813]: I1201 08:41:24.434829 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:24 crc kubenswrapper[4813]: I1201 08:41:24.434925 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:24 crc kubenswrapper[4813]: I1201 08:41:24.435029 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:24Z","lastTransitionTime":"2025-12-01T08:41:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:24 crc kubenswrapper[4813]: I1201 08:41:24.538099 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:24 crc kubenswrapper[4813]: I1201 08:41:24.538194 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:24 crc kubenswrapper[4813]: I1201 08:41:24.538228 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:24 crc kubenswrapper[4813]: I1201 08:41:24.538259 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:24 crc kubenswrapper[4813]: I1201 08:41:24.538284 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:24Z","lastTransitionTime":"2025-12-01T08:41:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:24 crc kubenswrapper[4813]: I1201 08:41:24.640400 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:24 crc kubenswrapper[4813]: I1201 08:41:24.640457 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:24 crc kubenswrapper[4813]: I1201 08:41:24.640466 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:24 crc kubenswrapper[4813]: I1201 08:41:24.640484 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:24 crc kubenswrapper[4813]: I1201 08:41:24.640494 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:24Z","lastTransitionTime":"2025-12-01T08:41:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:24 crc kubenswrapper[4813]: I1201 08:41:24.743261 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:24 crc kubenswrapper[4813]: I1201 08:41:24.743303 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:24 crc kubenswrapper[4813]: I1201 08:41:24.743311 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:24 crc kubenswrapper[4813]: I1201 08:41:24.743324 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:24 crc kubenswrapper[4813]: I1201 08:41:24.743336 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:24Z","lastTransitionTime":"2025-12-01T08:41:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:24 crc kubenswrapper[4813]: I1201 08:41:24.846697 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:24 crc kubenswrapper[4813]: I1201 08:41:24.846740 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:24 crc kubenswrapper[4813]: I1201 08:41:24.846752 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:24 crc kubenswrapper[4813]: I1201 08:41:24.846769 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:24 crc kubenswrapper[4813]: I1201 08:41:24.846781 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:24Z","lastTransitionTime":"2025-12-01T08:41:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:24 crc kubenswrapper[4813]: I1201 08:41:24.949841 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:24 crc kubenswrapper[4813]: I1201 08:41:24.949883 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:24 crc kubenswrapper[4813]: I1201 08:41:24.949891 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:24 crc kubenswrapper[4813]: I1201 08:41:24.949906 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:24 crc kubenswrapper[4813]: I1201 08:41:24.949916 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:24Z","lastTransitionTime":"2025-12-01T08:41:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:25 crc kubenswrapper[4813]: I1201 08:41:25.052144 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:25 crc kubenswrapper[4813]: I1201 08:41:25.052171 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:25 crc kubenswrapper[4813]: I1201 08:41:25.052178 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:25 crc kubenswrapper[4813]: I1201 08:41:25.052193 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:25 crc kubenswrapper[4813]: I1201 08:41:25.052201 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:25Z","lastTransitionTime":"2025-12-01T08:41:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:25 crc kubenswrapper[4813]: I1201 08:41:25.154536 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:25 crc kubenswrapper[4813]: I1201 08:41:25.154583 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:25 crc kubenswrapper[4813]: I1201 08:41:25.154600 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:25 crc kubenswrapper[4813]: I1201 08:41:25.154624 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:25 crc kubenswrapper[4813]: I1201 08:41:25.154641 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:25Z","lastTransitionTime":"2025-12-01T08:41:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:25 crc kubenswrapper[4813]: I1201 08:41:25.257337 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:25 crc kubenswrapper[4813]: I1201 08:41:25.257366 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:25 crc kubenswrapper[4813]: I1201 08:41:25.257375 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:25 crc kubenswrapper[4813]: I1201 08:41:25.257389 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:25 crc kubenswrapper[4813]: I1201 08:41:25.257401 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:25Z","lastTransitionTime":"2025-12-01T08:41:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:25 crc kubenswrapper[4813]: I1201 08:41:25.360843 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:25 crc kubenswrapper[4813]: I1201 08:41:25.360886 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:25 crc kubenswrapper[4813]: I1201 08:41:25.360896 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:25 crc kubenswrapper[4813]: I1201 08:41:25.360914 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:25 crc kubenswrapper[4813]: I1201 08:41:25.360925 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:25Z","lastTransitionTime":"2025-12-01T08:41:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:25 crc kubenswrapper[4813]: I1201 08:41:25.393533 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 08:41:25 crc kubenswrapper[4813]: E1201 08:41:25.393660 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 08:41:25 crc kubenswrapper[4813]: I1201 08:41:25.393716 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-96mbr" Dec 01 08:41:25 crc kubenswrapper[4813]: E1201 08:41:25.393906 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-96mbr" podUID="0789e3f7-5b08-4c04-8596-9ba29f50cdd0" Dec 01 08:41:25 crc kubenswrapper[4813]: I1201 08:41:25.462797 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:25 crc kubenswrapper[4813]: I1201 08:41:25.462838 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:25 crc kubenswrapper[4813]: I1201 08:41:25.462846 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:25 crc kubenswrapper[4813]: I1201 08:41:25.462860 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:25 crc kubenswrapper[4813]: I1201 08:41:25.462870 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:25Z","lastTransitionTime":"2025-12-01T08:41:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:25 crc kubenswrapper[4813]: I1201 08:41:25.565483 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:25 crc kubenswrapper[4813]: I1201 08:41:25.565514 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:25 crc kubenswrapper[4813]: I1201 08:41:25.565552 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:25 crc kubenswrapper[4813]: I1201 08:41:25.565569 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:25 crc kubenswrapper[4813]: I1201 08:41:25.565604 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:25Z","lastTransitionTime":"2025-12-01T08:41:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:25 crc kubenswrapper[4813]: I1201 08:41:25.668634 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:25 crc kubenswrapper[4813]: I1201 08:41:25.668685 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:25 crc kubenswrapper[4813]: I1201 08:41:25.668703 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:25 crc kubenswrapper[4813]: I1201 08:41:25.668724 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:25 crc kubenswrapper[4813]: I1201 08:41:25.668736 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:25Z","lastTransitionTime":"2025-12-01T08:41:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:25 crc kubenswrapper[4813]: I1201 08:41:25.771409 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:25 crc kubenswrapper[4813]: I1201 08:41:25.771459 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:25 crc kubenswrapper[4813]: I1201 08:41:25.771470 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:25 crc kubenswrapper[4813]: I1201 08:41:25.771492 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:25 crc kubenswrapper[4813]: I1201 08:41:25.771506 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:25Z","lastTransitionTime":"2025-12-01T08:41:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:25 crc kubenswrapper[4813]: I1201 08:41:25.874095 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:25 crc kubenswrapper[4813]: I1201 08:41:25.874133 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:25 crc kubenswrapper[4813]: I1201 08:41:25.874144 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:25 crc kubenswrapper[4813]: I1201 08:41:25.874162 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:25 crc kubenswrapper[4813]: I1201 08:41:25.874174 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:25Z","lastTransitionTime":"2025-12-01T08:41:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:25 crc kubenswrapper[4813]: I1201 08:41:25.976958 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:25 crc kubenswrapper[4813]: I1201 08:41:25.977070 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:25 crc kubenswrapper[4813]: I1201 08:41:25.977091 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:25 crc kubenswrapper[4813]: I1201 08:41:25.977116 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:25 crc kubenswrapper[4813]: I1201 08:41:25.977136 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:25Z","lastTransitionTime":"2025-12-01T08:41:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:26 crc kubenswrapper[4813]: I1201 08:41:26.079506 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:26 crc kubenswrapper[4813]: I1201 08:41:26.079549 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:26 crc kubenswrapper[4813]: I1201 08:41:26.079567 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:26 crc kubenswrapper[4813]: I1201 08:41:26.079585 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:26 crc kubenswrapper[4813]: I1201 08:41:26.079596 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:26Z","lastTransitionTime":"2025-12-01T08:41:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:26 crc kubenswrapper[4813]: I1201 08:41:26.182322 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:26 crc kubenswrapper[4813]: I1201 08:41:26.182368 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:26 crc kubenswrapper[4813]: I1201 08:41:26.182378 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:26 crc kubenswrapper[4813]: I1201 08:41:26.182402 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:26 crc kubenswrapper[4813]: I1201 08:41:26.182417 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:26Z","lastTransitionTime":"2025-12-01T08:41:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:26 crc kubenswrapper[4813]: I1201 08:41:26.290015 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:26 crc kubenswrapper[4813]: I1201 08:41:26.290072 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:26 crc kubenswrapper[4813]: I1201 08:41:26.290085 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:26 crc kubenswrapper[4813]: I1201 08:41:26.290103 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:26 crc kubenswrapper[4813]: I1201 08:41:26.290116 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:26Z","lastTransitionTime":"2025-12-01T08:41:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:26 crc kubenswrapper[4813]: I1201 08:41:26.392191 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:26 crc kubenswrapper[4813]: I1201 08:41:26.392215 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:26 crc kubenswrapper[4813]: I1201 08:41:26.392223 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:26 crc kubenswrapper[4813]: I1201 08:41:26.392235 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:26 crc kubenswrapper[4813]: I1201 08:41:26.392243 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:26Z","lastTransitionTime":"2025-12-01T08:41:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:26 crc kubenswrapper[4813]: I1201 08:41:26.392718 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 08:41:26 crc kubenswrapper[4813]: I1201 08:41:26.392759 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 08:41:26 crc kubenswrapper[4813]: E1201 08:41:26.392825 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 08:41:26 crc kubenswrapper[4813]: E1201 08:41:26.392908 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 08:41:26 crc kubenswrapper[4813]: I1201 08:41:26.406454 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-9wn4q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3d423402-e08f-42e9-b60c-6eb9bf2f6d48\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3049b842978c6580e1033a442a37766a2e76e278aface9a89fc0b2c57faf0df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rzxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c915733072b32a63237e2401cbf84b059aca396c2c0db884f0f439d599a35d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rzxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:41Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-9wn4q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:26Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:26 crc kubenswrapper[4813]: I1201 08:41:26.420021 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-8pg2p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dd53b38d-d9dd-4f98-9d06-dd95f1a4e3a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a214d71293e4e09573c665524d880fc319dfb1c0c6d95d11d78f253865660fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnjhq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ddf9d5989e4195b3074de08f99d53b48066f53fe93ce204145b3c714996fc57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnjhq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:53Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-8pg2p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:26Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:26 crc kubenswrapper[4813]: I1201 08:41:26.434603 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:26Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:26 crc kubenswrapper[4813]: I1201 08:41:26.447381 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-c576v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"17292cd0-1eac-40b5-a60e-67167c57409c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a42206c914d0493ca74788ea1a941de80920ab7b810fb95227cd9b8711e8f4b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7zwfx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:41Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-c576v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:26Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:26 crc kubenswrapper[4813]: I1201 08:41:26.479858 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-j4pcc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b7505e14-3560-4d0d-be2c-d6a7a5953865\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ccc30a63dc1c23ff9880fb04af1f31f0656b2696d359988e23deed23989f480a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://944977e9722025088720107530d0c91fc693593f380e4b49678f28a56da180cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc32d930ab23f30103d8fa42291909cbec89ebb75c5c53cd783cd27fe1a3f09e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f19c67b1c6c80c7f3223512102810d46ede1445464d21ae3bbea83424539b2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d984c90302e57f3e5a6f3ca3613b415db04ac74a5a228582f1b52ef1d59f93e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce8963bf1e9ee66926907aaf3a03d18977f4d0a1983d7121b62fbc8e02a37e4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb17fb9ae11fa62235328567418fb259f4f8540ebac635842d3afe751dcbde8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8790f4730881303e7743498b66fc17b4b0df5b6620c8ca640ad7b9c90e6898d2\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T08:40:56Z\\\",\\\"message\\\":\\\"I1201 08:40:56.782175 6307 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1201 08:40:56.782219 6307 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1201 08:40:56.782251 6307 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1201 08:40:56.782302 6307 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1201 08:40:56.782342 6307 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1201 08:40:56.782379 6307 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1201 08:40:56.782414 6307 factory.go:656] Stopping watch factory\\\\nI1201 08:40:56.782444 6307 ovnkube.go:599] Stopped ovnkube\\\\nI1201 08:40:56.782487 6307 handler.go:208] Removed *v1.Node event handler 7\\\\nI1201 08:40:56.782514 6307 handler.go:208] Removed *v1.Node event handler 2\\\\nI1201 08:40:56.782539 6307 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1201 08:40:56.782564 6307 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1201 08:40:56.782591 6307 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1201 08:40:56.782650 6307 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1201 08:40:56.782670 6307 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1201 08:40:56.782663 6307 handler.go:208] Removed *v1.EgressFirewa\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:55Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fb17fb9ae11fa62235328567418fb259f4f8540ebac635842d3afe751dcbde8c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T08:41:21Z\\\",\\\"message\\\":\\\"guring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-console-operator/metrics]} name:Service_openshift-console-operator/metrics_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.88:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {ebd4748e-0473-49fb-88ad-83dbb221791a}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1201 08:41:21.455187 6576 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1201 08:41:21.455499 6576 obj_retry.go:303] Retry object setup: *v1.Pod openshift-image-registry/node-ca-xpnlc\\\\nF1201 08:41:21.455509 6576 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal erro\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T08:41:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b45454b440c5108fa150a6b856afa1778ba6e9f89633433cb3117b68f00f72c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd67634e0c806323334495401a4e20b48fea4bc5013acd973d2abd3d70870f99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd67634e0c806323334495401a4e20b48fea4bc5013acd973d2abd3d70870f99\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:41Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-j4pcc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:26Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:26 crc kubenswrapper[4813]: I1201 08:41:26.494124 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:26 crc kubenswrapper[4813]: I1201 08:41:26.494167 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:26 crc kubenswrapper[4813]: I1201 08:41:26.494178 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:26 crc kubenswrapper[4813]: I1201 08:41:26.494195 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:26 crc kubenswrapper[4813]: I1201 08:41:26.494208 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:26Z","lastTransitionTime":"2025-12-01T08:41:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:26 crc kubenswrapper[4813]: I1201 08:41:26.497878 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-h7lvd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d431b613-a91b-4fff-88b9-447804766289\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a32c51c943fcf6ada7e3f614667eb023e29d94c11128d26c1fecca37b03963b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://495593f99e92a9c28100a95e27d06ff5847008c5d996366cb474bedd9b69d5ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://495593f99e92a9c28100a95e27d06ff5847008c5d996366cb474bedd9b69d5ef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://19358f83a6c52957f959fe935a6843fdeef59905cc395e6e52a1fb4a670ac193\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://19358f83a6c52957f959fe935a6843fdeef59905cc395e6e52a1fb4a670ac193\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f089c115d916e339994879d224e3db27642b99aaabf59bcafb2979549052945\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9f089c115d916e339994879d224e3db27642b99aaabf59bcafb2979549052945\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb66f224462305c2005b31ef7bb9ed01efde8101e18518dd42d1dfb0df765bb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb66f224462305c2005b31ef7bb9ed01efde8101e18518dd42d1dfb0df765bb9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c268236c9a155c9b1f47c48d30dbd23948ceb2545592b40f356c60ea1d12e8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c268236c9a155c9b1f47c48d30dbd23948ceb2545592b40f356c60ea1d12e8a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://458aedca5f186e9aaed7e4ee29fd5247221491888b130f55c1fe48d88d565f8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://458aedca5f186e9aaed7e4ee29fd5247221491888b130f55c1fe48d88d565f8f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:41Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-h7lvd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:26Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:26 crc kubenswrapper[4813]: I1201 08:41:26.513496 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"490e526a-902b-4aff-b547-0cefea562764\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb48bd28fd3d29781695a5ae630f4913ef2745e97fbe7d463cca39a31e43ba0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74f8cf73f07b4cde260af429d26413e4e6226e7b177b493dd9c0e5c043ab6e1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3aaf696cefbe398bb6b26acfc5be43bda7e2be18218c8060840350c4106e66a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53de1f0682f36aa83ad8e5c226b95daf9f97df4ae154ebfc2fc287f9f238a4a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53de1f0682f36aa83ad8e5c226b95daf9f97df4ae154ebfc2fc287f9f238a4a8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:17Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:16Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:26Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:26 crc kubenswrapper[4813]: I1201 08:41:26.532440 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0023b0e358d75aacc889ce364846dda89ea513832c87fa4b066aea7dac57e8a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:26Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:26 crc kubenswrapper[4813]: I1201 08:41:26.546098 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:26Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:26 crc kubenswrapper[4813]: I1201 08:41:26.560186 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c7574f892d5e9142296b26a0b484eb6759ab53f571139fb6e23b93dc3da05780\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:26Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:26 crc kubenswrapper[4813]: I1201 08:41:26.570698 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:26Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:26 crc kubenswrapper[4813]: I1201 08:41:26.582909 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-xpnlc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"22da1fcf-0c57-4708-90d8-77692af17344\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4a36d213bdad2da91d0d33f400d5a8248befb68d5b94d19347873a93c6712a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xc44d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-xpnlc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:26Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:26 crc kubenswrapper[4813]: I1201 08:41:26.597026 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6fb5a648-ae4f-4e46-a530-007e9bf84abb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e24ebfc52dcb28368a2d436e7d9ab78f07f2aec2a4c71405a6c7d3301eee96a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33898118bbf639c8e6eae5aafa10386b95795512444ed5f535db3c663f169df2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9902c00bf7738686c93f62e1772a484d05715d74b75ee98c84c8f23ed383824b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3657b47ec9f1563c83bfbe3d728c8fd123f7078e219d251b1874269e42c31eea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:16Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:26Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:26 crc kubenswrapper[4813]: I1201 08:41:26.597230 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:26 crc kubenswrapper[4813]: I1201 08:41:26.597287 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:26 crc kubenswrapper[4813]: I1201 08:41:26.597310 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:26 crc kubenswrapper[4813]: I1201 08:41:26.597342 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:26 crc kubenswrapper[4813]: I1201 08:41:26.597360 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:26Z","lastTransitionTime":"2025-12-01T08:41:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:26 crc kubenswrapper[4813]: I1201 08:41:26.611305 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"349332bd-6b4d-4306-84ae-07019ca3860f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d441a348087ae3cadda1ad8a7d1c4dff02cf3f6e1d346c294b278fdd9131edcc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://da67635daf6d9e32937f4d5d4ceeee3f8f09ae4fd2dd26c3e85be1286c5949de\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e97356ee274ea78d9f237ffbc756abb6ac8df0dbac244bf0a86ed215b8cbf43\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8dc6feb9817f4f356b8e36af43ec3070774f312fbbb4909fa05b8b3e8dd53582\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a0404f5a5ed5a9750d4c479b608e68e8960b448f455d5af9d219aa5edaeb789b\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"message\\\":\\\"\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764578421\\\\\\\\\\\\\\\" (2025-12-01 08:40:21 +0000 UTC to 2025-12-31 08:40:22 +0000 UTC (now=2025-12-01 08:40:38.066068451 +0000 UTC))\\\\\\\"\\\\nI1201 08:40:38.066144 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1201 08:40:38.066166 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1201 08:40:38.066220 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1201 08:40:38.066232 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1201 08:40:38.066338 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764578432\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764578432\\\\\\\\\\\\\\\" (2025-12-01 07:40:32 +0000 UTC to 2026-12-01 07:40:32 +0000 UTC (now=2025-12-01 08:40:38.066301217 +0000 UTC))\\\\\\\"\\\\nI1201 08:40:38.066357 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1201 08:40:38.066385 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1201 08:40:38.066402 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1201 08:40:38.066414 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-63890811/tls.crt::/tmp/serving-cert-63890811/tls.key\\\\\\\"\\\\nI1201 08:40:38.066419 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1201 08:40:38.066572 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1201 08:40:38.067159 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7354f9337d42115bb1d974bf3772f2d03365acba57f2155135c7a0bbab39879e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:19Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05492df5703392cdf0932d6bf66025b18e53bd05fa0857a7cb2ad70979cbdc3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05492df5703392cdf0932d6bf66025b18e53bd05fa0857a7cb2ad70979cbdc3e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:16Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:26Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:26 crc kubenswrapper[4813]: I1201 08:41:26.627166 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3264b3af996af452df2953c59bea6deeb5ea6af1b119ce89f7334595c27e70da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9e879b1f563bd66383d58fd7b435a8a81bd50cb387dc1fabb6fc47d5a3fa468\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:26Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:26 crc kubenswrapper[4813]: I1201 08:41:26.651806 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-b6qqv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9656a03-0096-4c89-b0da-e37e4103611b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4195aebc1b2e388d480920a7f72257f6660afc99ad8c1a09478ef445739b272\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cjvnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:41Z\\\"}}\" for pod \"openshift-multus\"/\"multus-b6qqv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:26Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:26 crc kubenswrapper[4813]: I1201 08:41:26.664190 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-96mbr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0789e3f7-5b08-4c04-8596-9ba29f50cdd0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xt4fj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xt4fj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:55Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-96mbr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:26Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:26 crc kubenswrapper[4813]: I1201 08:41:26.699583 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:26 crc kubenswrapper[4813]: I1201 08:41:26.699633 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:26 crc kubenswrapper[4813]: I1201 08:41:26.699649 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:26 crc kubenswrapper[4813]: I1201 08:41:26.699669 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:26 crc kubenswrapper[4813]: I1201 08:41:26.699683 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:26Z","lastTransitionTime":"2025-12-01T08:41:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:26 crc kubenswrapper[4813]: I1201 08:41:26.805187 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:26 crc kubenswrapper[4813]: I1201 08:41:26.805231 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:26 crc kubenswrapper[4813]: I1201 08:41:26.805241 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:26 crc kubenswrapper[4813]: I1201 08:41:26.805260 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:26 crc kubenswrapper[4813]: I1201 08:41:26.805271 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:26Z","lastTransitionTime":"2025-12-01T08:41:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:26 crc kubenswrapper[4813]: I1201 08:41:26.907328 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:26 crc kubenswrapper[4813]: I1201 08:41:26.907366 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:26 crc kubenswrapper[4813]: I1201 08:41:26.907376 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:26 crc kubenswrapper[4813]: I1201 08:41:26.907394 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:26 crc kubenswrapper[4813]: I1201 08:41:26.907405 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:26Z","lastTransitionTime":"2025-12-01T08:41:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:27 crc kubenswrapper[4813]: I1201 08:41:27.010090 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:27 crc kubenswrapper[4813]: I1201 08:41:27.010128 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:27 crc kubenswrapper[4813]: I1201 08:41:27.010139 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:27 crc kubenswrapper[4813]: I1201 08:41:27.010157 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:27 crc kubenswrapper[4813]: I1201 08:41:27.010171 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:27Z","lastTransitionTime":"2025-12-01T08:41:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:27 crc kubenswrapper[4813]: I1201 08:41:27.112140 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:27 crc kubenswrapper[4813]: I1201 08:41:27.112175 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:27 crc kubenswrapper[4813]: I1201 08:41:27.112183 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:27 crc kubenswrapper[4813]: I1201 08:41:27.112196 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:27 crc kubenswrapper[4813]: I1201 08:41:27.112205 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:27Z","lastTransitionTime":"2025-12-01T08:41:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:27 crc kubenswrapper[4813]: I1201 08:41:27.214445 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:27 crc kubenswrapper[4813]: I1201 08:41:27.214478 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:27 crc kubenswrapper[4813]: I1201 08:41:27.214487 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:27 crc kubenswrapper[4813]: I1201 08:41:27.214500 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:27 crc kubenswrapper[4813]: I1201 08:41:27.214510 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:27Z","lastTransitionTime":"2025-12-01T08:41:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:27 crc kubenswrapper[4813]: I1201 08:41:27.320519 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:27 crc kubenswrapper[4813]: I1201 08:41:27.320586 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:27 crc kubenswrapper[4813]: I1201 08:41:27.320597 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:27 crc kubenswrapper[4813]: I1201 08:41:27.320615 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:27 crc kubenswrapper[4813]: I1201 08:41:27.320631 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:27Z","lastTransitionTime":"2025-12-01T08:41:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:27 crc kubenswrapper[4813]: I1201 08:41:27.393269 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-96mbr" Dec 01 08:41:27 crc kubenswrapper[4813]: E1201 08:41:27.393434 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-96mbr" podUID="0789e3f7-5b08-4c04-8596-9ba29f50cdd0" Dec 01 08:41:27 crc kubenswrapper[4813]: I1201 08:41:27.393538 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 08:41:27 crc kubenswrapper[4813]: E1201 08:41:27.393843 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 08:41:27 crc kubenswrapper[4813]: I1201 08:41:27.423825 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:27 crc kubenswrapper[4813]: I1201 08:41:27.423864 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:27 crc kubenswrapper[4813]: I1201 08:41:27.423875 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:27 crc kubenswrapper[4813]: I1201 08:41:27.423891 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:27 crc kubenswrapper[4813]: I1201 08:41:27.423901 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:27Z","lastTransitionTime":"2025-12-01T08:41:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:27 crc kubenswrapper[4813]: I1201 08:41:27.527230 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:27 crc kubenswrapper[4813]: I1201 08:41:27.527294 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:27 crc kubenswrapper[4813]: I1201 08:41:27.527308 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:27 crc kubenswrapper[4813]: I1201 08:41:27.527324 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:27 crc kubenswrapper[4813]: I1201 08:41:27.527333 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:27Z","lastTransitionTime":"2025-12-01T08:41:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:27 crc kubenswrapper[4813]: I1201 08:41:27.630059 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:27 crc kubenswrapper[4813]: I1201 08:41:27.630122 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:27 crc kubenswrapper[4813]: I1201 08:41:27.630139 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:27 crc kubenswrapper[4813]: I1201 08:41:27.630167 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:27 crc kubenswrapper[4813]: I1201 08:41:27.630182 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:27Z","lastTransitionTime":"2025-12-01T08:41:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:27 crc kubenswrapper[4813]: I1201 08:41:27.733485 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:27 crc kubenswrapper[4813]: I1201 08:41:27.733649 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:27 crc kubenswrapper[4813]: I1201 08:41:27.733678 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:27 crc kubenswrapper[4813]: I1201 08:41:27.733712 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:27 crc kubenswrapper[4813]: I1201 08:41:27.733737 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:27Z","lastTransitionTime":"2025-12-01T08:41:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:27 crc kubenswrapper[4813]: I1201 08:41:27.809762 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0789e3f7-5b08-4c04-8596-9ba29f50cdd0-metrics-certs\") pod \"network-metrics-daemon-96mbr\" (UID: \"0789e3f7-5b08-4c04-8596-9ba29f50cdd0\") " pod="openshift-multus/network-metrics-daemon-96mbr" Dec 01 08:41:27 crc kubenswrapper[4813]: E1201 08:41:27.809947 4813 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 01 08:41:27 crc kubenswrapper[4813]: E1201 08:41:27.810071 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/0789e3f7-5b08-4c04-8596-9ba29f50cdd0-metrics-certs podName:0789e3f7-5b08-4c04-8596-9ba29f50cdd0 nodeName:}" failed. No retries permitted until 2025-12-01 08:41:59.810036124 +0000 UTC m=+104.432957710 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/0789e3f7-5b08-4c04-8596-9ba29f50cdd0-metrics-certs") pod "network-metrics-daemon-96mbr" (UID: "0789e3f7-5b08-4c04-8596-9ba29f50cdd0") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 01 08:41:27 crc kubenswrapper[4813]: I1201 08:41:27.837766 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:27 crc kubenswrapper[4813]: I1201 08:41:27.837825 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:27 crc kubenswrapper[4813]: I1201 08:41:27.837843 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:27 crc kubenswrapper[4813]: I1201 08:41:27.837872 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:27 crc kubenswrapper[4813]: I1201 08:41:27.837890 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:27Z","lastTransitionTime":"2025-12-01T08:41:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:27 crc kubenswrapper[4813]: I1201 08:41:27.941132 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:27 crc kubenswrapper[4813]: I1201 08:41:27.941172 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:27 crc kubenswrapper[4813]: I1201 08:41:27.941182 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:27 crc kubenswrapper[4813]: I1201 08:41:27.941199 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:27 crc kubenswrapper[4813]: I1201 08:41:27.941213 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:27Z","lastTransitionTime":"2025-12-01T08:41:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:28 crc kubenswrapper[4813]: I1201 08:41:28.044191 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:28 crc kubenswrapper[4813]: I1201 08:41:28.044267 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:28 crc kubenswrapper[4813]: I1201 08:41:28.044278 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:28 crc kubenswrapper[4813]: I1201 08:41:28.044301 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:28 crc kubenswrapper[4813]: I1201 08:41:28.044312 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:28Z","lastTransitionTime":"2025-12-01T08:41:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:28 crc kubenswrapper[4813]: I1201 08:41:28.147522 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:28 crc kubenswrapper[4813]: I1201 08:41:28.147558 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:28 crc kubenswrapper[4813]: I1201 08:41:28.147566 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:28 crc kubenswrapper[4813]: I1201 08:41:28.147581 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:28 crc kubenswrapper[4813]: I1201 08:41:28.147590 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:28Z","lastTransitionTime":"2025-12-01T08:41:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:28 crc kubenswrapper[4813]: I1201 08:41:28.249700 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:28 crc kubenswrapper[4813]: I1201 08:41:28.249739 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:28 crc kubenswrapper[4813]: I1201 08:41:28.249748 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:28 crc kubenswrapper[4813]: I1201 08:41:28.249762 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:28 crc kubenswrapper[4813]: I1201 08:41:28.249772 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:28Z","lastTransitionTime":"2025-12-01T08:41:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:28 crc kubenswrapper[4813]: I1201 08:41:28.352303 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:28 crc kubenswrapper[4813]: I1201 08:41:28.352359 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:28 crc kubenswrapper[4813]: I1201 08:41:28.352375 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:28 crc kubenswrapper[4813]: I1201 08:41:28.352401 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:28 crc kubenswrapper[4813]: I1201 08:41:28.352419 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:28Z","lastTransitionTime":"2025-12-01T08:41:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:28 crc kubenswrapper[4813]: I1201 08:41:28.394339 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 08:41:28 crc kubenswrapper[4813]: I1201 08:41:28.394365 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 08:41:28 crc kubenswrapper[4813]: E1201 08:41:28.394523 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 08:41:28 crc kubenswrapper[4813]: E1201 08:41:28.394711 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 08:41:28 crc kubenswrapper[4813]: I1201 08:41:28.455313 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:28 crc kubenswrapper[4813]: I1201 08:41:28.455357 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:28 crc kubenswrapper[4813]: I1201 08:41:28.455367 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:28 crc kubenswrapper[4813]: I1201 08:41:28.455381 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:28 crc kubenswrapper[4813]: I1201 08:41:28.455391 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:28Z","lastTransitionTime":"2025-12-01T08:41:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:28 crc kubenswrapper[4813]: I1201 08:41:28.558028 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:28 crc kubenswrapper[4813]: I1201 08:41:28.558061 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:28 crc kubenswrapper[4813]: I1201 08:41:28.558100 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:28 crc kubenswrapper[4813]: I1201 08:41:28.558141 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:28 crc kubenswrapper[4813]: I1201 08:41:28.558156 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:28Z","lastTransitionTime":"2025-12-01T08:41:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:28 crc kubenswrapper[4813]: I1201 08:41:28.660749 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:28 crc kubenswrapper[4813]: I1201 08:41:28.660783 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:28 crc kubenswrapper[4813]: I1201 08:41:28.660795 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:28 crc kubenswrapper[4813]: I1201 08:41:28.660815 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:28 crc kubenswrapper[4813]: I1201 08:41:28.660829 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:28Z","lastTransitionTime":"2025-12-01T08:41:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:28 crc kubenswrapper[4813]: I1201 08:41:28.764239 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:28 crc kubenswrapper[4813]: I1201 08:41:28.764281 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:28 crc kubenswrapper[4813]: I1201 08:41:28.764290 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:28 crc kubenswrapper[4813]: I1201 08:41:28.764306 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:28 crc kubenswrapper[4813]: I1201 08:41:28.764317 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:28Z","lastTransitionTime":"2025-12-01T08:41:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:28 crc kubenswrapper[4813]: I1201 08:41:28.866721 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:28 crc kubenswrapper[4813]: I1201 08:41:28.866765 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:28 crc kubenswrapper[4813]: I1201 08:41:28.866785 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:28 crc kubenswrapper[4813]: I1201 08:41:28.866805 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:28 crc kubenswrapper[4813]: I1201 08:41:28.866816 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:28Z","lastTransitionTime":"2025-12-01T08:41:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:28 crc kubenswrapper[4813]: I1201 08:41:28.971187 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:28 crc kubenswrapper[4813]: I1201 08:41:28.971238 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:28 crc kubenswrapper[4813]: I1201 08:41:28.971253 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:28 crc kubenswrapper[4813]: I1201 08:41:28.971280 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:28 crc kubenswrapper[4813]: I1201 08:41:28.971296 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:28Z","lastTransitionTime":"2025-12-01T08:41:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:29 crc kubenswrapper[4813]: I1201 08:41:29.074486 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:29 crc kubenswrapper[4813]: I1201 08:41:29.074535 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:29 crc kubenswrapper[4813]: I1201 08:41:29.074544 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:29 crc kubenswrapper[4813]: I1201 08:41:29.074560 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:29 crc kubenswrapper[4813]: I1201 08:41:29.074570 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:29Z","lastTransitionTime":"2025-12-01T08:41:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:29 crc kubenswrapper[4813]: I1201 08:41:29.177307 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:29 crc kubenswrapper[4813]: I1201 08:41:29.177357 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:29 crc kubenswrapper[4813]: I1201 08:41:29.177368 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:29 crc kubenswrapper[4813]: I1201 08:41:29.177385 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:29 crc kubenswrapper[4813]: I1201 08:41:29.177397 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:29Z","lastTransitionTime":"2025-12-01T08:41:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:29 crc kubenswrapper[4813]: I1201 08:41:29.280172 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:29 crc kubenswrapper[4813]: I1201 08:41:29.280230 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:29 crc kubenswrapper[4813]: I1201 08:41:29.280243 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:29 crc kubenswrapper[4813]: I1201 08:41:29.280268 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:29 crc kubenswrapper[4813]: I1201 08:41:29.280283 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:29Z","lastTransitionTime":"2025-12-01T08:41:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:29 crc kubenswrapper[4813]: I1201 08:41:29.383077 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:29 crc kubenswrapper[4813]: I1201 08:41:29.383112 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:29 crc kubenswrapper[4813]: I1201 08:41:29.383119 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:29 crc kubenswrapper[4813]: I1201 08:41:29.383132 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:29 crc kubenswrapper[4813]: I1201 08:41:29.383142 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:29Z","lastTransitionTime":"2025-12-01T08:41:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:29 crc kubenswrapper[4813]: I1201 08:41:29.393344 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 08:41:29 crc kubenswrapper[4813]: I1201 08:41:29.393409 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-96mbr" Dec 01 08:41:29 crc kubenswrapper[4813]: E1201 08:41:29.393457 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 08:41:29 crc kubenswrapper[4813]: E1201 08:41:29.393560 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-96mbr" podUID="0789e3f7-5b08-4c04-8596-9ba29f50cdd0" Dec 01 08:41:29 crc kubenswrapper[4813]: I1201 08:41:29.486133 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:29 crc kubenswrapper[4813]: I1201 08:41:29.486203 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:29 crc kubenswrapper[4813]: I1201 08:41:29.486227 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:29 crc kubenswrapper[4813]: I1201 08:41:29.486252 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:29 crc kubenswrapper[4813]: I1201 08:41:29.486269 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:29Z","lastTransitionTime":"2025-12-01T08:41:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:29 crc kubenswrapper[4813]: I1201 08:41:29.590350 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:29 crc kubenswrapper[4813]: I1201 08:41:29.590481 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:29 crc kubenswrapper[4813]: I1201 08:41:29.590522 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:29 crc kubenswrapper[4813]: I1201 08:41:29.590556 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:29 crc kubenswrapper[4813]: I1201 08:41:29.590580 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:29Z","lastTransitionTime":"2025-12-01T08:41:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:29 crc kubenswrapper[4813]: I1201 08:41:29.693767 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:29 crc kubenswrapper[4813]: I1201 08:41:29.693807 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:29 crc kubenswrapper[4813]: I1201 08:41:29.693818 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:29 crc kubenswrapper[4813]: I1201 08:41:29.693833 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:29 crc kubenswrapper[4813]: I1201 08:41:29.693842 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:29Z","lastTransitionTime":"2025-12-01T08:41:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:29 crc kubenswrapper[4813]: I1201 08:41:29.796904 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:29 crc kubenswrapper[4813]: I1201 08:41:29.796947 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:29 crc kubenswrapper[4813]: I1201 08:41:29.796955 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:29 crc kubenswrapper[4813]: I1201 08:41:29.796990 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:29 crc kubenswrapper[4813]: I1201 08:41:29.796999 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:29Z","lastTransitionTime":"2025-12-01T08:41:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:29 crc kubenswrapper[4813]: I1201 08:41:29.900278 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:29 crc kubenswrapper[4813]: I1201 08:41:29.900310 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:29 crc kubenswrapper[4813]: I1201 08:41:29.900321 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:29 crc kubenswrapper[4813]: I1201 08:41:29.900335 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:29 crc kubenswrapper[4813]: I1201 08:41:29.900343 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:29Z","lastTransitionTime":"2025-12-01T08:41:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:30 crc kubenswrapper[4813]: I1201 08:41:30.003499 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:30 crc kubenswrapper[4813]: I1201 08:41:30.003560 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:30 crc kubenswrapper[4813]: I1201 08:41:30.003580 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:30 crc kubenswrapper[4813]: I1201 08:41:30.003607 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:30 crc kubenswrapper[4813]: I1201 08:41:30.003627 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:30Z","lastTransitionTime":"2025-12-01T08:41:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:30 crc kubenswrapper[4813]: I1201 08:41:30.107136 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:30 crc kubenswrapper[4813]: I1201 08:41:30.107184 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:30 crc kubenswrapper[4813]: I1201 08:41:30.107199 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:30 crc kubenswrapper[4813]: I1201 08:41:30.107219 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:30 crc kubenswrapper[4813]: I1201 08:41:30.107230 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:30Z","lastTransitionTime":"2025-12-01T08:41:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:30 crc kubenswrapper[4813]: I1201 08:41:30.210914 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:30 crc kubenswrapper[4813]: I1201 08:41:30.210959 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:30 crc kubenswrapper[4813]: I1201 08:41:30.210989 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:30 crc kubenswrapper[4813]: I1201 08:41:30.211004 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:30 crc kubenswrapper[4813]: I1201 08:41:30.211012 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:30Z","lastTransitionTime":"2025-12-01T08:41:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:30 crc kubenswrapper[4813]: I1201 08:41:30.314540 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:30 crc kubenswrapper[4813]: I1201 08:41:30.314617 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:30 crc kubenswrapper[4813]: I1201 08:41:30.314627 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:30 crc kubenswrapper[4813]: I1201 08:41:30.314648 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:30 crc kubenswrapper[4813]: I1201 08:41:30.314660 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:30Z","lastTransitionTime":"2025-12-01T08:41:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:30 crc kubenswrapper[4813]: I1201 08:41:30.393742 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 08:41:30 crc kubenswrapper[4813]: I1201 08:41:30.394062 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 08:41:30 crc kubenswrapper[4813]: E1201 08:41:30.394076 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 08:41:30 crc kubenswrapper[4813]: E1201 08:41:30.394635 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 08:41:30 crc kubenswrapper[4813]: I1201 08:41:30.417246 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:30 crc kubenswrapper[4813]: I1201 08:41:30.417294 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:30 crc kubenswrapper[4813]: I1201 08:41:30.417307 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:30 crc kubenswrapper[4813]: I1201 08:41:30.417326 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:30 crc kubenswrapper[4813]: I1201 08:41:30.417340 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:30Z","lastTransitionTime":"2025-12-01T08:41:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:30 crc kubenswrapper[4813]: I1201 08:41:30.520696 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:30 crc kubenswrapper[4813]: I1201 08:41:30.520773 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:30 crc kubenswrapper[4813]: I1201 08:41:30.520791 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:30 crc kubenswrapper[4813]: I1201 08:41:30.520816 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:30 crc kubenswrapper[4813]: I1201 08:41:30.520838 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:30Z","lastTransitionTime":"2025-12-01T08:41:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:30 crc kubenswrapper[4813]: I1201 08:41:30.625321 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:30 crc kubenswrapper[4813]: I1201 08:41:30.625356 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:30 crc kubenswrapper[4813]: I1201 08:41:30.625366 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:30 crc kubenswrapper[4813]: I1201 08:41:30.625383 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:30 crc kubenswrapper[4813]: I1201 08:41:30.625394 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:30Z","lastTransitionTime":"2025-12-01T08:41:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:30 crc kubenswrapper[4813]: I1201 08:41:30.728373 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:30 crc kubenswrapper[4813]: I1201 08:41:30.728424 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:30 crc kubenswrapper[4813]: I1201 08:41:30.728445 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:30 crc kubenswrapper[4813]: I1201 08:41:30.728477 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:30 crc kubenswrapper[4813]: I1201 08:41:30.728501 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:30Z","lastTransitionTime":"2025-12-01T08:41:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:30 crc kubenswrapper[4813]: I1201 08:41:30.831435 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:30 crc kubenswrapper[4813]: I1201 08:41:30.831468 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:30 crc kubenswrapper[4813]: I1201 08:41:30.831481 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:30 crc kubenswrapper[4813]: I1201 08:41:30.831499 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:30 crc kubenswrapper[4813]: I1201 08:41:30.831511 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:30Z","lastTransitionTime":"2025-12-01T08:41:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:30 crc kubenswrapper[4813]: I1201 08:41:30.934821 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:30 crc kubenswrapper[4813]: I1201 08:41:30.934871 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:30 crc kubenswrapper[4813]: I1201 08:41:30.934885 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:30 crc kubenswrapper[4813]: I1201 08:41:30.934904 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:30 crc kubenswrapper[4813]: I1201 08:41:30.934915 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:30Z","lastTransitionTime":"2025-12-01T08:41:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:30 crc kubenswrapper[4813]: I1201 08:41:30.994561 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:30 crc kubenswrapper[4813]: I1201 08:41:30.994616 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:30 crc kubenswrapper[4813]: I1201 08:41:30.994638 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:30 crc kubenswrapper[4813]: I1201 08:41:30.994666 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:30 crc kubenswrapper[4813]: I1201 08:41:30.994687 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:30Z","lastTransitionTime":"2025-12-01T08:41:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:31 crc kubenswrapper[4813]: E1201 08:41:31.013534 4813 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:41:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:41:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:30Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:41:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:41:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:30Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cd271ba1-a5c5-44e0-b03e-cfd1f5e8cb76\\\",\\\"systemUUID\\\":\\\"906e3783-eb6d-4f21-889c-7989aeaa25ba\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:31Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:31 crc kubenswrapper[4813]: I1201 08:41:31.019244 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:31 crc kubenswrapper[4813]: I1201 08:41:31.019305 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:31 crc kubenswrapper[4813]: I1201 08:41:31.019332 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:31 crc kubenswrapper[4813]: I1201 08:41:31.019363 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:31 crc kubenswrapper[4813]: I1201 08:41:31.019384 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:31Z","lastTransitionTime":"2025-12-01T08:41:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:31 crc kubenswrapper[4813]: E1201 08:41:31.040191 4813 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:41:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:41:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:31Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:41:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:41:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:31Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cd271ba1-a5c5-44e0-b03e-cfd1f5e8cb76\\\",\\\"systemUUID\\\":\\\"906e3783-eb6d-4f21-889c-7989aeaa25ba\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:31Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:31 crc kubenswrapper[4813]: I1201 08:41:31.044571 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:31 crc kubenswrapper[4813]: I1201 08:41:31.044620 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:31 crc kubenswrapper[4813]: I1201 08:41:31.044638 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:31 crc kubenswrapper[4813]: I1201 08:41:31.044663 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:31 crc kubenswrapper[4813]: I1201 08:41:31.044679 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:31Z","lastTransitionTime":"2025-12-01T08:41:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:31 crc kubenswrapper[4813]: E1201 08:41:31.068607 4813 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:41:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:41:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:31Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:41:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:41:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:31Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cd271ba1-a5c5-44e0-b03e-cfd1f5e8cb76\\\",\\\"systemUUID\\\":\\\"906e3783-eb6d-4f21-889c-7989aeaa25ba\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:31Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:31 crc kubenswrapper[4813]: I1201 08:41:31.074737 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:31 crc kubenswrapper[4813]: I1201 08:41:31.074844 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:31 crc kubenswrapper[4813]: I1201 08:41:31.074909 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:31 crc kubenswrapper[4813]: I1201 08:41:31.074945 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:31 crc kubenswrapper[4813]: I1201 08:41:31.074988 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:31Z","lastTransitionTime":"2025-12-01T08:41:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:31 crc kubenswrapper[4813]: E1201 08:41:31.095837 4813 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:41:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:41:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:31Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:41:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:41:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:31Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cd271ba1-a5c5-44e0-b03e-cfd1f5e8cb76\\\",\\\"systemUUID\\\":\\\"906e3783-eb6d-4f21-889c-7989aeaa25ba\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:31Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:31 crc kubenswrapper[4813]: I1201 08:41:31.100886 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:31 crc kubenswrapper[4813]: I1201 08:41:31.100930 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:31 crc kubenswrapper[4813]: I1201 08:41:31.100948 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:31 crc kubenswrapper[4813]: I1201 08:41:31.101005 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:31 crc kubenswrapper[4813]: I1201 08:41:31.101024 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:31Z","lastTransitionTime":"2025-12-01T08:41:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:31 crc kubenswrapper[4813]: E1201 08:41:31.123176 4813 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:41:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:41:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:31Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:41:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:41:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:31Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cd271ba1-a5c5-44e0-b03e-cfd1f5e8cb76\\\",\\\"systemUUID\\\":\\\"906e3783-eb6d-4f21-889c-7989aeaa25ba\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:31Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:31 crc kubenswrapper[4813]: E1201 08:41:31.123416 4813 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 01 08:41:31 crc kubenswrapper[4813]: I1201 08:41:31.125533 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:31 crc kubenswrapper[4813]: I1201 08:41:31.125575 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:31 crc kubenswrapper[4813]: I1201 08:41:31.125591 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:31 crc kubenswrapper[4813]: I1201 08:41:31.125613 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:31 crc kubenswrapper[4813]: I1201 08:41:31.125631 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:31Z","lastTransitionTime":"2025-12-01T08:41:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:31 crc kubenswrapper[4813]: I1201 08:41:31.228682 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:31 crc kubenswrapper[4813]: I1201 08:41:31.228746 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:31 crc kubenswrapper[4813]: I1201 08:41:31.228766 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:31 crc kubenswrapper[4813]: I1201 08:41:31.228791 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:31 crc kubenswrapper[4813]: I1201 08:41:31.228808 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:31Z","lastTransitionTime":"2025-12-01T08:41:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:31 crc kubenswrapper[4813]: I1201 08:41:31.334754 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:31 crc kubenswrapper[4813]: I1201 08:41:31.334811 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:31 crc kubenswrapper[4813]: I1201 08:41:31.334828 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:31 crc kubenswrapper[4813]: I1201 08:41:31.334851 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:31 crc kubenswrapper[4813]: I1201 08:41:31.334868 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:31Z","lastTransitionTime":"2025-12-01T08:41:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:31 crc kubenswrapper[4813]: I1201 08:41:31.393132 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 08:41:31 crc kubenswrapper[4813]: I1201 08:41:31.393132 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-96mbr" Dec 01 08:41:31 crc kubenswrapper[4813]: E1201 08:41:31.393345 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 08:41:31 crc kubenswrapper[4813]: E1201 08:41:31.393512 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-96mbr" podUID="0789e3f7-5b08-4c04-8596-9ba29f50cdd0" Dec 01 08:41:31 crc kubenswrapper[4813]: I1201 08:41:31.438270 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:31 crc kubenswrapper[4813]: I1201 08:41:31.438324 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:31 crc kubenswrapper[4813]: I1201 08:41:31.438340 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:31 crc kubenswrapper[4813]: I1201 08:41:31.438364 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:31 crc kubenswrapper[4813]: I1201 08:41:31.438381 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:31Z","lastTransitionTime":"2025-12-01T08:41:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:31 crc kubenswrapper[4813]: I1201 08:41:31.542200 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:31 crc kubenswrapper[4813]: I1201 08:41:31.542263 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:31 crc kubenswrapper[4813]: I1201 08:41:31.542280 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:31 crc kubenswrapper[4813]: I1201 08:41:31.542304 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:31 crc kubenswrapper[4813]: I1201 08:41:31.542321 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:31Z","lastTransitionTime":"2025-12-01T08:41:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:31 crc kubenswrapper[4813]: I1201 08:41:31.604721 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-b6qqv_b9656a03-0096-4c89-b0da-e37e4103611b/kube-multus/0.log" Dec 01 08:41:31 crc kubenswrapper[4813]: I1201 08:41:31.604801 4813 generic.go:334] "Generic (PLEG): container finished" podID="b9656a03-0096-4c89-b0da-e37e4103611b" containerID="c4195aebc1b2e388d480920a7f72257f6660afc99ad8c1a09478ef445739b272" exitCode=1 Dec 01 08:41:31 crc kubenswrapper[4813]: I1201 08:41:31.604848 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-b6qqv" event={"ID":"b9656a03-0096-4c89-b0da-e37e4103611b","Type":"ContainerDied","Data":"c4195aebc1b2e388d480920a7f72257f6660afc99ad8c1a09478ef445739b272"} Dec 01 08:41:31 crc kubenswrapper[4813]: I1201 08:41:31.605466 4813 scope.go:117] "RemoveContainer" containerID="c4195aebc1b2e388d480920a7f72257f6660afc99ad8c1a09478ef445739b272" Dec 01 08:41:31 crc kubenswrapper[4813]: I1201 08:41:31.631270 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"349332bd-6b4d-4306-84ae-07019ca3860f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d441a348087ae3cadda1ad8a7d1c4dff02cf3f6e1d346c294b278fdd9131edcc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://da67635daf6d9e32937f4d5d4ceeee3f8f09ae4fd2dd26c3e85be1286c5949de\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e97356ee274ea78d9f237ffbc756abb6ac8df0dbac244bf0a86ed215b8cbf43\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8dc6feb9817f4f356b8e36af43ec3070774f312fbbb4909fa05b8b3e8dd53582\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a0404f5a5ed5a9750d4c479b608e68e8960b448f455d5af9d219aa5edaeb789b\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"message\\\":\\\"\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764578421\\\\\\\\\\\\\\\" (2025-12-01 08:40:21 +0000 UTC to 2025-12-31 08:40:22 +0000 UTC (now=2025-12-01 08:40:38.066068451 +0000 UTC))\\\\\\\"\\\\nI1201 08:40:38.066144 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1201 08:40:38.066166 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1201 08:40:38.066220 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1201 08:40:38.066232 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1201 08:40:38.066338 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764578432\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764578432\\\\\\\\\\\\\\\" (2025-12-01 07:40:32 +0000 UTC to 2026-12-01 07:40:32 +0000 UTC (now=2025-12-01 08:40:38.066301217 +0000 UTC))\\\\\\\"\\\\nI1201 08:40:38.066357 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1201 08:40:38.066385 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1201 08:40:38.066402 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1201 08:40:38.066414 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-63890811/tls.crt::/tmp/serving-cert-63890811/tls.key\\\\\\\"\\\\nI1201 08:40:38.066419 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1201 08:40:38.066572 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1201 08:40:38.067159 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7354f9337d42115bb1d974bf3772f2d03365acba57f2155135c7a0bbab39879e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:19Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05492df5703392cdf0932d6bf66025b18e53bd05fa0857a7cb2ad70979cbdc3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05492df5703392cdf0932d6bf66025b18e53bd05fa0857a7cb2ad70979cbdc3e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:16Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:31Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:31 crc kubenswrapper[4813]: I1201 08:41:31.645815 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:31 crc kubenswrapper[4813]: I1201 08:41:31.645869 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:31 crc kubenswrapper[4813]: I1201 08:41:31.645896 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:31 crc kubenswrapper[4813]: I1201 08:41:31.645928 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:31 crc kubenswrapper[4813]: I1201 08:41:31.645949 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:31Z","lastTransitionTime":"2025-12-01T08:41:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:31 crc kubenswrapper[4813]: I1201 08:41:31.658433 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3264b3af996af452df2953c59bea6deeb5ea6af1b119ce89f7334595c27e70da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9e879b1f563bd66383d58fd7b435a8a81bd50cb387dc1fabb6fc47d5a3fa468\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:31Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:31 crc kubenswrapper[4813]: I1201 08:41:31.680993 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-b6qqv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9656a03-0096-4c89-b0da-e37e4103611b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4195aebc1b2e388d480920a7f72257f6660afc99ad8c1a09478ef445739b272\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c4195aebc1b2e388d480920a7f72257f6660afc99ad8c1a09478ef445739b272\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T08:41:31Z\\\",\\\"message\\\":\\\"2025-12-01T08:40:45+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_038ba082-9223-4817-aa2f-727d3b2383fe\\\\n2025-12-01T08:40:45+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_038ba082-9223-4817-aa2f-727d3b2383fe to /host/opt/cni/bin/\\\\n2025-12-01T08:40:46Z [verbose] multus-daemon started\\\\n2025-12-01T08:40:46Z [verbose] Readiness Indicator file check\\\\n2025-12-01T08:41:31Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cjvnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:41Z\\\"}}\" for pod \"openshift-multus\"/\"multus-b6qqv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:31Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:31 crc kubenswrapper[4813]: I1201 08:41:31.695676 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-96mbr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0789e3f7-5b08-4c04-8596-9ba29f50cdd0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xt4fj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xt4fj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:55Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-96mbr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:31Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:31 crc kubenswrapper[4813]: I1201 08:41:31.709638 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-9wn4q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3d423402-e08f-42e9-b60c-6eb9bf2f6d48\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3049b842978c6580e1033a442a37766a2e76e278aface9a89fc0b2c57faf0df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rzxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c915733072b32a63237e2401cbf84b059aca396c2c0db884f0f439d599a35d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rzxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:41Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-9wn4q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:31Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:31 crc kubenswrapper[4813]: I1201 08:41:31.722358 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-8pg2p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dd53b38d-d9dd-4f98-9d06-dd95f1a4e3a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a214d71293e4e09573c665524d880fc319dfb1c0c6d95d11d78f253865660fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnjhq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ddf9d5989e4195b3074de08f99d53b48066f53fe93ce204145b3c714996fc57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnjhq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:53Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-8pg2p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:31Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:31 crc kubenswrapper[4813]: I1201 08:41:31.736001 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"490e526a-902b-4aff-b547-0cefea562764\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb48bd28fd3d29781695a5ae630f4913ef2745e97fbe7d463cca39a31e43ba0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74f8cf73f07b4cde260af429d26413e4e6226e7b177b493dd9c0e5c043ab6e1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3aaf696cefbe398bb6b26acfc5be43bda7e2be18218c8060840350c4106e66a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53de1f0682f36aa83ad8e5c226b95daf9f97df4ae154ebfc2fc287f9f238a4a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53de1f0682f36aa83ad8e5c226b95daf9f97df4ae154ebfc2fc287f9f238a4a8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:17Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:16Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:31Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:31 crc kubenswrapper[4813]: I1201 08:41:31.749397 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:31 crc kubenswrapper[4813]: I1201 08:41:31.749440 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:31 crc kubenswrapper[4813]: I1201 08:41:31.749452 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:31 crc kubenswrapper[4813]: I1201 08:41:31.749474 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:31 crc kubenswrapper[4813]: I1201 08:41:31.749486 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:31Z","lastTransitionTime":"2025-12-01T08:41:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:31 crc kubenswrapper[4813]: I1201 08:41:31.755383 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:31Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:31 crc kubenswrapper[4813]: I1201 08:41:31.774834 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-c576v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"17292cd0-1eac-40b5-a60e-67167c57409c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a42206c914d0493ca74788ea1a941de80920ab7b810fb95227cd9b8711e8f4b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7zwfx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:41Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-c576v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:31Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:31 crc kubenswrapper[4813]: I1201 08:41:31.801094 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-j4pcc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b7505e14-3560-4d0d-be2c-d6a7a5953865\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ccc30a63dc1c23ff9880fb04af1f31f0656b2696d359988e23deed23989f480a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://944977e9722025088720107530d0c91fc693593f380e4b49678f28a56da180cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc32d930ab23f30103d8fa42291909cbec89ebb75c5c53cd783cd27fe1a3f09e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f19c67b1c6c80c7f3223512102810d46ede1445464d21ae3bbea83424539b2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d984c90302e57f3e5a6f3ca3613b415db04ac74a5a228582f1b52ef1d59f93e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce8963bf1e9ee66926907aaf3a03d18977f4d0a1983d7121b62fbc8e02a37e4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb17fb9ae11fa62235328567418fb259f4f8540ebac635842d3afe751dcbde8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8790f4730881303e7743498b66fc17b4b0df5b6620c8ca640ad7b9c90e6898d2\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T08:40:56Z\\\",\\\"message\\\":\\\"I1201 08:40:56.782175 6307 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1201 08:40:56.782219 6307 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1201 08:40:56.782251 6307 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1201 08:40:56.782302 6307 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1201 08:40:56.782342 6307 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1201 08:40:56.782379 6307 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1201 08:40:56.782414 6307 factory.go:656] Stopping watch factory\\\\nI1201 08:40:56.782444 6307 ovnkube.go:599] Stopped ovnkube\\\\nI1201 08:40:56.782487 6307 handler.go:208] Removed *v1.Node event handler 7\\\\nI1201 08:40:56.782514 6307 handler.go:208] Removed *v1.Node event handler 2\\\\nI1201 08:40:56.782539 6307 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1201 08:40:56.782564 6307 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1201 08:40:56.782591 6307 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1201 08:40:56.782650 6307 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1201 08:40:56.782670 6307 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1201 08:40:56.782663 6307 handler.go:208] Removed *v1.EgressFirewa\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:55Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fb17fb9ae11fa62235328567418fb259f4f8540ebac635842d3afe751dcbde8c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T08:41:21Z\\\",\\\"message\\\":\\\"guring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-console-operator/metrics]} name:Service_openshift-console-operator/metrics_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.88:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {ebd4748e-0473-49fb-88ad-83dbb221791a}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1201 08:41:21.455187 6576 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1201 08:41:21.455499 6576 obj_retry.go:303] Retry object setup: *v1.Pod openshift-image-registry/node-ca-xpnlc\\\\nF1201 08:41:21.455509 6576 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal erro\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T08:41:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b45454b440c5108fa150a6b856afa1778ba6e9f89633433cb3117b68f00f72c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd67634e0c806323334495401a4e20b48fea4bc5013acd973d2abd3d70870f99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd67634e0c806323334495401a4e20b48fea4bc5013acd973d2abd3d70870f99\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:41Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-j4pcc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:31Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:31 crc kubenswrapper[4813]: I1201 08:41:31.827442 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-h7lvd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d431b613-a91b-4fff-88b9-447804766289\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a32c51c943fcf6ada7e3f614667eb023e29d94c11128d26c1fecca37b03963b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://495593f99e92a9c28100a95e27d06ff5847008c5d996366cb474bedd9b69d5ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://495593f99e92a9c28100a95e27d06ff5847008c5d996366cb474bedd9b69d5ef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://19358f83a6c52957f959fe935a6843fdeef59905cc395e6e52a1fb4a670ac193\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://19358f83a6c52957f959fe935a6843fdeef59905cc395e6e52a1fb4a670ac193\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f089c115d916e339994879d224e3db27642b99aaabf59bcafb2979549052945\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9f089c115d916e339994879d224e3db27642b99aaabf59bcafb2979549052945\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb66f224462305c2005b31ef7bb9ed01efde8101e18518dd42d1dfb0df765bb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb66f224462305c2005b31ef7bb9ed01efde8101e18518dd42d1dfb0df765bb9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c268236c9a155c9b1f47c48d30dbd23948ceb2545592b40f356c60ea1d12e8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c268236c9a155c9b1f47c48d30dbd23948ceb2545592b40f356c60ea1d12e8a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://458aedca5f186e9aaed7e4ee29fd5247221491888b130f55c1fe48d88d565f8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://458aedca5f186e9aaed7e4ee29fd5247221491888b130f55c1fe48d88d565f8f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:41Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-h7lvd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:31Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:31 crc kubenswrapper[4813]: I1201 08:41:31.846282 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6fb5a648-ae4f-4e46-a530-007e9bf84abb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e24ebfc52dcb28368a2d436e7d9ab78f07f2aec2a4c71405a6c7d3301eee96a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33898118bbf639c8e6eae5aafa10386b95795512444ed5f535db3c663f169df2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9902c00bf7738686c93f62e1772a484d05715d74b75ee98c84c8f23ed383824b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3657b47ec9f1563c83bfbe3d728c8fd123f7078e219d251b1874269e42c31eea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:16Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:31Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:31 crc kubenswrapper[4813]: I1201 08:41:31.852449 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:31 crc kubenswrapper[4813]: I1201 08:41:31.852485 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:31 crc kubenswrapper[4813]: I1201 08:41:31.852496 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:31 crc kubenswrapper[4813]: I1201 08:41:31.852513 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:31 crc kubenswrapper[4813]: I1201 08:41:31.852524 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:31Z","lastTransitionTime":"2025-12-01T08:41:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:31 crc kubenswrapper[4813]: I1201 08:41:31.866025 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0023b0e358d75aacc889ce364846dda89ea513832c87fa4b066aea7dac57e8a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:31Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:31 crc kubenswrapper[4813]: I1201 08:41:31.887361 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:31Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:31 crc kubenswrapper[4813]: I1201 08:41:31.910329 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c7574f892d5e9142296b26a0b484eb6759ab53f571139fb6e23b93dc3da05780\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:31Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:31 crc kubenswrapper[4813]: I1201 08:41:31.926102 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:31Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:31 crc kubenswrapper[4813]: I1201 08:41:31.939581 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-xpnlc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"22da1fcf-0c57-4708-90d8-77692af17344\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4a36d213bdad2da91d0d33f400d5a8248befb68d5b94d19347873a93c6712a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xc44d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-xpnlc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:31Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:31 crc kubenswrapper[4813]: I1201 08:41:31.956164 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:31 crc kubenswrapper[4813]: I1201 08:41:31.956201 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:31 crc kubenswrapper[4813]: I1201 08:41:31.956213 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:31 crc kubenswrapper[4813]: I1201 08:41:31.956229 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:31 crc kubenswrapper[4813]: I1201 08:41:31.956241 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:31Z","lastTransitionTime":"2025-12-01T08:41:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:32 crc kubenswrapper[4813]: I1201 08:41:32.059699 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:32 crc kubenswrapper[4813]: I1201 08:41:32.059749 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:32 crc kubenswrapper[4813]: I1201 08:41:32.059765 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:32 crc kubenswrapper[4813]: I1201 08:41:32.059835 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:32 crc kubenswrapper[4813]: I1201 08:41:32.059854 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:32Z","lastTransitionTime":"2025-12-01T08:41:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:32 crc kubenswrapper[4813]: I1201 08:41:32.163773 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:32 crc kubenswrapper[4813]: I1201 08:41:32.163839 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:32 crc kubenswrapper[4813]: I1201 08:41:32.163864 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:32 crc kubenswrapper[4813]: I1201 08:41:32.163894 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:32 crc kubenswrapper[4813]: I1201 08:41:32.163918 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:32Z","lastTransitionTime":"2025-12-01T08:41:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:32 crc kubenswrapper[4813]: I1201 08:41:32.267293 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:32 crc kubenswrapper[4813]: I1201 08:41:32.267346 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:32 crc kubenswrapper[4813]: I1201 08:41:32.267369 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:32 crc kubenswrapper[4813]: I1201 08:41:32.267396 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:32 crc kubenswrapper[4813]: I1201 08:41:32.267412 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:32Z","lastTransitionTime":"2025-12-01T08:41:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:32 crc kubenswrapper[4813]: I1201 08:41:32.372062 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:32 crc kubenswrapper[4813]: I1201 08:41:32.372126 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:32 crc kubenswrapper[4813]: I1201 08:41:32.372197 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:32 crc kubenswrapper[4813]: I1201 08:41:32.372225 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:32 crc kubenswrapper[4813]: I1201 08:41:32.372284 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:32Z","lastTransitionTime":"2025-12-01T08:41:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:32 crc kubenswrapper[4813]: I1201 08:41:32.393734 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 08:41:32 crc kubenswrapper[4813]: E1201 08:41:32.393902 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 08:41:32 crc kubenswrapper[4813]: I1201 08:41:32.393735 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 08:41:32 crc kubenswrapper[4813]: E1201 08:41:32.394162 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 08:41:32 crc kubenswrapper[4813]: I1201 08:41:32.475946 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:32 crc kubenswrapper[4813]: I1201 08:41:32.476050 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:32 crc kubenswrapper[4813]: I1201 08:41:32.476077 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:32 crc kubenswrapper[4813]: I1201 08:41:32.476110 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:32 crc kubenswrapper[4813]: I1201 08:41:32.476132 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:32Z","lastTransitionTime":"2025-12-01T08:41:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:32 crc kubenswrapper[4813]: I1201 08:41:32.579958 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:32 crc kubenswrapper[4813]: I1201 08:41:32.580067 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:32 crc kubenswrapper[4813]: I1201 08:41:32.580090 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:32 crc kubenswrapper[4813]: I1201 08:41:32.580119 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:32 crc kubenswrapper[4813]: I1201 08:41:32.580141 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:32Z","lastTransitionTime":"2025-12-01T08:41:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:32 crc kubenswrapper[4813]: I1201 08:41:32.612110 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-b6qqv_b9656a03-0096-4c89-b0da-e37e4103611b/kube-multus/0.log" Dec 01 08:41:32 crc kubenswrapper[4813]: I1201 08:41:32.612172 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-b6qqv" event={"ID":"b9656a03-0096-4c89-b0da-e37e4103611b","Type":"ContainerStarted","Data":"2b730cdf02e6456d9a1d02d66b94064ca71052eb8f373eaabfe327369eb03a15"} Dec 01 08:41:32 crc kubenswrapper[4813]: I1201 08:41:32.638290 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-h7lvd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d431b613-a91b-4fff-88b9-447804766289\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a32c51c943fcf6ada7e3f614667eb023e29d94c11128d26c1fecca37b03963b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://495593f99e92a9c28100a95e27d06ff5847008c5d996366cb474bedd9b69d5ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://495593f99e92a9c28100a95e27d06ff5847008c5d996366cb474bedd9b69d5ef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://19358f83a6c52957f959fe935a6843fdeef59905cc395e6e52a1fb4a670ac193\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://19358f83a6c52957f959fe935a6843fdeef59905cc395e6e52a1fb4a670ac193\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f089c115d916e339994879d224e3db27642b99aaabf59bcafb2979549052945\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9f089c115d916e339994879d224e3db27642b99aaabf59bcafb2979549052945\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb66f224462305c2005b31ef7bb9ed01efde8101e18518dd42d1dfb0df765bb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb66f224462305c2005b31ef7bb9ed01efde8101e18518dd42d1dfb0df765bb9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c268236c9a155c9b1f47c48d30dbd23948ceb2545592b40f356c60ea1d12e8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c268236c9a155c9b1f47c48d30dbd23948ceb2545592b40f356c60ea1d12e8a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://458aedca5f186e9aaed7e4ee29fd5247221491888b130f55c1fe48d88d565f8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://458aedca5f186e9aaed7e4ee29fd5247221491888b130f55c1fe48d88d565f8f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:41Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-h7lvd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:32Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:32 crc kubenswrapper[4813]: I1201 08:41:32.655758 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"490e526a-902b-4aff-b547-0cefea562764\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb48bd28fd3d29781695a5ae630f4913ef2745e97fbe7d463cca39a31e43ba0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74f8cf73f07b4cde260af429d26413e4e6226e7b177b493dd9c0e5c043ab6e1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3aaf696cefbe398bb6b26acfc5be43bda7e2be18218c8060840350c4106e66a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53de1f0682f36aa83ad8e5c226b95daf9f97df4ae154ebfc2fc287f9f238a4a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53de1f0682f36aa83ad8e5c226b95daf9f97df4ae154ebfc2fc287f9f238a4a8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:17Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:16Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:32Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:32 crc kubenswrapper[4813]: I1201 08:41:32.673883 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:32Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:32 crc kubenswrapper[4813]: I1201 08:41:32.682726 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:32 crc kubenswrapper[4813]: I1201 08:41:32.682766 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:32 crc kubenswrapper[4813]: I1201 08:41:32.682774 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:32 crc kubenswrapper[4813]: I1201 08:41:32.682790 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:32 crc kubenswrapper[4813]: I1201 08:41:32.682799 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:32Z","lastTransitionTime":"2025-12-01T08:41:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:32 crc kubenswrapper[4813]: I1201 08:41:32.688688 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-c576v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"17292cd0-1eac-40b5-a60e-67167c57409c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a42206c914d0493ca74788ea1a941de80920ab7b810fb95227cd9b8711e8f4b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7zwfx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:41Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-c576v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:32Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:32 crc kubenswrapper[4813]: I1201 08:41:32.712459 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-j4pcc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b7505e14-3560-4d0d-be2c-d6a7a5953865\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ccc30a63dc1c23ff9880fb04af1f31f0656b2696d359988e23deed23989f480a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://944977e9722025088720107530d0c91fc693593f380e4b49678f28a56da180cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc32d930ab23f30103d8fa42291909cbec89ebb75c5c53cd783cd27fe1a3f09e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f19c67b1c6c80c7f3223512102810d46ede1445464d21ae3bbea83424539b2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d984c90302e57f3e5a6f3ca3613b415db04ac74a5a228582f1b52ef1d59f93e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce8963bf1e9ee66926907aaf3a03d18977f4d0a1983d7121b62fbc8e02a37e4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb17fb9ae11fa62235328567418fb259f4f8540ebac635842d3afe751dcbde8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8790f4730881303e7743498b66fc17b4b0df5b6620c8ca640ad7b9c90e6898d2\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T08:40:56Z\\\",\\\"message\\\":\\\"I1201 08:40:56.782175 6307 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1201 08:40:56.782219 6307 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1201 08:40:56.782251 6307 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1201 08:40:56.782302 6307 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1201 08:40:56.782342 6307 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1201 08:40:56.782379 6307 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1201 08:40:56.782414 6307 factory.go:656] Stopping watch factory\\\\nI1201 08:40:56.782444 6307 ovnkube.go:599] Stopped ovnkube\\\\nI1201 08:40:56.782487 6307 handler.go:208] Removed *v1.Node event handler 7\\\\nI1201 08:40:56.782514 6307 handler.go:208] Removed *v1.Node event handler 2\\\\nI1201 08:40:56.782539 6307 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1201 08:40:56.782564 6307 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1201 08:40:56.782591 6307 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1201 08:40:56.782650 6307 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1201 08:40:56.782670 6307 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1201 08:40:56.782663 6307 handler.go:208] Removed *v1.EgressFirewa\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:55Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fb17fb9ae11fa62235328567418fb259f4f8540ebac635842d3afe751dcbde8c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T08:41:21Z\\\",\\\"message\\\":\\\"guring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-console-operator/metrics]} name:Service_openshift-console-operator/metrics_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.88:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {ebd4748e-0473-49fb-88ad-83dbb221791a}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1201 08:41:21.455187 6576 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1201 08:41:21.455499 6576 obj_retry.go:303] Retry object setup: *v1.Pod openshift-image-registry/node-ca-xpnlc\\\\nF1201 08:41:21.455509 6576 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal erro\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T08:41:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b45454b440c5108fa150a6b856afa1778ba6e9f89633433cb3117b68f00f72c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd67634e0c806323334495401a4e20b48fea4bc5013acd973d2abd3d70870f99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd67634e0c806323334495401a4e20b48fea4bc5013acd973d2abd3d70870f99\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:41Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-j4pcc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:32Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:32 crc kubenswrapper[4813]: I1201 08:41:32.757358 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:32Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:32 crc kubenswrapper[4813]: I1201 08:41:32.782007 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-xpnlc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"22da1fcf-0c57-4708-90d8-77692af17344\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4a36d213bdad2da91d0d33f400d5a8248befb68d5b94d19347873a93c6712a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xc44d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-xpnlc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:32Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:32 crc kubenswrapper[4813]: I1201 08:41:32.784955 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:32 crc kubenswrapper[4813]: I1201 08:41:32.785021 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:32 crc kubenswrapper[4813]: I1201 08:41:32.785034 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:32 crc kubenswrapper[4813]: I1201 08:41:32.785055 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:32 crc kubenswrapper[4813]: I1201 08:41:32.785069 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:32Z","lastTransitionTime":"2025-12-01T08:41:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:32 crc kubenswrapper[4813]: I1201 08:41:32.800391 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6fb5a648-ae4f-4e46-a530-007e9bf84abb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e24ebfc52dcb28368a2d436e7d9ab78f07f2aec2a4c71405a6c7d3301eee96a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33898118bbf639c8e6eae5aafa10386b95795512444ed5f535db3c663f169df2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9902c00bf7738686c93f62e1772a484d05715d74b75ee98c84c8f23ed383824b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3657b47ec9f1563c83bfbe3d728c8fd123f7078e219d251b1874269e42c31eea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:16Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:32Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:32 crc kubenswrapper[4813]: I1201 08:41:32.817574 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0023b0e358d75aacc889ce364846dda89ea513832c87fa4b066aea7dac57e8a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:32Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:32 crc kubenswrapper[4813]: I1201 08:41:32.835367 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:32Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:32 crc kubenswrapper[4813]: I1201 08:41:32.848300 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c7574f892d5e9142296b26a0b484eb6759ab53f571139fb6e23b93dc3da05780\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:32Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:32 crc kubenswrapper[4813]: I1201 08:41:32.859377 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-96mbr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0789e3f7-5b08-4c04-8596-9ba29f50cdd0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xt4fj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xt4fj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:55Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-96mbr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:32Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:32 crc kubenswrapper[4813]: I1201 08:41:32.872593 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"349332bd-6b4d-4306-84ae-07019ca3860f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d441a348087ae3cadda1ad8a7d1c4dff02cf3f6e1d346c294b278fdd9131edcc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://da67635daf6d9e32937f4d5d4ceeee3f8f09ae4fd2dd26c3e85be1286c5949de\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e97356ee274ea78d9f237ffbc756abb6ac8df0dbac244bf0a86ed215b8cbf43\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8dc6feb9817f4f356b8e36af43ec3070774f312fbbb4909fa05b8b3e8dd53582\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a0404f5a5ed5a9750d4c479b608e68e8960b448f455d5af9d219aa5edaeb789b\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"message\\\":\\\"\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764578421\\\\\\\\\\\\\\\" (2025-12-01 08:40:21 +0000 UTC to 2025-12-31 08:40:22 +0000 UTC (now=2025-12-01 08:40:38.066068451 +0000 UTC))\\\\\\\"\\\\nI1201 08:40:38.066144 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1201 08:40:38.066166 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1201 08:40:38.066220 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1201 08:40:38.066232 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1201 08:40:38.066338 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764578432\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764578432\\\\\\\\\\\\\\\" (2025-12-01 07:40:32 +0000 UTC to 2026-12-01 07:40:32 +0000 UTC (now=2025-12-01 08:40:38.066301217 +0000 UTC))\\\\\\\"\\\\nI1201 08:40:38.066357 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1201 08:40:38.066385 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1201 08:40:38.066402 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1201 08:40:38.066414 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-63890811/tls.crt::/tmp/serving-cert-63890811/tls.key\\\\\\\"\\\\nI1201 08:40:38.066419 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1201 08:40:38.066572 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1201 08:40:38.067159 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7354f9337d42115bb1d974bf3772f2d03365acba57f2155135c7a0bbab39879e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:19Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05492df5703392cdf0932d6bf66025b18e53bd05fa0857a7cb2ad70979cbdc3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05492df5703392cdf0932d6bf66025b18e53bd05fa0857a7cb2ad70979cbdc3e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:16Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:32Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:32 crc kubenswrapper[4813]: I1201 08:41:32.883221 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3264b3af996af452df2953c59bea6deeb5ea6af1b119ce89f7334595c27e70da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9e879b1f563bd66383d58fd7b435a8a81bd50cb387dc1fabb6fc47d5a3fa468\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:32Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:32 crc kubenswrapper[4813]: I1201 08:41:32.890356 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:32 crc kubenswrapper[4813]: I1201 08:41:32.890433 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:32 crc kubenswrapper[4813]: I1201 08:41:32.890486 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:32 crc kubenswrapper[4813]: I1201 08:41:32.890573 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:32 crc kubenswrapper[4813]: I1201 08:41:32.890607 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:32Z","lastTransitionTime":"2025-12-01T08:41:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:32 crc kubenswrapper[4813]: I1201 08:41:32.896997 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-b6qqv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9656a03-0096-4c89-b0da-e37e4103611b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b730cdf02e6456d9a1d02d66b94064ca71052eb8f373eaabfe327369eb03a15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c4195aebc1b2e388d480920a7f72257f6660afc99ad8c1a09478ef445739b272\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T08:41:31Z\\\",\\\"message\\\":\\\"2025-12-01T08:40:45+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_038ba082-9223-4817-aa2f-727d3b2383fe\\\\n2025-12-01T08:40:45+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_038ba082-9223-4817-aa2f-727d3b2383fe to /host/opt/cni/bin/\\\\n2025-12-01T08:40:46Z [verbose] multus-daemon started\\\\n2025-12-01T08:40:46Z [verbose] Readiness Indicator file check\\\\n2025-12-01T08:41:31Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:41:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cjvnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:41Z\\\"}}\" for pod \"openshift-multus\"/\"multus-b6qqv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:32Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:32 crc kubenswrapper[4813]: I1201 08:41:32.908754 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-9wn4q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3d423402-e08f-42e9-b60c-6eb9bf2f6d48\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3049b842978c6580e1033a442a37766a2e76e278aface9a89fc0b2c57faf0df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rzxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c915733072b32a63237e2401cbf84b059aca396c2c0db884f0f439d599a35d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rzxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:41Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-9wn4q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:32Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:32 crc kubenswrapper[4813]: I1201 08:41:32.918617 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-8pg2p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dd53b38d-d9dd-4f98-9d06-dd95f1a4e3a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a214d71293e4e09573c665524d880fc319dfb1c0c6d95d11d78f253865660fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnjhq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ddf9d5989e4195b3074de08f99d53b48066f53fe93ce204145b3c714996fc57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnjhq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:53Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-8pg2p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:32Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:32 crc kubenswrapper[4813]: I1201 08:41:32.993803 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:32 crc kubenswrapper[4813]: I1201 08:41:32.993842 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:32 crc kubenswrapper[4813]: I1201 08:41:32.993853 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:32 crc kubenswrapper[4813]: I1201 08:41:32.993870 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:32 crc kubenswrapper[4813]: I1201 08:41:32.993882 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:32Z","lastTransitionTime":"2025-12-01T08:41:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:33 crc kubenswrapper[4813]: I1201 08:41:33.098862 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:33 crc kubenswrapper[4813]: I1201 08:41:33.098932 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:33 crc kubenswrapper[4813]: I1201 08:41:33.098955 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:33 crc kubenswrapper[4813]: I1201 08:41:33.099024 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:33 crc kubenswrapper[4813]: I1201 08:41:33.099049 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:33Z","lastTransitionTime":"2025-12-01T08:41:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:33 crc kubenswrapper[4813]: I1201 08:41:33.201784 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:33 crc kubenswrapper[4813]: I1201 08:41:33.201832 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:33 crc kubenswrapper[4813]: I1201 08:41:33.201845 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:33 crc kubenswrapper[4813]: I1201 08:41:33.201861 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:33 crc kubenswrapper[4813]: I1201 08:41:33.201872 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:33Z","lastTransitionTime":"2025-12-01T08:41:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:33 crc kubenswrapper[4813]: I1201 08:41:33.304626 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:33 crc kubenswrapper[4813]: I1201 08:41:33.304750 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:33 crc kubenswrapper[4813]: I1201 08:41:33.304771 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:33 crc kubenswrapper[4813]: I1201 08:41:33.304797 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:33 crc kubenswrapper[4813]: I1201 08:41:33.304815 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:33Z","lastTransitionTime":"2025-12-01T08:41:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:33 crc kubenswrapper[4813]: I1201 08:41:33.392766 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-96mbr" Dec 01 08:41:33 crc kubenswrapper[4813]: I1201 08:41:33.392788 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 08:41:33 crc kubenswrapper[4813]: E1201 08:41:33.393083 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-96mbr" podUID="0789e3f7-5b08-4c04-8596-9ba29f50cdd0" Dec 01 08:41:33 crc kubenswrapper[4813]: E1201 08:41:33.393206 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 08:41:33 crc kubenswrapper[4813]: I1201 08:41:33.408603 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:33 crc kubenswrapper[4813]: I1201 08:41:33.408666 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:33 crc kubenswrapper[4813]: I1201 08:41:33.408682 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:33 crc kubenswrapper[4813]: I1201 08:41:33.408707 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:33 crc kubenswrapper[4813]: I1201 08:41:33.408725 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:33Z","lastTransitionTime":"2025-12-01T08:41:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:33 crc kubenswrapper[4813]: I1201 08:41:33.513064 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:33 crc kubenswrapper[4813]: I1201 08:41:33.513124 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:33 crc kubenswrapper[4813]: I1201 08:41:33.513146 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:33 crc kubenswrapper[4813]: I1201 08:41:33.513176 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:33 crc kubenswrapper[4813]: I1201 08:41:33.513199 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:33Z","lastTransitionTime":"2025-12-01T08:41:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:33 crc kubenswrapper[4813]: I1201 08:41:33.616332 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:33 crc kubenswrapper[4813]: I1201 08:41:33.616406 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:33 crc kubenswrapper[4813]: I1201 08:41:33.616424 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:33 crc kubenswrapper[4813]: I1201 08:41:33.616451 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:33 crc kubenswrapper[4813]: I1201 08:41:33.616471 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:33Z","lastTransitionTime":"2025-12-01T08:41:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:33 crc kubenswrapper[4813]: I1201 08:41:33.720137 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:33 crc kubenswrapper[4813]: I1201 08:41:33.720247 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:33 crc kubenswrapper[4813]: I1201 08:41:33.720271 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:33 crc kubenswrapper[4813]: I1201 08:41:33.720302 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:33 crc kubenswrapper[4813]: I1201 08:41:33.720389 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:33Z","lastTransitionTime":"2025-12-01T08:41:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:33 crc kubenswrapper[4813]: I1201 08:41:33.823625 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:33 crc kubenswrapper[4813]: I1201 08:41:33.823669 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:33 crc kubenswrapper[4813]: I1201 08:41:33.823682 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:33 crc kubenswrapper[4813]: I1201 08:41:33.823702 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:33 crc kubenswrapper[4813]: I1201 08:41:33.823712 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:33Z","lastTransitionTime":"2025-12-01T08:41:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:33 crc kubenswrapper[4813]: I1201 08:41:33.926665 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:33 crc kubenswrapper[4813]: I1201 08:41:33.926714 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:33 crc kubenswrapper[4813]: I1201 08:41:33.926725 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:33 crc kubenswrapper[4813]: I1201 08:41:33.926743 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:33 crc kubenswrapper[4813]: I1201 08:41:33.926756 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:33Z","lastTransitionTime":"2025-12-01T08:41:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:34 crc kubenswrapper[4813]: I1201 08:41:34.029845 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:34 crc kubenswrapper[4813]: I1201 08:41:34.029916 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:34 crc kubenswrapper[4813]: I1201 08:41:34.029941 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:34 crc kubenswrapper[4813]: I1201 08:41:34.030016 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:34 crc kubenswrapper[4813]: I1201 08:41:34.030045 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:34Z","lastTransitionTime":"2025-12-01T08:41:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:34 crc kubenswrapper[4813]: I1201 08:41:34.132204 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:34 crc kubenswrapper[4813]: I1201 08:41:34.132238 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:34 crc kubenswrapper[4813]: I1201 08:41:34.132245 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:34 crc kubenswrapper[4813]: I1201 08:41:34.132260 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:34 crc kubenswrapper[4813]: I1201 08:41:34.132269 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:34Z","lastTransitionTime":"2025-12-01T08:41:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:34 crc kubenswrapper[4813]: I1201 08:41:34.236021 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:34 crc kubenswrapper[4813]: I1201 08:41:34.236066 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:34 crc kubenswrapper[4813]: I1201 08:41:34.236078 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:34 crc kubenswrapper[4813]: I1201 08:41:34.236096 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:34 crc kubenswrapper[4813]: I1201 08:41:34.236111 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:34Z","lastTransitionTime":"2025-12-01T08:41:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:34 crc kubenswrapper[4813]: I1201 08:41:34.340146 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:34 crc kubenswrapper[4813]: I1201 08:41:34.340218 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:34 crc kubenswrapper[4813]: I1201 08:41:34.340236 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:34 crc kubenswrapper[4813]: I1201 08:41:34.340262 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:34 crc kubenswrapper[4813]: I1201 08:41:34.340279 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:34Z","lastTransitionTime":"2025-12-01T08:41:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:34 crc kubenswrapper[4813]: I1201 08:41:34.394231 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 08:41:34 crc kubenswrapper[4813]: E1201 08:41:34.394832 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 08:41:34 crc kubenswrapper[4813]: I1201 08:41:34.394640 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 08:41:34 crc kubenswrapper[4813]: E1201 08:41:34.395425 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 08:41:34 crc kubenswrapper[4813]: I1201 08:41:34.422523 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Dec 01 08:41:34 crc kubenswrapper[4813]: I1201 08:41:34.443703 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:34 crc kubenswrapper[4813]: I1201 08:41:34.443773 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:34 crc kubenswrapper[4813]: I1201 08:41:34.443799 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:34 crc kubenswrapper[4813]: I1201 08:41:34.443868 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:34 crc kubenswrapper[4813]: I1201 08:41:34.443894 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:34Z","lastTransitionTime":"2025-12-01T08:41:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:34 crc kubenswrapper[4813]: I1201 08:41:34.547093 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:34 crc kubenswrapper[4813]: I1201 08:41:34.547173 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:34 crc kubenswrapper[4813]: I1201 08:41:34.547196 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:34 crc kubenswrapper[4813]: I1201 08:41:34.547227 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:34 crc kubenswrapper[4813]: I1201 08:41:34.547254 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:34Z","lastTransitionTime":"2025-12-01T08:41:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:34 crc kubenswrapper[4813]: I1201 08:41:34.649652 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:34 crc kubenswrapper[4813]: I1201 08:41:34.649745 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:34 crc kubenswrapper[4813]: I1201 08:41:34.649769 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:34 crc kubenswrapper[4813]: I1201 08:41:34.649799 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:34 crc kubenswrapper[4813]: I1201 08:41:34.649822 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:34Z","lastTransitionTime":"2025-12-01T08:41:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:34 crc kubenswrapper[4813]: I1201 08:41:34.753162 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:34 crc kubenswrapper[4813]: I1201 08:41:34.753491 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:34 crc kubenswrapper[4813]: I1201 08:41:34.753672 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:34 crc kubenswrapper[4813]: I1201 08:41:34.753814 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:34 crc kubenswrapper[4813]: I1201 08:41:34.753959 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:34Z","lastTransitionTime":"2025-12-01T08:41:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:34 crc kubenswrapper[4813]: I1201 08:41:34.856388 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:34 crc kubenswrapper[4813]: I1201 08:41:34.856426 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:34 crc kubenswrapper[4813]: I1201 08:41:34.856438 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:34 crc kubenswrapper[4813]: I1201 08:41:34.856452 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:34 crc kubenswrapper[4813]: I1201 08:41:34.856463 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:34Z","lastTransitionTime":"2025-12-01T08:41:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:34 crc kubenswrapper[4813]: I1201 08:41:34.959788 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:34 crc kubenswrapper[4813]: I1201 08:41:34.960122 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:34 crc kubenswrapper[4813]: I1201 08:41:34.960321 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:34 crc kubenswrapper[4813]: I1201 08:41:34.960974 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:34 crc kubenswrapper[4813]: I1201 08:41:34.961210 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:34Z","lastTransitionTime":"2025-12-01T08:41:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:35 crc kubenswrapper[4813]: I1201 08:41:35.064468 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:35 crc kubenswrapper[4813]: I1201 08:41:35.064543 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:35 crc kubenswrapper[4813]: I1201 08:41:35.064568 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:35 crc kubenswrapper[4813]: I1201 08:41:35.064603 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:35 crc kubenswrapper[4813]: I1201 08:41:35.064624 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:35Z","lastTransitionTime":"2025-12-01T08:41:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:35 crc kubenswrapper[4813]: I1201 08:41:35.167484 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:35 crc kubenswrapper[4813]: I1201 08:41:35.167528 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:35 crc kubenswrapper[4813]: I1201 08:41:35.167544 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:35 crc kubenswrapper[4813]: I1201 08:41:35.167566 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:35 crc kubenswrapper[4813]: I1201 08:41:35.167584 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:35Z","lastTransitionTime":"2025-12-01T08:41:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:35 crc kubenswrapper[4813]: I1201 08:41:35.271437 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:35 crc kubenswrapper[4813]: I1201 08:41:35.271518 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:35 crc kubenswrapper[4813]: I1201 08:41:35.271541 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:35 crc kubenswrapper[4813]: I1201 08:41:35.271572 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:35 crc kubenswrapper[4813]: I1201 08:41:35.271596 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:35Z","lastTransitionTime":"2025-12-01T08:41:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:35 crc kubenswrapper[4813]: I1201 08:41:35.375636 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:35 crc kubenswrapper[4813]: I1201 08:41:35.375719 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:35 crc kubenswrapper[4813]: I1201 08:41:35.375745 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:35 crc kubenswrapper[4813]: I1201 08:41:35.375776 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:35 crc kubenswrapper[4813]: I1201 08:41:35.375800 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:35Z","lastTransitionTime":"2025-12-01T08:41:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:35 crc kubenswrapper[4813]: I1201 08:41:35.393533 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-96mbr" Dec 01 08:41:35 crc kubenswrapper[4813]: I1201 08:41:35.393548 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 08:41:35 crc kubenswrapper[4813]: E1201 08:41:35.394154 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-96mbr" podUID="0789e3f7-5b08-4c04-8596-9ba29f50cdd0" Dec 01 08:41:35 crc kubenswrapper[4813]: E1201 08:41:35.394400 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 08:41:35 crc kubenswrapper[4813]: I1201 08:41:35.394603 4813 scope.go:117] "RemoveContainer" containerID="fb17fb9ae11fa62235328567418fb259f4f8540ebac635842d3afe751dcbde8c" Dec 01 08:41:35 crc kubenswrapper[4813]: E1201 08:41:35.395030 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-j4pcc_openshift-ovn-kubernetes(b7505e14-3560-4d0d-be2c-d6a7a5953865)\"" pod="openshift-ovn-kubernetes/ovnkube-node-j4pcc" podUID="b7505e14-3560-4d0d-be2c-d6a7a5953865" Dec 01 08:41:35 crc kubenswrapper[4813]: I1201 08:41:35.417154 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:35Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:35 crc kubenswrapper[4813]: I1201 08:41:35.443897 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c7574f892d5e9142296b26a0b484eb6759ab53f571139fb6e23b93dc3da05780\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:35Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:35 crc kubenswrapper[4813]: I1201 08:41:35.465929 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:35Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:35 crc kubenswrapper[4813]: I1201 08:41:35.479643 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:35 crc kubenswrapper[4813]: I1201 08:41:35.479700 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:35 crc kubenswrapper[4813]: I1201 08:41:35.479722 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:35 crc kubenswrapper[4813]: I1201 08:41:35.479753 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:35 crc kubenswrapper[4813]: I1201 08:41:35.479775 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:35Z","lastTransitionTime":"2025-12-01T08:41:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:35 crc kubenswrapper[4813]: I1201 08:41:35.480414 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-xpnlc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"22da1fcf-0c57-4708-90d8-77692af17344\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4a36d213bdad2da91d0d33f400d5a8248befb68d5b94d19347873a93c6712a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xc44d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-xpnlc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:35Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:35 crc kubenswrapper[4813]: I1201 08:41:35.499847 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6fb5a648-ae4f-4e46-a530-007e9bf84abb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e24ebfc52dcb28368a2d436e7d9ab78f07f2aec2a4c71405a6c7d3301eee96a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33898118bbf639c8e6eae5aafa10386b95795512444ed5f535db3c663f169df2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9902c00bf7738686c93f62e1772a484d05715d74b75ee98c84c8f23ed383824b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3657b47ec9f1563c83bfbe3d728c8fd123f7078e219d251b1874269e42c31eea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:16Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:35Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:35 crc kubenswrapper[4813]: I1201 08:41:35.516367 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0023b0e358d75aacc889ce364846dda89ea513832c87fa4b066aea7dac57e8a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:35Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:35 crc kubenswrapper[4813]: I1201 08:41:35.529493 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3264b3af996af452df2953c59bea6deeb5ea6af1b119ce89f7334595c27e70da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9e879b1f563bd66383d58fd7b435a8a81bd50cb387dc1fabb6fc47d5a3fa468\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:35Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:35 crc kubenswrapper[4813]: I1201 08:41:35.549633 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-b6qqv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9656a03-0096-4c89-b0da-e37e4103611b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b730cdf02e6456d9a1d02d66b94064ca71052eb8f373eaabfe327369eb03a15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c4195aebc1b2e388d480920a7f72257f6660afc99ad8c1a09478ef445739b272\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T08:41:31Z\\\",\\\"message\\\":\\\"2025-12-01T08:40:45+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_038ba082-9223-4817-aa2f-727d3b2383fe\\\\n2025-12-01T08:40:45+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_038ba082-9223-4817-aa2f-727d3b2383fe to /host/opt/cni/bin/\\\\n2025-12-01T08:40:46Z [verbose] multus-daemon started\\\\n2025-12-01T08:40:46Z [verbose] Readiness Indicator file check\\\\n2025-12-01T08:41:31Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:41:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cjvnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:41Z\\\"}}\" for pod \"openshift-multus\"/\"multus-b6qqv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:35Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:35 crc kubenswrapper[4813]: I1201 08:41:35.564810 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-96mbr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0789e3f7-5b08-4c04-8596-9ba29f50cdd0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xt4fj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xt4fj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:55Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-96mbr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:35Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:35 crc kubenswrapper[4813]: I1201 08:41:35.575952 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a6a31593-c1c1-4aab-9e99-f683e908c07f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://86abafcbe7fed4bfbaa6c3b2f64d6a314afa7ed0fe2709bc8a01ba9fb17c1a75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://62135a7a7fa690060354d059965c832020df757ae8b6b0f9cefdb21f154b820e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://62135a7a7fa690060354d059965c832020df757ae8b6b0f9cefdb21f154b820e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:16Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:35Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:35 crc kubenswrapper[4813]: I1201 08:41:35.584143 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:35 crc kubenswrapper[4813]: I1201 08:41:35.584211 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:35 crc kubenswrapper[4813]: I1201 08:41:35.584230 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:35 crc kubenswrapper[4813]: I1201 08:41:35.584261 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:35 crc kubenswrapper[4813]: I1201 08:41:35.584279 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:35Z","lastTransitionTime":"2025-12-01T08:41:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:35 crc kubenswrapper[4813]: I1201 08:41:35.597169 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"349332bd-6b4d-4306-84ae-07019ca3860f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d441a348087ae3cadda1ad8a7d1c4dff02cf3f6e1d346c294b278fdd9131edcc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://da67635daf6d9e32937f4d5d4ceeee3f8f09ae4fd2dd26c3e85be1286c5949de\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e97356ee274ea78d9f237ffbc756abb6ac8df0dbac244bf0a86ed215b8cbf43\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8dc6feb9817f4f356b8e36af43ec3070774f312fbbb4909fa05b8b3e8dd53582\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a0404f5a5ed5a9750d4c479b608e68e8960b448f455d5af9d219aa5edaeb789b\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"message\\\":\\\"\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764578421\\\\\\\\\\\\\\\" (2025-12-01 08:40:21 +0000 UTC to 2025-12-31 08:40:22 +0000 UTC (now=2025-12-01 08:40:38.066068451 +0000 UTC))\\\\\\\"\\\\nI1201 08:40:38.066144 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1201 08:40:38.066166 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1201 08:40:38.066220 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1201 08:40:38.066232 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1201 08:40:38.066338 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764578432\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764578432\\\\\\\\\\\\\\\" (2025-12-01 07:40:32 +0000 UTC to 2026-12-01 07:40:32 +0000 UTC (now=2025-12-01 08:40:38.066301217 +0000 UTC))\\\\\\\"\\\\nI1201 08:40:38.066357 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1201 08:40:38.066385 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1201 08:40:38.066402 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1201 08:40:38.066414 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-63890811/tls.crt::/tmp/serving-cert-63890811/tls.key\\\\\\\"\\\\nI1201 08:40:38.066419 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1201 08:40:38.066572 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1201 08:40:38.067159 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7354f9337d42115bb1d974bf3772f2d03365acba57f2155135c7a0bbab39879e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:19Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05492df5703392cdf0932d6bf66025b18e53bd05fa0857a7cb2ad70979cbdc3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05492df5703392cdf0932d6bf66025b18e53bd05fa0857a7cb2ad70979cbdc3e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:16Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:35Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:35 crc kubenswrapper[4813]: I1201 08:41:35.617798 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-8pg2p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dd53b38d-d9dd-4f98-9d06-dd95f1a4e3a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a214d71293e4e09573c665524d880fc319dfb1c0c6d95d11d78f253865660fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnjhq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ddf9d5989e4195b3074de08f99d53b48066f53fe93ce204145b3c714996fc57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnjhq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:53Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-8pg2p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:35Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:35 crc kubenswrapper[4813]: I1201 08:41:35.638312 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-9wn4q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3d423402-e08f-42e9-b60c-6eb9bf2f6d48\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3049b842978c6580e1033a442a37766a2e76e278aface9a89fc0b2c57faf0df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rzxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c915733072b32a63237e2401cbf84b059aca396c2c0db884f0f439d599a35d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rzxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:41Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-9wn4q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:35Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:35 crc kubenswrapper[4813]: I1201 08:41:35.652419 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-c576v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"17292cd0-1eac-40b5-a60e-67167c57409c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a42206c914d0493ca74788ea1a941de80920ab7b810fb95227cd9b8711e8f4b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7zwfx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:41Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-c576v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:35Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:35 crc kubenswrapper[4813]: I1201 08:41:35.676433 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-j4pcc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b7505e14-3560-4d0d-be2c-d6a7a5953865\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ccc30a63dc1c23ff9880fb04af1f31f0656b2696d359988e23deed23989f480a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://944977e9722025088720107530d0c91fc693593f380e4b49678f28a56da180cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc32d930ab23f30103d8fa42291909cbec89ebb75c5c53cd783cd27fe1a3f09e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f19c67b1c6c80c7f3223512102810d46ede1445464d21ae3bbea83424539b2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d984c90302e57f3e5a6f3ca3613b415db04ac74a5a228582f1b52ef1d59f93e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce8963bf1e9ee66926907aaf3a03d18977f4d0a1983d7121b62fbc8e02a37e4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb17fb9ae11fa62235328567418fb259f4f8540ebac635842d3afe751dcbde8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fb17fb9ae11fa62235328567418fb259f4f8540ebac635842d3afe751dcbde8c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T08:41:21Z\\\",\\\"message\\\":\\\"guring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-console-operator/metrics]} name:Service_openshift-console-operator/metrics_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.88:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {ebd4748e-0473-49fb-88ad-83dbb221791a}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1201 08:41:21.455187 6576 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1201 08:41:21.455499 6576 obj_retry.go:303] Retry object setup: *v1.Pod openshift-image-registry/node-ca-xpnlc\\\\nF1201 08:41:21.455509 6576 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal erro\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T08:41:20Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-j4pcc_openshift-ovn-kubernetes(b7505e14-3560-4d0d-be2c-d6a7a5953865)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b45454b440c5108fa150a6b856afa1778ba6e9f89633433cb3117b68f00f72c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd67634e0c806323334495401a4e20b48fea4bc5013acd973d2abd3d70870f99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd67634e0c806323334495401a4e20b48fea4bc5013acd973d2abd3d70870f99\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:41Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-j4pcc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:35Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:35 crc kubenswrapper[4813]: I1201 08:41:35.686916 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:35 crc kubenswrapper[4813]: I1201 08:41:35.686997 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:35 crc kubenswrapper[4813]: I1201 08:41:35.687017 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:35 crc kubenswrapper[4813]: I1201 08:41:35.687040 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:35 crc kubenswrapper[4813]: I1201 08:41:35.687055 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:35Z","lastTransitionTime":"2025-12-01T08:41:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:35 crc kubenswrapper[4813]: I1201 08:41:35.692042 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-h7lvd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d431b613-a91b-4fff-88b9-447804766289\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a32c51c943fcf6ada7e3f614667eb023e29d94c11128d26c1fecca37b03963b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://495593f99e92a9c28100a95e27d06ff5847008c5d996366cb474bedd9b69d5ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://495593f99e92a9c28100a95e27d06ff5847008c5d996366cb474bedd9b69d5ef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://19358f83a6c52957f959fe935a6843fdeef59905cc395e6e52a1fb4a670ac193\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://19358f83a6c52957f959fe935a6843fdeef59905cc395e6e52a1fb4a670ac193\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f089c115d916e339994879d224e3db27642b99aaabf59bcafb2979549052945\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9f089c115d916e339994879d224e3db27642b99aaabf59bcafb2979549052945\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb66f224462305c2005b31ef7bb9ed01efde8101e18518dd42d1dfb0df765bb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb66f224462305c2005b31ef7bb9ed01efde8101e18518dd42d1dfb0df765bb9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c268236c9a155c9b1f47c48d30dbd23948ceb2545592b40f356c60ea1d12e8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c268236c9a155c9b1f47c48d30dbd23948ceb2545592b40f356c60ea1d12e8a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://458aedca5f186e9aaed7e4ee29fd5247221491888b130f55c1fe48d88d565f8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://458aedca5f186e9aaed7e4ee29fd5247221491888b130f55c1fe48d88d565f8f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:41Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-h7lvd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:35Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:35 crc kubenswrapper[4813]: I1201 08:41:35.706828 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"490e526a-902b-4aff-b547-0cefea562764\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb48bd28fd3d29781695a5ae630f4913ef2745e97fbe7d463cca39a31e43ba0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74f8cf73f07b4cde260af429d26413e4e6226e7b177b493dd9c0e5c043ab6e1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3aaf696cefbe398bb6b26acfc5be43bda7e2be18218c8060840350c4106e66a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53de1f0682f36aa83ad8e5c226b95daf9f97df4ae154ebfc2fc287f9f238a4a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53de1f0682f36aa83ad8e5c226b95daf9f97df4ae154ebfc2fc287f9f238a4a8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:17Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:16Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:35Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:35 crc kubenswrapper[4813]: I1201 08:41:35.720248 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:35Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:35 crc kubenswrapper[4813]: I1201 08:41:35.789723 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:35 crc kubenswrapper[4813]: I1201 08:41:35.789754 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:35 crc kubenswrapper[4813]: I1201 08:41:35.789766 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:35 crc kubenswrapper[4813]: I1201 08:41:35.789781 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:35 crc kubenswrapper[4813]: I1201 08:41:35.789791 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:35Z","lastTransitionTime":"2025-12-01T08:41:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:35 crc kubenswrapper[4813]: I1201 08:41:35.893082 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:35 crc kubenswrapper[4813]: I1201 08:41:35.893733 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:35 crc kubenswrapper[4813]: I1201 08:41:35.893825 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:35 crc kubenswrapper[4813]: I1201 08:41:35.893911 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:35 crc kubenswrapper[4813]: I1201 08:41:35.894053 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:35Z","lastTransitionTime":"2025-12-01T08:41:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:35 crc kubenswrapper[4813]: I1201 08:41:35.997083 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:35 crc kubenswrapper[4813]: I1201 08:41:35.997118 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:35 crc kubenswrapper[4813]: I1201 08:41:35.997129 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:35 crc kubenswrapper[4813]: I1201 08:41:35.997143 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:35 crc kubenswrapper[4813]: I1201 08:41:35.997152 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:35Z","lastTransitionTime":"2025-12-01T08:41:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:36 crc kubenswrapper[4813]: I1201 08:41:36.100027 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:36 crc kubenswrapper[4813]: I1201 08:41:36.100261 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:36 crc kubenswrapper[4813]: I1201 08:41:36.100375 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:36 crc kubenswrapper[4813]: I1201 08:41:36.100448 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:36 crc kubenswrapper[4813]: I1201 08:41:36.100504 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:36Z","lastTransitionTime":"2025-12-01T08:41:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:36 crc kubenswrapper[4813]: I1201 08:41:36.203952 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:36 crc kubenswrapper[4813]: I1201 08:41:36.204015 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:36 crc kubenswrapper[4813]: I1201 08:41:36.204024 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:36 crc kubenswrapper[4813]: I1201 08:41:36.204040 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:36 crc kubenswrapper[4813]: I1201 08:41:36.204051 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:36Z","lastTransitionTime":"2025-12-01T08:41:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:36 crc kubenswrapper[4813]: I1201 08:41:36.310211 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:36 crc kubenswrapper[4813]: I1201 08:41:36.310282 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:36 crc kubenswrapper[4813]: I1201 08:41:36.310307 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:36 crc kubenswrapper[4813]: I1201 08:41:36.310345 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:36 crc kubenswrapper[4813]: I1201 08:41:36.310379 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:36Z","lastTransitionTime":"2025-12-01T08:41:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:36 crc kubenswrapper[4813]: I1201 08:41:36.423462 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 08:41:36 crc kubenswrapper[4813]: I1201 08:41:36.423468 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 08:41:36 crc kubenswrapper[4813]: E1201 08:41:36.423654 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 08:41:36 crc kubenswrapper[4813]: E1201 08:41:36.424039 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 08:41:36 crc kubenswrapper[4813]: I1201 08:41:36.425387 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:36 crc kubenswrapper[4813]: I1201 08:41:36.425418 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:36 crc kubenswrapper[4813]: I1201 08:41:36.425429 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:36 crc kubenswrapper[4813]: I1201 08:41:36.425447 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:36 crc kubenswrapper[4813]: I1201 08:41:36.425459 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:36Z","lastTransitionTime":"2025-12-01T08:41:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:36 crc kubenswrapper[4813]: I1201 08:41:36.440429 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-9wn4q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3d423402-e08f-42e9-b60c-6eb9bf2f6d48\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3049b842978c6580e1033a442a37766a2e76e278aface9a89fc0b2c57faf0df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rzxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c915733072b32a63237e2401cbf84b059aca396c2c0db884f0f439d599a35d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rzxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:41Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-9wn4q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:36Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:36 crc kubenswrapper[4813]: I1201 08:41:36.452299 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-8pg2p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dd53b38d-d9dd-4f98-9d06-dd95f1a4e3a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a214d71293e4e09573c665524d880fc319dfb1c0c6d95d11d78f253865660fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnjhq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ddf9d5989e4195b3074de08f99d53b48066f53fe93ce204145b3c714996fc57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnjhq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:53Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-8pg2p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:36Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:36 crc kubenswrapper[4813]: I1201 08:41:36.464966 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"490e526a-902b-4aff-b547-0cefea562764\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb48bd28fd3d29781695a5ae630f4913ef2745e97fbe7d463cca39a31e43ba0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74f8cf73f07b4cde260af429d26413e4e6226e7b177b493dd9c0e5c043ab6e1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3aaf696cefbe398bb6b26acfc5be43bda7e2be18218c8060840350c4106e66a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53de1f0682f36aa83ad8e5c226b95daf9f97df4ae154ebfc2fc287f9f238a4a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53de1f0682f36aa83ad8e5c226b95daf9f97df4ae154ebfc2fc287f9f238a4a8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:17Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:16Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:36Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:36 crc kubenswrapper[4813]: I1201 08:41:36.477278 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:36Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:36 crc kubenswrapper[4813]: I1201 08:41:36.487159 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-c576v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"17292cd0-1eac-40b5-a60e-67167c57409c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a42206c914d0493ca74788ea1a941de80920ab7b810fb95227cd9b8711e8f4b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7zwfx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:41Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-c576v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:36Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:36 crc kubenswrapper[4813]: I1201 08:41:36.508208 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-j4pcc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b7505e14-3560-4d0d-be2c-d6a7a5953865\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ccc30a63dc1c23ff9880fb04af1f31f0656b2696d359988e23deed23989f480a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://944977e9722025088720107530d0c91fc693593f380e4b49678f28a56da180cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc32d930ab23f30103d8fa42291909cbec89ebb75c5c53cd783cd27fe1a3f09e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f19c67b1c6c80c7f3223512102810d46ede1445464d21ae3bbea83424539b2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d984c90302e57f3e5a6f3ca3613b415db04ac74a5a228582f1b52ef1d59f93e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce8963bf1e9ee66926907aaf3a03d18977f4d0a1983d7121b62fbc8e02a37e4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb17fb9ae11fa62235328567418fb259f4f8540ebac635842d3afe751dcbde8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fb17fb9ae11fa62235328567418fb259f4f8540ebac635842d3afe751dcbde8c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T08:41:21Z\\\",\\\"message\\\":\\\"guring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-console-operator/metrics]} name:Service_openshift-console-operator/metrics_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.88:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {ebd4748e-0473-49fb-88ad-83dbb221791a}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1201 08:41:21.455187 6576 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1201 08:41:21.455499 6576 obj_retry.go:303] Retry object setup: *v1.Pod openshift-image-registry/node-ca-xpnlc\\\\nF1201 08:41:21.455509 6576 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal erro\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T08:41:20Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-j4pcc_openshift-ovn-kubernetes(b7505e14-3560-4d0d-be2c-d6a7a5953865)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b45454b440c5108fa150a6b856afa1778ba6e9f89633433cb3117b68f00f72c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd67634e0c806323334495401a4e20b48fea4bc5013acd973d2abd3d70870f99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd67634e0c806323334495401a4e20b48fea4bc5013acd973d2abd3d70870f99\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dgqwb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:41Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-j4pcc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:36Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:36 crc kubenswrapper[4813]: I1201 08:41:36.527033 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-h7lvd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d431b613-a91b-4fff-88b9-447804766289\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a32c51c943fcf6ada7e3f614667eb023e29d94c11128d26c1fecca37b03963b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://495593f99e92a9c28100a95e27d06ff5847008c5d996366cb474bedd9b69d5ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://495593f99e92a9c28100a95e27d06ff5847008c5d996366cb474bedd9b69d5ef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://19358f83a6c52957f959fe935a6843fdeef59905cc395e6e52a1fb4a670ac193\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://19358f83a6c52957f959fe935a6843fdeef59905cc395e6e52a1fb4a670ac193\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f089c115d916e339994879d224e3db27642b99aaabf59bcafb2979549052945\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9f089c115d916e339994879d224e3db27642b99aaabf59bcafb2979549052945\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb66f224462305c2005b31ef7bb9ed01efde8101e18518dd42d1dfb0df765bb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb66f224462305c2005b31ef7bb9ed01efde8101e18518dd42d1dfb0df765bb9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c268236c9a155c9b1f47c48d30dbd23948ceb2545592b40f356c60ea1d12e8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c268236c9a155c9b1f47c48d30dbd23948ceb2545592b40f356c60ea1d12e8a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://458aedca5f186e9aaed7e4ee29fd5247221491888b130f55c1fe48d88d565f8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://458aedca5f186e9aaed7e4ee29fd5247221491888b130f55c1fe48d88d565f8f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bq6d5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:41Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-h7lvd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:36Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:36 crc kubenswrapper[4813]: I1201 08:41:36.527761 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:36 crc kubenswrapper[4813]: I1201 08:41:36.527782 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:36 crc kubenswrapper[4813]: I1201 08:41:36.527792 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:36 crc kubenswrapper[4813]: I1201 08:41:36.527816 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:36 crc kubenswrapper[4813]: I1201 08:41:36.527830 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:36Z","lastTransitionTime":"2025-12-01T08:41:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:36 crc kubenswrapper[4813]: I1201 08:41:36.539943 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6fb5a648-ae4f-4e46-a530-007e9bf84abb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e24ebfc52dcb28368a2d436e7d9ab78f07f2aec2a4c71405a6c7d3301eee96a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33898118bbf639c8e6eae5aafa10386b95795512444ed5f535db3c663f169df2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9902c00bf7738686c93f62e1772a484d05715d74b75ee98c84c8f23ed383824b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3657b47ec9f1563c83bfbe3d728c8fd123f7078e219d251b1874269e42c31eea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:16Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:36Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:36 crc kubenswrapper[4813]: I1201 08:41:36.553485 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0023b0e358d75aacc889ce364846dda89ea513832c87fa4b066aea7dac57e8a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:36Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:36 crc kubenswrapper[4813]: I1201 08:41:36.567102 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:36Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:36 crc kubenswrapper[4813]: I1201 08:41:36.580596 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c7574f892d5e9142296b26a0b484eb6759ab53f571139fb6e23b93dc3da05780\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:36Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:36 crc kubenswrapper[4813]: I1201 08:41:36.592242 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:36Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:36 crc kubenswrapper[4813]: I1201 08:41:36.603435 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-xpnlc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"22da1fcf-0c57-4708-90d8-77692af17344\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4a36d213bdad2da91d0d33f400d5a8248befb68d5b94d19347873a93c6712a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xc44d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-xpnlc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:36Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:36 crc kubenswrapper[4813]: I1201 08:41:36.622238 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a6a31593-c1c1-4aab-9e99-f683e908c07f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://86abafcbe7fed4bfbaa6c3b2f64d6a314afa7ed0fe2709bc8a01ba9fb17c1a75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://62135a7a7fa690060354d059965c832020df757ae8b6b0f9cefdb21f154b820e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://62135a7a7fa690060354d059965c832020df757ae8b6b0f9cefdb21f154b820e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:16Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:36Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:36 crc kubenswrapper[4813]: I1201 08:41:36.633004 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:36 crc kubenswrapper[4813]: I1201 08:41:36.633504 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:36 crc kubenswrapper[4813]: I1201 08:41:36.633580 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:36 crc kubenswrapper[4813]: I1201 08:41:36.633694 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:36 crc kubenswrapper[4813]: I1201 08:41:36.633778 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:36Z","lastTransitionTime":"2025-12-01T08:41:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:36 crc kubenswrapper[4813]: I1201 08:41:36.639484 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"349332bd-6b4d-4306-84ae-07019ca3860f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d441a348087ae3cadda1ad8a7d1c4dff02cf3f6e1d346c294b278fdd9131edcc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://da67635daf6d9e32937f4d5d4ceeee3f8f09ae4fd2dd26c3e85be1286c5949de\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e97356ee274ea78d9f237ffbc756abb6ac8df0dbac244bf0a86ed215b8cbf43\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8dc6feb9817f4f356b8e36af43ec3070774f312fbbb4909fa05b8b3e8dd53582\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a0404f5a5ed5a9750d4c479b608e68e8960b448f455d5af9d219aa5edaeb789b\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"message\\\":\\\"\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764578421\\\\\\\\\\\\\\\" (2025-12-01 08:40:21 +0000 UTC to 2025-12-31 08:40:22 +0000 UTC (now=2025-12-01 08:40:38.066068451 +0000 UTC))\\\\\\\"\\\\nI1201 08:40:38.066144 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1201 08:40:38.066166 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1201 08:40:38.066220 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1201 08:40:38.066232 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1201 08:40:38.066338 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764578432\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764578432\\\\\\\\\\\\\\\" (2025-12-01 07:40:32 +0000 UTC to 2026-12-01 07:40:32 +0000 UTC (now=2025-12-01 08:40:38.066301217 +0000 UTC))\\\\\\\"\\\\nI1201 08:40:38.066357 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1201 08:40:38.066385 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1201 08:40:38.066402 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1201 08:40:38.066414 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-63890811/tls.crt::/tmp/serving-cert-63890811/tls.key\\\\\\\"\\\\nI1201 08:40:38.066419 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1201 08:40:38.066572 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1201 08:40:38.067159 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7354f9337d42115bb1d974bf3772f2d03365acba57f2155135c7a0bbab39879e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:19Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05492df5703392cdf0932d6bf66025b18e53bd05fa0857a7cb2ad70979cbdc3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05492df5703392cdf0932d6bf66025b18e53bd05fa0857a7cb2ad70979cbdc3e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:16Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:36Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:36 crc kubenswrapper[4813]: I1201 08:41:36.652744 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3264b3af996af452df2953c59bea6deeb5ea6af1b119ce89f7334595c27e70da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9e879b1f563bd66383d58fd7b435a8a81bd50cb387dc1fabb6fc47d5a3fa468\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:36Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:36 crc kubenswrapper[4813]: I1201 08:41:36.666372 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-b6qqv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9656a03-0096-4c89-b0da-e37e4103611b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b730cdf02e6456d9a1d02d66b94064ca71052eb8f373eaabfe327369eb03a15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c4195aebc1b2e388d480920a7f72257f6660afc99ad8c1a09478ef445739b272\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T08:41:31Z\\\",\\\"message\\\":\\\"2025-12-01T08:40:45+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_038ba082-9223-4817-aa2f-727d3b2383fe\\\\n2025-12-01T08:40:45+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_038ba082-9223-4817-aa2f-727d3b2383fe to /host/opt/cni/bin/\\\\n2025-12-01T08:40:46Z [verbose] multus-daemon started\\\\n2025-12-01T08:40:46Z [verbose] Readiness Indicator file check\\\\n2025-12-01T08:41:31Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:42Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:41:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cjvnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:41Z\\\"}}\" for pod \"openshift-multus\"/\"multus-b6qqv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:36Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:36 crc kubenswrapper[4813]: I1201 08:41:36.680523 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-96mbr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0789e3f7-5b08-4c04-8596-9ba29f50cdd0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xt4fj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xt4fj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:55Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-96mbr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:36Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:36 crc kubenswrapper[4813]: I1201 08:41:36.736206 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:36 crc kubenswrapper[4813]: I1201 08:41:36.736256 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:36 crc kubenswrapper[4813]: I1201 08:41:36.736269 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:36 crc kubenswrapper[4813]: I1201 08:41:36.736305 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:36 crc kubenswrapper[4813]: I1201 08:41:36.736319 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:36Z","lastTransitionTime":"2025-12-01T08:41:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:36 crc kubenswrapper[4813]: I1201 08:41:36.840536 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:36 crc kubenswrapper[4813]: I1201 08:41:36.840602 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:36 crc kubenswrapper[4813]: I1201 08:41:36.840621 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:36 crc kubenswrapper[4813]: I1201 08:41:36.840651 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:36 crc kubenswrapper[4813]: I1201 08:41:36.840668 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:36Z","lastTransitionTime":"2025-12-01T08:41:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:36 crc kubenswrapper[4813]: I1201 08:41:36.944389 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:36 crc kubenswrapper[4813]: I1201 08:41:36.944449 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:36 crc kubenswrapper[4813]: I1201 08:41:36.944463 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:36 crc kubenswrapper[4813]: I1201 08:41:36.944481 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:36 crc kubenswrapper[4813]: I1201 08:41:36.944494 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:36Z","lastTransitionTime":"2025-12-01T08:41:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:37 crc kubenswrapper[4813]: I1201 08:41:37.048184 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:37 crc kubenswrapper[4813]: I1201 08:41:37.048539 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:37 crc kubenswrapper[4813]: I1201 08:41:37.048945 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:37 crc kubenswrapper[4813]: I1201 08:41:37.049447 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:37 crc kubenswrapper[4813]: I1201 08:41:37.049912 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:37Z","lastTransitionTime":"2025-12-01T08:41:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:37 crc kubenswrapper[4813]: I1201 08:41:37.153384 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:37 crc kubenswrapper[4813]: I1201 08:41:37.153676 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:37 crc kubenswrapper[4813]: I1201 08:41:37.153738 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:37 crc kubenswrapper[4813]: I1201 08:41:37.153800 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:37 crc kubenswrapper[4813]: I1201 08:41:37.154086 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:37Z","lastTransitionTime":"2025-12-01T08:41:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:37 crc kubenswrapper[4813]: I1201 08:41:37.257696 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:37 crc kubenswrapper[4813]: I1201 08:41:37.257735 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:37 crc kubenswrapper[4813]: I1201 08:41:37.257744 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:37 crc kubenswrapper[4813]: I1201 08:41:37.257760 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:37 crc kubenswrapper[4813]: I1201 08:41:37.257769 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:37Z","lastTransitionTime":"2025-12-01T08:41:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:37 crc kubenswrapper[4813]: I1201 08:41:37.362499 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:37 crc kubenswrapper[4813]: I1201 08:41:37.362962 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:37 crc kubenswrapper[4813]: I1201 08:41:37.363356 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:37 crc kubenswrapper[4813]: I1201 08:41:37.363587 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:37 crc kubenswrapper[4813]: I1201 08:41:37.363804 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:37Z","lastTransitionTime":"2025-12-01T08:41:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:37 crc kubenswrapper[4813]: I1201 08:41:37.394632 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 08:41:37 crc kubenswrapper[4813]: I1201 08:41:37.394625 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-96mbr" Dec 01 08:41:37 crc kubenswrapper[4813]: E1201 08:41:37.395839 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 08:41:37 crc kubenswrapper[4813]: E1201 08:41:37.396001 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-96mbr" podUID="0789e3f7-5b08-4c04-8596-9ba29f50cdd0" Dec 01 08:41:37 crc kubenswrapper[4813]: I1201 08:41:37.466906 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:37 crc kubenswrapper[4813]: I1201 08:41:37.467344 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:37 crc kubenswrapper[4813]: I1201 08:41:37.467471 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:37 crc kubenswrapper[4813]: I1201 08:41:37.467561 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:37 crc kubenswrapper[4813]: I1201 08:41:37.467623 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:37Z","lastTransitionTime":"2025-12-01T08:41:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:37 crc kubenswrapper[4813]: I1201 08:41:37.572215 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:37 crc kubenswrapper[4813]: I1201 08:41:37.572723 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:37 crc kubenswrapper[4813]: I1201 08:41:37.572796 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:37 crc kubenswrapper[4813]: I1201 08:41:37.572872 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:37 crc kubenswrapper[4813]: I1201 08:41:37.572970 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:37Z","lastTransitionTime":"2025-12-01T08:41:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:37 crc kubenswrapper[4813]: I1201 08:41:37.676638 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:37 crc kubenswrapper[4813]: I1201 08:41:37.676719 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:37 crc kubenswrapper[4813]: I1201 08:41:37.676751 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:37 crc kubenswrapper[4813]: I1201 08:41:37.676785 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:37 crc kubenswrapper[4813]: I1201 08:41:37.676808 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:37Z","lastTransitionTime":"2025-12-01T08:41:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:37 crc kubenswrapper[4813]: I1201 08:41:37.781307 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:37 crc kubenswrapper[4813]: I1201 08:41:37.781806 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:37 crc kubenswrapper[4813]: I1201 08:41:37.782220 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:37 crc kubenswrapper[4813]: I1201 08:41:37.782389 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:37 crc kubenswrapper[4813]: I1201 08:41:37.782480 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:37Z","lastTransitionTime":"2025-12-01T08:41:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:37 crc kubenswrapper[4813]: I1201 08:41:37.885078 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:37 crc kubenswrapper[4813]: I1201 08:41:37.885136 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:37 crc kubenswrapper[4813]: I1201 08:41:37.885155 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:37 crc kubenswrapper[4813]: I1201 08:41:37.885195 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:37 crc kubenswrapper[4813]: I1201 08:41:37.885209 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:37Z","lastTransitionTime":"2025-12-01T08:41:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:37 crc kubenswrapper[4813]: I1201 08:41:37.988806 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:37 crc kubenswrapper[4813]: I1201 08:41:37.989606 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:37 crc kubenswrapper[4813]: I1201 08:41:37.989716 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:37 crc kubenswrapper[4813]: I1201 08:41:37.989823 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:37 crc kubenswrapper[4813]: I1201 08:41:37.989921 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:37Z","lastTransitionTime":"2025-12-01T08:41:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:38 crc kubenswrapper[4813]: I1201 08:41:38.092836 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:38 crc kubenswrapper[4813]: I1201 08:41:38.093172 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:38 crc kubenswrapper[4813]: I1201 08:41:38.093278 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:38 crc kubenswrapper[4813]: I1201 08:41:38.093385 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:38 crc kubenswrapper[4813]: I1201 08:41:38.093476 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:38Z","lastTransitionTime":"2025-12-01T08:41:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:38 crc kubenswrapper[4813]: I1201 08:41:38.196504 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:38 crc kubenswrapper[4813]: I1201 08:41:38.196883 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:38 crc kubenswrapper[4813]: I1201 08:41:38.196993 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:38 crc kubenswrapper[4813]: I1201 08:41:38.197136 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:38 crc kubenswrapper[4813]: I1201 08:41:38.197254 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:38Z","lastTransitionTime":"2025-12-01T08:41:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:38 crc kubenswrapper[4813]: I1201 08:41:38.305514 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:38 crc kubenswrapper[4813]: I1201 08:41:38.305769 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:38 crc kubenswrapper[4813]: I1201 08:41:38.305838 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:38 crc kubenswrapper[4813]: I1201 08:41:38.305933 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:38 crc kubenswrapper[4813]: I1201 08:41:38.306061 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:38Z","lastTransitionTime":"2025-12-01T08:41:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:38 crc kubenswrapper[4813]: I1201 08:41:38.393207 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 08:41:38 crc kubenswrapper[4813]: I1201 08:41:38.393243 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 08:41:38 crc kubenswrapper[4813]: E1201 08:41:38.393772 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 08:41:38 crc kubenswrapper[4813]: E1201 08:41:38.393797 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 08:41:38 crc kubenswrapper[4813]: I1201 08:41:38.408535 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:38 crc kubenswrapper[4813]: I1201 08:41:38.408785 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:38 crc kubenswrapper[4813]: I1201 08:41:38.408879 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:38 crc kubenswrapper[4813]: I1201 08:41:38.409071 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:38 crc kubenswrapper[4813]: I1201 08:41:38.409159 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:38Z","lastTransitionTime":"2025-12-01T08:41:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:38 crc kubenswrapper[4813]: I1201 08:41:38.512114 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:38 crc kubenswrapper[4813]: I1201 08:41:38.512159 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:38 crc kubenswrapper[4813]: I1201 08:41:38.512176 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:38 crc kubenswrapper[4813]: I1201 08:41:38.512198 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:38 crc kubenswrapper[4813]: I1201 08:41:38.512213 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:38Z","lastTransitionTime":"2025-12-01T08:41:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:38 crc kubenswrapper[4813]: I1201 08:41:38.614962 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:38 crc kubenswrapper[4813]: I1201 08:41:38.615078 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:38 crc kubenswrapper[4813]: I1201 08:41:38.615100 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:38 crc kubenswrapper[4813]: I1201 08:41:38.615130 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:38 crc kubenswrapper[4813]: I1201 08:41:38.615153 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:38Z","lastTransitionTime":"2025-12-01T08:41:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:38 crc kubenswrapper[4813]: I1201 08:41:38.718446 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:38 crc kubenswrapper[4813]: I1201 08:41:38.718804 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:38 crc kubenswrapper[4813]: I1201 08:41:38.719017 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:38 crc kubenswrapper[4813]: I1201 08:41:38.719179 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:38 crc kubenswrapper[4813]: I1201 08:41:38.719297 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:38Z","lastTransitionTime":"2025-12-01T08:41:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:38 crc kubenswrapper[4813]: I1201 08:41:38.822106 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:38 crc kubenswrapper[4813]: I1201 08:41:38.822141 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:38 crc kubenswrapper[4813]: I1201 08:41:38.822154 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:38 crc kubenswrapper[4813]: I1201 08:41:38.822171 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:38 crc kubenswrapper[4813]: I1201 08:41:38.822183 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:38Z","lastTransitionTime":"2025-12-01T08:41:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:38 crc kubenswrapper[4813]: I1201 08:41:38.924676 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:38 crc kubenswrapper[4813]: I1201 08:41:38.924765 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:38 crc kubenswrapper[4813]: I1201 08:41:38.924788 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:38 crc kubenswrapper[4813]: I1201 08:41:38.924818 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:38 crc kubenswrapper[4813]: I1201 08:41:38.924838 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:38Z","lastTransitionTime":"2025-12-01T08:41:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:39 crc kubenswrapper[4813]: I1201 08:41:39.028894 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:39 crc kubenswrapper[4813]: I1201 08:41:39.028952 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:39 crc kubenswrapper[4813]: I1201 08:41:39.029014 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:39 crc kubenswrapper[4813]: I1201 08:41:39.029046 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:39 crc kubenswrapper[4813]: I1201 08:41:39.029063 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:39Z","lastTransitionTime":"2025-12-01T08:41:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:39 crc kubenswrapper[4813]: I1201 08:41:39.132300 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:39 crc kubenswrapper[4813]: I1201 08:41:39.132807 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:39 crc kubenswrapper[4813]: I1201 08:41:39.132959 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:39 crc kubenswrapper[4813]: I1201 08:41:39.133303 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:39 crc kubenswrapper[4813]: I1201 08:41:39.133501 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:39Z","lastTransitionTime":"2025-12-01T08:41:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:39 crc kubenswrapper[4813]: I1201 08:41:39.264329 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:39 crc kubenswrapper[4813]: I1201 08:41:39.264543 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:39 crc kubenswrapper[4813]: I1201 08:41:39.264644 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:39 crc kubenswrapper[4813]: I1201 08:41:39.264712 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:39 crc kubenswrapper[4813]: I1201 08:41:39.264770 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:39Z","lastTransitionTime":"2025-12-01T08:41:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:39 crc kubenswrapper[4813]: I1201 08:41:39.368067 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:39 crc kubenswrapper[4813]: I1201 08:41:39.368146 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:39 crc kubenswrapper[4813]: I1201 08:41:39.368171 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:39 crc kubenswrapper[4813]: I1201 08:41:39.368200 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:39 crc kubenswrapper[4813]: I1201 08:41:39.368217 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:39Z","lastTransitionTime":"2025-12-01T08:41:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:39 crc kubenswrapper[4813]: I1201 08:41:39.393541 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-96mbr" Dec 01 08:41:39 crc kubenswrapper[4813]: E1201 08:41:39.394071 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-96mbr" podUID="0789e3f7-5b08-4c04-8596-9ba29f50cdd0" Dec 01 08:41:39 crc kubenswrapper[4813]: I1201 08:41:39.393571 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 08:41:39 crc kubenswrapper[4813]: E1201 08:41:39.407268 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 08:41:39 crc kubenswrapper[4813]: I1201 08:41:39.471479 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:39 crc kubenswrapper[4813]: I1201 08:41:39.471548 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:39 crc kubenswrapper[4813]: I1201 08:41:39.471565 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:39 crc kubenswrapper[4813]: I1201 08:41:39.471597 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:39 crc kubenswrapper[4813]: I1201 08:41:39.471615 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:39Z","lastTransitionTime":"2025-12-01T08:41:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:39 crc kubenswrapper[4813]: I1201 08:41:39.575172 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:39 crc kubenswrapper[4813]: I1201 08:41:39.575220 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:39 crc kubenswrapper[4813]: I1201 08:41:39.575232 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:39 crc kubenswrapper[4813]: I1201 08:41:39.575254 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:39 crc kubenswrapper[4813]: I1201 08:41:39.575269 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:39Z","lastTransitionTime":"2025-12-01T08:41:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:39 crc kubenswrapper[4813]: I1201 08:41:39.685291 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:39 crc kubenswrapper[4813]: I1201 08:41:39.685602 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:39 crc kubenswrapper[4813]: I1201 08:41:39.685610 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:39 crc kubenswrapper[4813]: I1201 08:41:39.685624 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:39 crc kubenswrapper[4813]: I1201 08:41:39.685633 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:39Z","lastTransitionTime":"2025-12-01T08:41:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:39 crc kubenswrapper[4813]: I1201 08:41:39.789119 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:39 crc kubenswrapper[4813]: I1201 08:41:39.789166 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:39 crc kubenswrapper[4813]: I1201 08:41:39.789176 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:39 crc kubenswrapper[4813]: I1201 08:41:39.789193 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:39 crc kubenswrapper[4813]: I1201 08:41:39.789203 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:39Z","lastTransitionTime":"2025-12-01T08:41:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:39 crc kubenswrapper[4813]: I1201 08:41:39.892127 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:39 crc kubenswrapper[4813]: I1201 08:41:39.892172 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:39 crc kubenswrapper[4813]: I1201 08:41:39.892183 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:39 crc kubenswrapper[4813]: I1201 08:41:39.892214 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:39 crc kubenswrapper[4813]: I1201 08:41:39.892226 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:39Z","lastTransitionTime":"2025-12-01T08:41:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:39 crc kubenswrapper[4813]: I1201 08:41:39.995285 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:39 crc kubenswrapper[4813]: I1201 08:41:39.995328 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:39 crc kubenswrapper[4813]: I1201 08:41:39.995340 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:39 crc kubenswrapper[4813]: I1201 08:41:39.995358 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:39 crc kubenswrapper[4813]: I1201 08:41:39.995373 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:39Z","lastTransitionTime":"2025-12-01T08:41:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:40 crc kubenswrapper[4813]: I1201 08:41:40.098273 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:40 crc kubenswrapper[4813]: I1201 08:41:40.098319 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:40 crc kubenswrapper[4813]: I1201 08:41:40.098335 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:40 crc kubenswrapper[4813]: I1201 08:41:40.098353 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:40 crc kubenswrapper[4813]: I1201 08:41:40.098365 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:40Z","lastTransitionTime":"2025-12-01T08:41:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:40 crc kubenswrapper[4813]: I1201 08:41:40.201720 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:40 crc kubenswrapper[4813]: I1201 08:41:40.201761 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:40 crc kubenswrapper[4813]: I1201 08:41:40.201772 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:40 crc kubenswrapper[4813]: I1201 08:41:40.201789 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:40 crc kubenswrapper[4813]: I1201 08:41:40.201801 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:40Z","lastTransitionTime":"2025-12-01T08:41:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:40 crc kubenswrapper[4813]: I1201 08:41:40.304454 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:40 crc kubenswrapper[4813]: I1201 08:41:40.304522 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:40 crc kubenswrapper[4813]: I1201 08:41:40.304535 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:40 crc kubenswrapper[4813]: I1201 08:41:40.304556 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:40 crc kubenswrapper[4813]: I1201 08:41:40.304570 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:40Z","lastTransitionTime":"2025-12-01T08:41:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:40 crc kubenswrapper[4813]: I1201 08:41:40.393409 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 08:41:40 crc kubenswrapper[4813]: E1201 08:41:40.393563 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 08:41:40 crc kubenswrapper[4813]: I1201 08:41:40.393874 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 08:41:40 crc kubenswrapper[4813]: E1201 08:41:40.393948 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 08:41:40 crc kubenswrapper[4813]: I1201 08:41:40.406525 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:40 crc kubenswrapper[4813]: I1201 08:41:40.406570 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:40 crc kubenswrapper[4813]: I1201 08:41:40.406578 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:40 crc kubenswrapper[4813]: I1201 08:41:40.406589 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:40 crc kubenswrapper[4813]: I1201 08:41:40.406598 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:40Z","lastTransitionTime":"2025-12-01T08:41:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:40 crc kubenswrapper[4813]: I1201 08:41:40.509670 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:40 crc kubenswrapper[4813]: I1201 08:41:40.509743 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:40 crc kubenswrapper[4813]: I1201 08:41:40.509762 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:40 crc kubenswrapper[4813]: I1201 08:41:40.509788 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:40 crc kubenswrapper[4813]: I1201 08:41:40.509806 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:40Z","lastTransitionTime":"2025-12-01T08:41:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:40 crc kubenswrapper[4813]: I1201 08:41:40.613211 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:40 crc kubenswrapper[4813]: I1201 08:41:40.613259 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:40 crc kubenswrapper[4813]: I1201 08:41:40.613271 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:40 crc kubenswrapper[4813]: I1201 08:41:40.613291 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:40 crc kubenswrapper[4813]: I1201 08:41:40.613315 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:40Z","lastTransitionTime":"2025-12-01T08:41:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:40 crc kubenswrapper[4813]: I1201 08:41:40.717828 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:40 crc kubenswrapper[4813]: I1201 08:41:40.717895 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:40 crc kubenswrapper[4813]: I1201 08:41:40.717913 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:40 crc kubenswrapper[4813]: I1201 08:41:40.717941 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:40 crc kubenswrapper[4813]: I1201 08:41:40.717992 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:40Z","lastTransitionTime":"2025-12-01T08:41:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:40 crc kubenswrapper[4813]: I1201 08:41:40.825538 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:40 crc kubenswrapper[4813]: I1201 08:41:40.825864 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:40 crc kubenswrapper[4813]: I1201 08:41:40.825874 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:40 crc kubenswrapper[4813]: I1201 08:41:40.825891 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:40 crc kubenswrapper[4813]: I1201 08:41:40.825901 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:40Z","lastTransitionTime":"2025-12-01T08:41:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:40 crc kubenswrapper[4813]: I1201 08:41:40.929130 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:40 crc kubenswrapper[4813]: I1201 08:41:40.929224 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:40 crc kubenswrapper[4813]: I1201 08:41:40.929242 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:40 crc kubenswrapper[4813]: I1201 08:41:40.929267 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:40 crc kubenswrapper[4813]: I1201 08:41:40.929284 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:40Z","lastTransitionTime":"2025-12-01T08:41:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:41 crc kubenswrapper[4813]: I1201 08:41:41.031725 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:41 crc kubenswrapper[4813]: I1201 08:41:41.031769 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:41 crc kubenswrapper[4813]: I1201 08:41:41.031777 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:41 crc kubenswrapper[4813]: I1201 08:41:41.031792 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:41 crc kubenswrapper[4813]: I1201 08:41:41.031801 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:41Z","lastTransitionTime":"2025-12-01T08:41:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:41 crc kubenswrapper[4813]: I1201 08:41:41.134913 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:41 crc kubenswrapper[4813]: I1201 08:41:41.134995 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:41 crc kubenswrapper[4813]: I1201 08:41:41.135013 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:41 crc kubenswrapper[4813]: I1201 08:41:41.135035 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:41 crc kubenswrapper[4813]: I1201 08:41:41.135049 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:41Z","lastTransitionTime":"2025-12-01T08:41:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:41 crc kubenswrapper[4813]: I1201 08:41:41.238466 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:41 crc kubenswrapper[4813]: I1201 08:41:41.238517 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:41 crc kubenswrapper[4813]: I1201 08:41:41.238532 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:41 crc kubenswrapper[4813]: I1201 08:41:41.238553 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:41 crc kubenswrapper[4813]: I1201 08:41:41.238568 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:41Z","lastTransitionTime":"2025-12-01T08:41:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:41 crc kubenswrapper[4813]: I1201 08:41:41.342418 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:41 crc kubenswrapper[4813]: I1201 08:41:41.342471 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:41 crc kubenswrapper[4813]: I1201 08:41:41.342483 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:41 crc kubenswrapper[4813]: I1201 08:41:41.342504 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:41 crc kubenswrapper[4813]: I1201 08:41:41.342517 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:41Z","lastTransitionTime":"2025-12-01T08:41:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:41 crc kubenswrapper[4813]: I1201 08:41:41.376953 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:41 crc kubenswrapper[4813]: I1201 08:41:41.377088 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:41 crc kubenswrapper[4813]: I1201 08:41:41.377108 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:41 crc kubenswrapper[4813]: I1201 08:41:41.377135 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:41 crc kubenswrapper[4813]: I1201 08:41:41.377153 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:41Z","lastTransitionTime":"2025-12-01T08:41:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:41 crc kubenswrapper[4813]: I1201 08:41:41.393474 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 08:41:41 crc kubenswrapper[4813]: I1201 08:41:41.393474 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-96mbr" Dec 01 08:41:41 crc kubenswrapper[4813]: E1201 08:41:41.393786 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 08:41:41 crc kubenswrapper[4813]: E1201 08:41:41.393900 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-96mbr" podUID="0789e3f7-5b08-4c04-8596-9ba29f50cdd0" Dec 01 08:41:41 crc kubenswrapper[4813]: E1201 08:41:41.403309 4813 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:41:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:41:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:41Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:41:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:41:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:41Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cd271ba1-a5c5-44e0-b03e-cfd1f5e8cb76\\\",\\\"systemUUID\\\":\\\"906e3783-eb6d-4f21-889c-7989aeaa25ba\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:41Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:41 crc kubenswrapper[4813]: I1201 08:41:41.410624 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:41 crc kubenswrapper[4813]: I1201 08:41:41.410696 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:41 crc kubenswrapper[4813]: I1201 08:41:41.410713 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:41 crc kubenswrapper[4813]: I1201 08:41:41.410741 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:41 crc kubenswrapper[4813]: I1201 08:41:41.410759 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:41Z","lastTransitionTime":"2025-12-01T08:41:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:41 crc kubenswrapper[4813]: E1201 08:41:41.431795 4813 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:41:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:41:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:41Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:41:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:41:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:41Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cd271ba1-a5c5-44e0-b03e-cfd1f5e8cb76\\\",\\\"systemUUID\\\":\\\"906e3783-eb6d-4f21-889c-7989aeaa25ba\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:41Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:41 crc kubenswrapper[4813]: I1201 08:41:41.437535 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:41 crc kubenswrapper[4813]: I1201 08:41:41.437596 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:41 crc kubenswrapper[4813]: I1201 08:41:41.437613 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:41 crc kubenswrapper[4813]: I1201 08:41:41.437640 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:41 crc kubenswrapper[4813]: I1201 08:41:41.437659 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:41Z","lastTransitionTime":"2025-12-01T08:41:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:41 crc kubenswrapper[4813]: E1201 08:41:41.456722 4813 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:41:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:41:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:41Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:41:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:41:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:41Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cd271ba1-a5c5-44e0-b03e-cfd1f5e8cb76\\\",\\\"systemUUID\\\":\\\"906e3783-eb6d-4f21-889c-7989aeaa25ba\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:41Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:41 crc kubenswrapper[4813]: I1201 08:41:41.462056 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:41 crc kubenswrapper[4813]: I1201 08:41:41.462118 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:41 crc kubenswrapper[4813]: I1201 08:41:41.462136 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:41 crc kubenswrapper[4813]: I1201 08:41:41.462163 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:41 crc kubenswrapper[4813]: I1201 08:41:41.462180 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:41Z","lastTransitionTime":"2025-12-01T08:41:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:41 crc kubenswrapper[4813]: E1201 08:41:41.485025 4813 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:41:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:41:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:41Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:41:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:41:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:41Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cd271ba1-a5c5-44e0-b03e-cfd1f5e8cb76\\\",\\\"systemUUID\\\":\\\"906e3783-eb6d-4f21-889c-7989aeaa25ba\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:41Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:41 crc kubenswrapper[4813]: I1201 08:41:41.490323 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:41 crc kubenswrapper[4813]: I1201 08:41:41.490390 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:41 crc kubenswrapper[4813]: I1201 08:41:41.490414 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:41 crc kubenswrapper[4813]: I1201 08:41:41.490441 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:41 crc kubenswrapper[4813]: I1201 08:41:41.490464 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:41Z","lastTransitionTime":"2025-12-01T08:41:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:41 crc kubenswrapper[4813]: E1201 08:41:41.510440 4813 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:41:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:41:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:41Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:41:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:41:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:41Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cd271ba1-a5c5-44e0-b03e-cfd1f5e8cb76\\\",\\\"systemUUID\\\":\\\"906e3783-eb6d-4f21-889c-7989aeaa25ba\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:41Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:41 crc kubenswrapper[4813]: E1201 08:41:41.510665 4813 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 01 08:41:41 crc kubenswrapper[4813]: I1201 08:41:41.513224 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:41 crc kubenswrapper[4813]: I1201 08:41:41.513273 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:41 crc kubenswrapper[4813]: I1201 08:41:41.513290 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:41 crc kubenswrapper[4813]: I1201 08:41:41.513329 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:41 crc kubenswrapper[4813]: I1201 08:41:41.513346 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:41Z","lastTransitionTime":"2025-12-01T08:41:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:41 crc kubenswrapper[4813]: I1201 08:41:41.616490 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:41 crc kubenswrapper[4813]: I1201 08:41:41.616541 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:41 crc kubenswrapper[4813]: I1201 08:41:41.616558 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:41 crc kubenswrapper[4813]: I1201 08:41:41.616582 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:41 crc kubenswrapper[4813]: I1201 08:41:41.616600 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:41Z","lastTransitionTime":"2025-12-01T08:41:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:41 crc kubenswrapper[4813]: I1201 08:41:41.719615 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:41 crc kubenswrapper[4813]: I1201 08:41:41.719729 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:41 crc kubenswrapper[4813]: I1201 08:41:41.719796 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:41 crc kubenswrapper[4813]: I1201 08:41:41.719844 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:41 crc kubenswrapper[4813]: I1201 08:41:41.719870 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:41Z","lastTransitionTime":"2025-12-01T08:41:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:41 crc kubenswrapper[4813]: I1201 08:41:41.823330 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:41 crc kubenswrapper[4813]: I1201 08:41:41.823402 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:41 crc kubenswrapper[4813]: I1201 08:41:41.823428 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:41 crc kubenswrapper[4813]: I1201 08:41:41.823462 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:41 crc kubenswrapper[4813]: I1201 08:41:41.823481 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:41Z","lastTransitionTime":"2025-12-01T08:41:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:41 crc kubenswrapper[4813]: I1201 08:41:41.948221 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:41 crc kubenswrapper[4813]: I1201 08:41:41.948260 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:41 crc kubenswrapper[4813]: I1201 08:41:41.948269 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:41 crc kubenswrapper[4813]: I1201 08:41:41.948284 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:41 crc kubenswrapper[4813]: I1201 08:41:41.948293 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:41Z","lastTransitionTime":"2025-12-01T08:41:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:42 crc kubenswrapper[4813]: I1201 08:41:42.051192 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:42 crc kubenswrapper[4813]: I1201 08:41:42.051240 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:42 crc kubenswrapper[4813]: I1201 08:41:42.051250 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:42 crc kubenswrapper[4813]: I1201 08:41:42.051271 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:42 crc kubenswrapper[4813]: I1201 08:41:42.051282 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:42Z","lastTransitionTime":"2025-12-01T08:41:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:42 crc kubenswrapper[4813]: I1201 08:41:42.154170 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:42 crc kubenswrapper[4813]: I1201 08:41:42.154246 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:42 crc kubenswrapper[4813]: I1201 08:41:42.154265 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:42 crc kubenswrapper[4813]: I1201 08:41:42.154295 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:42 crc kubenswrapper[4813]: I1201 08:41:42.154313 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:42Z","lastTransitionTime":"2025-12-01T08:41:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:42 crc kubenswrapper[4813]: I1201 08:41:42.256896 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:42 crc kubenswrapper[4813]: I1201 08:41:42.256956 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:42 crc kubenswrapper[4813]: I1201 08:41:42.257031 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:42 crc kubenswrapper[4813]: I1201 08:41:42.257064 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:42 crc kubenswrapper[4813]: I1201 08:41:42.257081 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:42Z","lastTransitionTime":"2025-12-01T08:41:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:42 crc kubenswrapper[4813]: I1201 08:41:42.306579 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 08:41:42 crc kubenswrapper[4813]: I1201 08:41:42.306815 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 08:41:42 crc kubenswrapper[4813]: I1201 08:41:42.306854 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 08:41:42 crc kubenswrapper[4813]: E1201 08:41:42.306911 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 08:42:46.306854796 +0000 UTC m=+150.929776392 (durationBeforeRetry 1m4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:41:42 crc kubenswrapper[4813]: E1201 08:41:42.307232 4813 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 01 08:41:42 crc kubenswrapper[4813]: E1201 08:41:42.307236 4813 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 01 08:41:42 crc kubenswrapper[4813]: E1201 08:41:42.307335 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-01 08:42:46.307316198 +0000 UTC m=+150.930237804 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 01 08:41:42 crc kubenswrapper[4813]: E1201 08:41:42.307371 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-01 08:42:46.30734592 +0000 UTC m=+150.930267536 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 01 08:41:42 crc kubenswrapper[4813]: I1201 08:41:42.359694 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:42 crc kubenswrapper[4813]: I1201 08:41:42.359739 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:42 crc kubenswrapper[4813]: I1201 08:41:42.359751 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:42 crc kubenswrapper[4813]: I1201 08:41:42.359771 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:42 crc kubenswrapper[4813]: I1201 08:41:42.359783 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:42Z","lastTransitionTime":"2025-12-01T08:41:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:42 crc kubenswrapper[4813]: I1201 08:41:42.393497 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 08:41:42 crc kubenswrapper[4813]: I1201 08:41:42.393539 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 08:41:42 crc kubenswrapper[4813]: E1201 08:41:42.393656 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 08:41:42 crc kubenswrapper[4813]: E1201 08:41:42.393804 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 08:41:42 crc kubenswrapper[4813]: I1201 08:41:42.407409 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 08:41:42 crc kubenswrapper[4813]: I1201 08:41:42.407476 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 08:41:42 crc kubenswrapper[4813]: E1201 08:41:42.407720 4813 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 01 08:41:42 crc kubenswrapper[4813]: E1201 08:41:42.407795 4813 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 01 08:41:42 crc kubenswrapper[4813]: E1201 08:41:42.407836 4813 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 08:41:42 crc kubenswrapper[4813]: E1201 08:41:42.407727 4813 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 01 08:41:42 crc kubenswrapper[4813]: E1201 08:41:42.407916 4813 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 01 08:41:42 crc kubenswrapper[4813]: E1201 08:41:42.407942 4813 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 08:41:42 crc kubenswrapper[4813]: E1201 08:41:42.407924 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-01 08:42:46.407898433 +0000 UTC m=+151.030820059 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 08:41:42 crc kubenswrapper[4813]: E1201 08:41:42.408071 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-01 08:42:46.408046691 +0000 UTC m=+151.030968307 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 08:41:42 crc kubenswrapper[4813]: I1201 08:41:42.462504 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:42 crc kubenswrapper[4813]: I1201 08:41:42.462583 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:42 crc kubenswrapper[4813]: I1201 08:41:42.462596 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:42 crc kubenswrapper[4813]: I1201 08:41:42.462638 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:42 crc kubenswrapper[4813]: I1201 08:41:42.462652 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:42Z","lastTransitionTime":"2025-12-01T08:41:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:42 crc kubenswrapper[4813]: I1201 08:41:42.566124 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:42 crc kubenswrapper[4813]: I1201 08:41:42.566200 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:42 crc kubenswrapper[4813]: I1201 08:41:42.566250 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:42 crc kubenswrapper[4813]: I1201 08:41:42.566329 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:42 crc kubenswrapper[4813]: I1201 08:41:42.566356 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:42Z","lastTransitionTime":"2025-12-01T08:41:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:42 crc kubenswrapper[4813]: I1201 08:41:42.675064 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:42 crc kubenswrapper[4813]: I1201 08:41:42.675144 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:42 crc kubenswrapper[4813]: I1201 08:41:42.675168 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:42 crc kubenswrapper[4813]: I1201 08:41:42.675200 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:42 crc kubenswrapper[4813]: I1201 08:41:42.675258 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:42Z","lastTransitionTime":"2025-12-01T08:41:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:42 crc kubenswrapper[4813]: I1201 08:41:42.777580 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:42 crc kubenswrapper[4813]: I1201 08:41:42.777620 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:42 crc kubenswrapper[4813]: I1201 08:41:42.777628 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:42 crc kubenswrapper[4813]: I1201 08:41:42.777642 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:42 crc kubenswrapper[4813]: I1201 08:41:42.777650 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:42Z","lastTransitionTime":"2025-12-01T08:41:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:42 crc kubenswrapper[4813]: I1201 08:41:42.880965 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:42 crc kubenswrapper[4813]: I1201 08:41:42.881091 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:42 crc kubenswrapper[4813]: I1201 08:41:42.881108 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:42 crc kubenswrapper[4813]: I1201 08:41:42.881136 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:42 crc kubenswrapper[4813]: I1201 08:41:42.881155 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:42Z","lastTransitionTime":"2025-12-01T08:41:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:42 crc kubenswrapper[4813]: I1201 08:41:42.984707 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:42 crc kubenswrapper[4813]: I1201 08:41:42.984768 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:42 crc kubenswrapper[4813]: I1201 08:41:42.984785 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:42 crc kubenswrapper[4813]: I1201 08:41:42.984809 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:42 crc kubenswrapper[4813]: I1201 08:41:42.984826 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:42Z","lastTransitionTime":"2025-12-01T08:41:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:43 crc kubenswrapper[4813]: I1201 08:41:43.088290 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:43 crc kubenswrapper[4813]: I1201 08:41:43.088374 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:43 crc kubenswrapper[4813]: I1201 08:41:43.088390 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:43 crc kubenswrapper[4813]: I1201 08:41:43.088407 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:43 crc kubenswrapper[4813]: I1201 08:41:43.088417 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:43Z","lastTransitionTime":"2025-12-01T08:41:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:43 crc kubenswrapper[4813]: I1201 08:41:43.190226 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:43 crc kubenswrapper[4813]: I1201 08:41:43.190287 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:43 crc kubenswrapper[4813]: I1201 08:41:43.190306 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:43 crc kubenswrapper[4813]: I1201 08:41:43.190329 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:43 crc kubenswrapper[4813]: I1201 08:41:43.190348 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:43Z","lastTransitionTime":"2025-12-01T08:41:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:43 crc kubenswrapper[4813]: I1201 08:41:43.293118 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:43 crc kubenswrapper[4813]: I1201 08:41:43.293152 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:43 crc kubenswrapper[4813]: I1201 08:41:43.293161 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:43 crc kubenswrapper[4813]: I1201 08:41:43.293173 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:43 crc kubenswrapper[4813]: I1201 08:41:43.293181 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:43Z","lastTransitionTime":"2025-12-01T08:41:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:43 crc kubenswrapper[4813]: I1201 08:41:43.393921 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-96mbr" Dec 01 08:41:43 crc kubenswrapper[4813]: I1201 08:41:43.394052 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 08:41:43 crc kubenswrapper[4813]: E1201 08:41:43.394276 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-96mbr" podUID="0789e3f7-5b08-4c04-8596-9ba29f50cdd0" Dec 01 08:41:43 crc kubenswrapper[4813]: E1201 08:41:43.394729 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 08:41:43 crc kubenswrapper[4813]: I1201 08:41:43.397281 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:43 crc kubenswrapper[4813]: I1201 08:41:43.397341 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:43 crc kubenswrapper[4813]: I1201 08:41:43.397368 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:43 crc kubenswrapper[4813]: I1201 08:41:43.397400 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:43 crc kubenswrapper[4813]: I1201 08:41:43.397425 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:43Z","lastTransitionTime":"2025-12-01T08:41:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:43 crc kubenswrapper[4813]: I1201 08:41:43.500361 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:43 crc kubenswrapper[4813]: I1201 08:41:43.500441 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:43 crc kubenswrapper[4813]: I1201 08:41:43.500455 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:43 crc kubenswrapper[4813]: I1201 08:41:43.500500 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:43 crc kubenswrapper[4813]: I1201 08:41:43.500515 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:43Z","lastTransitionTime":"2025-12-01T08:41:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:43 crc kubenswrapper[4813]: I1201 08:41:43.604735 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:43 crc kubenswrapper[4813]: I1201 08:41:43.604933 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:43 crc kubenswrapper[4813]: I1201 08:41:43.605010 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:43 crc kubenswrapper[4813]: I1201 08:41:43.605044 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:43 crc kubenswrapper[4813]: I1201 08:41:43.605062 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:43Z","lastTransitionTime":"2025-12-01T08:41:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:43 crc kubenswrapper[4813]: I1201 08:41:43.706897 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:43 crc kubenswrapper[4813]: I1201 08:41:43.706957 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:43 crc kubenswrapper[4813]: I1201 08:41:43.707008 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:43 crc kubenswrapper[4813]: I1201 08:41:43.707034 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:43 crc kubenswrapper[4813]: I1201 08:41:43.707048 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:43Z","lastTransitionTime":"2025-12-01T08:41:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:43 crc kubenswrapper[4813]: I1201 08:41:43.810132 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:43 crc kubenswrapper[4813]: I1201 08:41:43.810248 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:43 crc kubenswrapper[4813]: I1201 08:41:43.810321 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:43 crc kubenswrapper[4813]: I1201 08:41:43.810358 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:43 crc kubenswrapper[4813]: I1201 08:41:43.810431 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:43Z","lastTransitionTime":"2025-12-01T08:41:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:43 crc kubenswrapper[4813]: I1201 08:41:43.914374 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:43 crc kubenswrapper[4813]: I1201 08:41:43.914420 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:43 crc kubenswrapper[4813]: I1201 08:41:43.914430 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:43 crc kubenswrapper[4813]: I1201 08:41:43.914449 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:43 crc kubenswrapper[4813]: I1201 08:41:43.914460 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:43Z","lastTransitionTime":"2025-12-01T08:41:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:44 crc kubenswrapper[4813]: I1201 08:41:44.017751 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:44 crc kubenswrapper[4813]: I1201 08:41:44.017816 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:44 crc kubenswrapper[4813]: I1201 08:41:44.017833 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:44 crc kubenswrapper[4813]: I1201 08:41:44.017857 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:44 crc kubenswrapper[4813]: I1201 08:41:44.017875 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:44Z","lastTransitionTime":"2025-12-01T08:41:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:44 crc kubenswrapper[4813]: I1201 08:41:44.121018 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:44 crc kubenswrapper[4813]: I1201 08:41:44.121060 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:44 crc kubenswrapper[4813]: I1201 08:41:44.121073 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:44 crc kubenswrapper[4813]: I1201 08:41:44.121093 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:44 crc kubenswrapper[4813]: I1201 08:41:44.121108 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:44Z","lastTransitionTime":"2025-12-01T08:41:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:44 crc kubenswrapper[4813]: I1201 08:41:44.224595 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:44 crc kubenswrapper[4813]: I1201 08:41:44.224667 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:44 crc kubenswrapper[4813]: I1201 08:41:44.224691 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:44 crc kubenswrapper[4813]: I1201 08:41:44.224719 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:44 crc kubenswrapper[4813]: I1201 08:41:44.224740 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:44Z","lastTransitionTime":"2025-12-01T08:41:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:44 crc kubenswrapper[4813]: I1201 08:41:44.328157 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:44 crc kubenswrapper[4813]: I1201 08:41:44.328235 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:44 crc kubenswrapper[4813]: I1201 08:41:44.328254 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:44 crc kubenswrapper[4813]: I1201 08:41:44.328277 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:44 crc kubenswrapper[4813]: I1201 08:41:44.328295 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:44Z","lastTransitionTime":"2025-12-01T08:41:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:44 crc kubenswrapper[4813]: I1201 08:41:44.393132 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 08:41:44 crc kubenswrapper[4813]: I1201 08:41:44.393200 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 08:41:44 crc kubenswrapper[4813]: E1201 08:41:44.393407 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 08:41:44 crc kubenswrapper[4813]: E1201 08:41:44.393594 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 08:41:44 crc kubenswrapper[4813]: I1201 08:41:44.431065 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:44 crc kubenswrapper[4813]: I1201 08:41:44.431130 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:44 crc kubenswrapper[4813]: I1201 08:41:44.431147 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:44 crc kubenswrapper[4813]: I1201 08:41:44.431171 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:44 crc kubenswrapper[4813]: I1201 08:41:44.431188 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:44Z","lastTransitionTime":"2025-12-01T08:41:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:44 crc kubenswrapper[4813]: I1201 08:41:44.534628 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:44 crc kubenswrapper[4813]: I1201 08:41:44.534693 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:44 crc kubenswrapper[4813]: I1201 08:41:44.534709 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:44 crc kubenswrapper[4813]: I1201 08:41:44.534731 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:44 crc kubenswrapper[4813]: I1201 08:41:44.534748 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:44Z","lastTransitionTime":"2025-12-01T08:41:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:44 crc kubenswrapper[4813]: I1201 08:41:44.640070 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:44 crc kubenswrapper[4813]: I1201 08:41:44.640114 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:44 crc kubenswrapper[4813]: I1201 08:41:44.640124 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:44 crc kubenswrapper[4813]: I1201 08:41:44.640145 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:44 crc kubenswrapper[4813]: I1201 08:41:44.640163 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:44Z","lastTransitionTime":"2025-12-01T08:41:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:44 crc kubenswrapper[4813]: I1201 08:41:44.743797 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:44 crc kubenswrapper[4813]: I1201 08:41:44.743849 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:44 crc kubenswrapper[4813]: I1201 08:41:44.743859 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:44 crc kubenswrapper[4813]: I1201 08:41:44.743875 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:44 crc kubenswrapper[4813]: I1201 08:41:44.743886 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:44Z","lastTransitionTime":"2025-12-01T08:41:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:44 crc kubenswrapper[4813]: I1201 08:41:44.846247 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:44 crc kubenswrapper[4813]: I1201 08:41:44.846285 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:44 crc kubenswrapper[4813]: I1201 08:41:44.846297 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:44 crc kubenswrapper[4813]: I1201 08:41:44.846312 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:44 crc kubenswrapper[4813]: I1201 08:41:44.846321 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:44Z","lastTransitionTime":"2025-12-01T08:41:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:44 crc kubenswrapper[4813]: I1201 08:41:44.949134 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:44 crc kubenswrapper[4813]: I1201 08:41:44.949198 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:44 crc kubenswrapper[4813]: I1201 08:41:44.949220 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:44 crc kubenswrapper[4813]: I1201 08:41:44.949246 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:44 crc kubenswrapper[4813]: I1201 08:41:44.949264 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:44Z","lastTransitionTime":"2025-12-01T08:41:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:45 crc kubenswrapper[4813]: I1201 08:41:45.052424 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:45 crc kubenswrapper[4813]: I1201 08:41:45.052502 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:45 crc kubenswrapper[4813]: I1201 08:41:45.052526 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:45 crc kubenswrapper[4813]: I1201 08:41:45.052557 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:45 crc kubenswrapper[4813]: I1201 08:41:45.052580 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:45Z","lastTransitionTime":"2025-12-01T08:41:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:45 crc kubenswrapper[4813]: I1201 08:41:45.155001 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:45 crc kubenswrapper[4813]: I1201 08:41:45.155051 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:45 crc kubenswrapper[4813]: I1201 08:41:45.155063 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:45 crc kubenswrapper[4813]: I1201 08:41:45.155082 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:45 crc kubenswrapper[4813]: I1201 08:41:45.155094 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:45Z","lastTransitionTime":"2025-12-01T08:41:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:45 crc kubenswrapper[4813]: I1201 08:41:45.258154 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:45 crc kubenswrapper[4813]: I1201 08:41:45.258235 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:45 crc kubenswrapper[4813]: I1201 08:41:45.258260 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:45 crc kubenswrapper[4813]: I1201 08:41:45.258293 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:45 crc kubenswrapper[4813]: I1201 08:41:45.258316 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:45Z","lastTransitionTime":"2025-12-01T08:41:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:45 crc kubenswrapper[4813]: I1201 08:41:45.362456 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:45 crc kubenswrapper[4813]: I1201 08:41:45.362529 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:45 crc kubenswrapper[4813]: I1201 08:41:45.362554 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:45 crc kubenswrapper[4813]: I1201 08:41:45.362583 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:45 crc kubenswrapper[4813]: I1201 08:41:45.362605 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:45Z","lastTransitionTime":"2025-12-01T08:41:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:45 crc kubenswrapper[4813]: I1201 08:41:45.393171 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-96mbr" Dec 01 08:41:45 crc kubenswrapper[4813]: I1201 08:41:45.393182 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 08:41:45 crc kubenswrapper[4813]: E1201 08:41:45.393419 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-96mbr" podUID="0789e3f7-5b08-4c04-8596-9ba29f50cdd0" Dec 01 08:41:45 crc kubenswrapper[4813]: E1201 08:41:45.393501 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 08:41:45 crc kubenswrapper[4813]: I1201 08:41:45.469165 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:45 crc kubenswrapper[4813]: I1201 08:41:45.469261 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:45 crc kubenswrapper[4813]: I1201 08:41:45.469286 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:45 crc kubenswrapper[4813]: I1201 08:41:45.469326 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:45 crc kubenswrapper[4813]: I1201 08:41:45.469351 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:45Z","lastTransitionTime":"2025-12-01T08:41:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:45 crc kubenswrapper[4813]: I1201 08:41:45.572702 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:45 crc kubenswrapper[4813]: I1201 08:41:45.572748 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:45 crc kubenswrapper[4813]: I1201 08:41:45.572760 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:45 crc kubenswrapper[4813]: I1201 08:41:45.572778 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:45 crc kubenswrapper[4813]: I1201 08:41:45.572794 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:45Z","lastTransitionTime":"2025-12-01T08:41:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:45 crc kubenswrapper[4813]: I1201 08:41:45.675910 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:45 crc kubenswrapper[4813]: I1201 08:41:45.675943 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:45 crc kubenswrapper[4813]: I1201 08:41:45.675953 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:45 crc kubenswrapper[4813]: I1201 08:41:45.676026 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:45 crc kubenswrapper[4813]: I1201 08:41:45.676046 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:45Z","lastTransitionTime":"2025-12-01T08:41:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:45 crc kubenswrapper[4813]: I1201 08:41:45.780125 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:45 crc kubenswrapper[4813]: I1201 08:41:45.780230 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:45 crc kubenswrapper[4813]: I1201 08:41:45.780261 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:45 crc kubenswrapper[4813]: I1201 08:41:45.780332 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:45 crc kubenswrapper[4813]: I1201 08:41:45.780361 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:45Z","lastTransitionTime":"2025-12-01T08:41:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:45 crc kubenswrapper[4813]: I1201 08:41:45.884707 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:45 crc kubenswrapper[4813]: I1201 08:41:45.884798 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:45 crc kubenswrapper[4813]: I1201 08:41:45.884821 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:45 crc kubenswrapper[4813]: I1201 08:41:45.884855 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:45 crc kubenswrapper[4813]: I1201 08:41:45.884893 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:45Z","lastTransitionTime":"2025-12-01T08:41:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:45 crc kubenswrapper[4813]: I1201 08:41:45.989192 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:45 crc kubenswrapper[4813]: I1201 08:41:45.989268 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:45 crc kubenswrapper[4813]: I1201 08:41:45.989287 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:45 crc kubenswrapper[4813]: I1201 08:41:45.989312 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:45 crc kubenswrapper[4813]: I1201 08:41:45.989330 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:45Z","lastTransitionTime":"2025-12-01T08:41:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:46 crc kubenswrapper[4813]: I1201 08:41:46.092398 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:46 crc kubenswrapper[4813]: I1201 08:41:46.092481 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:46 crc kubenswrapper[4813]: I1201 08:41:46.092518 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:46 crc kubenswrapper[4813]: I1201 08:41:46.092552 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:46 crc kubenswrapper[4813]: I1201 08:41:46.092576 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:46Z","lastTransitionTime":"2025-12-01T08:41:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:46 crc kubenswrapper[4813]: I1201 08:41:46.195855 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:46 crc kubenswrapper[4813]: I1201 08:41:46.195911 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:46 crc kubenswrapper[4813]: I1201 08:41:46.195928 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:46 crc kubenswrapper[4813]: I1201 08:41:46.195951 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:46 crc kubenswrapper[4813]: I1201 08:41:46.196025 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:46Z","lastTransitionTime":"2025-12-01T08:41:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:46 crc kubenswrapper[4813]: I1201 08:41:46.299320 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:46 crc kubenswrapper[4813]: I1201 08:41:46.299385 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:46 crc kubenswrapper[4813]: I1201 08:41:46.299402 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:46 crc kubenswrapper[4813]: I1201 08:41:46.299425 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:46 crc kubenswrapper[4813]: I1201 08:41:46.299440 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:46Z","lastTransitionTime":"2025-12-01T08:41:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:46 crc kubenswrapper[4813]: I1201 08:41:46.394732 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 08:41:46 crc kubenswrapper[4813]: I1201 08:41:46.394728 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 08:41:46 crc kubenswrapper[4813]: E1201 08:41:46.395776 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 08:41:46 crc kubenswrapper[4813]: E1201 08:41:46.396742 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 08:41:46 crc kubenswrapper[4813]: I1201 08:41:46.399207 4813 scope.go:117] "RemoveContainer" containerID="fb17fb9ae11fa62235328567418fb259f4f8540ebac635842d3afe751dcbde8c" Dec 01 08:41:46 crc kubenswrapper[4813]: I1201 08:41:46.409521 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:46 crc kubenswrapper[4813]: I1201 08:41:46.409566 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:46 crc kubenswrapper[4813]: I1201 08:41:46.409579 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:46 crc kubenswrapper[4813]: I1201 08:41:46.409598 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:46 crc kubenswrapper[4813]: I1201 08:41:46.409607 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:46Z","lastTransitionTime":"2025-12-01T08:41:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:46 crc kubenswrapper[4813]: I1201 08:41:46.420349 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"490e526a-902b-4aff-b547-0cefea562764\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb48bd28fd3d29781695a5ae630f4913ef2745e97fbe7d463cca39a31e43ba0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74f8cf73f07b4cde260af429d26413e4e6226e7b177b493dd9c0e5c043ab6e1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3aaf696cefbe398bb6b26acfc5be43bda7e2be18218c8060840350c4106e66a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53de1f0682f36aa83ad8e5c226b95daf9f97df4ae154ebfc2fc287f9f238a4a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53de1f0682f36aa83ad8e5c226b95daf9f97df4ae154ebfc2fc287f9f238a4a8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T08:40:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T08:40:17Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:16Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:46Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:46 crc kubenswrapper[4813]: I1201 08:41:46.430393 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"] Dec 01 08:41:46 crc kubenswrapper[4813]: I1201 08:41:46.454031 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:46Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:46 crc kubenswrapper[4813]: I1201 08:41:46.470794 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-c576v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"17292cd0-1eac-40b5-a60e-67167c57409c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T08:40:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a42206c914d0493ca74788ea1a941de80920ab7b810fb95227cd9b8711e8f4b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T08:40:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7zwfx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T08:40:41Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-c576v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T08:41:46Z is after 2025-08-24T17:21:41Z" Dec 01 08:41:46 crc kubenswrapper[4813]: I1201 08:41:46.512171 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:46 crc kubenswrapper[4813]: I1201 08:41:46.512224 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:46 crc kubenswrapper[4813]: I1201 08:41:46.512238 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:46 crc kubenswrapper[4813]: I1201 08:41:46.512257 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:46 crc kubenswrapper[4813]: I1201 08:41:46.512269 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:46Z","lastTransitionTime":"2025-12-01T08:41:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:46 crc kubenswrapper[4813]: I1201 08:41:46.570225 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-xpnlc" podStartSLOduration=66.570092619 podStartE2EDuration="1m6.570092619s" podCreationTimestamp="2025-12-01 08:40:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:41:46.569437987 +0000 UTC m=+91.192359583" watchObservedRunningTime="2025-12-01 08:41:46.570092619 +0000 UTC m=+91.193014225" Dec 01 08:41:46 crc kubenswrapper[4813]: I1201 08:41:46.572227 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-h7lvd" podStartSLOduration=66.572196449 podStartE2EDuration="1m6.572196449s" podCreationTimestamp="2025-12-01 08:40:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:41:46.554958578 +0000 UTC m=+91.177880184" watchObservedRunningTime="2025-12-01 08:41:46.572196449 +0000 UTC m=+91.195118055" Dec 01 08:41:46 crc kubenswrapper[4813]: I1201 08:41:46.609266 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=65.60921867 podStartE2EDuration="1m5.60921867s" podCreationTimestamp="2025-12-01 08:40:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:41:46.59575829 +0000 UTC m=+91.218679906" watchObservedRunningTime="2025-12-01 08:41:46.60921867 +0000 UTC m=+91.232140276" Dec 01 08:41:46 crc kubenswrapper[4813]: I1201 08:41:46.615386 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:46 crc kubenswrapper[4813]: I1201 08:41:46.615432 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:46 crc kubenswrapper[4813]: I1201 08:41:46.615441 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:46 crc kubenswrapper[4813]: I1201 08:41:46.615457 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:46 crc kubenswrapper[4813]: I1201 08:41:46.615466 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:46Z","lastTransitionTime":"2025-12-01T08:41:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:46 crc kubenswrapper[4813]: I1201 08:41:46.677045 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=12.677026666 podStartE2EDuration="12.677026666s" podCreationTimestamp="2025-12-01 08:41:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:41:46.676393996 +0000 UTC m=+91.299315622" watchObservedRunningTime="2025-12-01 08:41:46.677026666 +0000 UTC m=+91.299948252" Dec 01 08:41:46 crc kubenswrapper[4813]: I1201 08:41:46.707367 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=68.707350869 podStartE2EDuration="1m8.707350869s" podCreationTimestamp="2025-12-01 08:40:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:41:46.692752184 +0000 UTC m=+91.315673770" watchObservedRunningTime="2025-12-01 08:41:46.707350869 +0000 UTC m=+91.330272455" Dec 01 08:41:46 crc kubenswrapper[4813]: I1201 08:41:46.718270 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:46 crc kubenswrapper[4813]: I1201 08:41:46.718312 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:46 crc kubenswrapper[4813]: I1201 08:41:46.718323 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:46 crc kubenswrapper[4813]: I1201 08:41:46.718335 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:46 crc kubenswrapper[4813]: I1201 08:41:46.718344 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:46Z","lastTransitionTime":"2025-12-01T08:41:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:46 crc kubenswrapper[4813]: I1201 08:41:46.733620 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-b6qqv" podStartSLOduration=66.733606568 podStartE2EDuration="1m6.733606568s" podCreationTimestamp="2025-12-01 08:40:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:41:46.721233519 +0000 UTC m=+91.344155135" watchObservedRunningTime="2025-12-01 08:41:46.733606568 +0000 UTC m=+91.356528154" Dec 01 08:41:46 crc kubenswrapper[4813]: I1201 08:41:46.747125 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-9wn4q" podStartSLOduration=66.747091749 podStartE2EDuration="1m6.747091749s" podCreationTimestamp="2025-12-01 08:40:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:41:46.746714421 +0000 UTC m=+91.369636017" watchObservedRunningTime="2025-12-01 08:41:46.747091749 +0000 UTC m=+91.370013375" Dec 01 08:41:46 crc kubenswrapper[4813]: I1201 08:41:46.761013 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-8pg2p" podStartSLOduration=66.76098655 podStartE2EDuration="1m6.76098655s" podCreationTimestamp="2025-12-01 08:40:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:41:46.760781971 +0000 UTC m=+91.383703567" watchObservedRunningTime="2025-12-01 08:41:46.76098655 +0000 UTC m=+91.383908146" Dec 01 08:41:46 crc kubenswrapper[4813]: I1201 08:41:46.820549 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:46 crc kubenswrapper[4813]: I1201 08:41:46.820579 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:46 crc kubenswrapper[4813]: I1201 08:41:46.820606 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:46 crc kubenswrapper[4813]: I1201 08:41:46.820620 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:46 crc kubenswrapper[4813]: I1201 08:41:46.820630 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:46Z","lastTransitionTime":"2025-12-01T08:41:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:46 crc kubenswrapper[4813]: I1201 08:41:46.923567 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:46 crc kubenswrapper[4813]: I1201 08:41:46.923613 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:46 crc kubenswrapper[4813]: I1201 08:41:46.923632 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:46 crc kubenswrapper[4813]: I1201 08:41:46.923656 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:46 crc kubenswrapper[4813]: I1201 08:41:46.923674 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:46Z","lastTransitionTime":"2025-12-01T08:41:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:47 crc kubenswrapper[4813]: I1201 08:41:47.026720 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:47 crc kubenswrapper[4813]: I1201 08:41:47.026794 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:47 crc kubenswrapper[4813]: I1201 08:41:47.026812 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:47 crc kubenswrapper[4813]: I1201 08:41:47.026838 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:47 crc kubenswrapper[4813]: I1201 08:41:47.026879 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:47Z","lastTransitionTime":"2025-12-01T08:41:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:47 crc kubenswrapper[4813]: I1201 08:41:47.130015 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:47 crc kubenswrapper[4813]: I1201 08:41:47.130079 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:47 crc kubenswrapper[4813]: I1201 08:41:47.130099 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:47 crc kubenswrapper[4813]: I1201 08:41:47.130128 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:47 crc kubenswrapper[4813]: I1201 08:41:47.130155 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:47Z","lastTransitionTime":"2025-12-01T08:41:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:47 crc kubenswrapper[4813]: I1201 08:41:47.235501 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:47 crc kubenswrapper[4813]: I1201 08:41:47.235550 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:47 crc kubenswrapper[4813]: I1201 08:41:47.235564 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:47 crc kubenswrapper[4813]: I1201 08:41:47.235585 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:47 crc kubenswrapper[4813]: I1201 08:41:47.235601 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:47Z","lastTransitionTime":"2025-12-01T08:41:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:47 crc kubenswrapper[4813]: I1201 08:41:47.338564 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:47 crc kubenswrapper[4813]: I1201 08:41:47.338784 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:47 crc kubenswrapper[4813]: I1201 08:41:47.338795 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:47 crc kubenswrapper[4813]: I1201 08:41:47.338810 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:47 crc kubenswrapper[4813]: I1201 08:41:47.338820 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:47Z","lastTransitionTime":"2025-12-01T08:41:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:47 crc kubenswrapper[4813]: I1201 08:41:47.393616 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 08:41:47 crc kubenswrapper[4813]: I1201 08:41:47.393753 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-96mbr" Dec 01 08:41:47 crc kubenswrapper[4813]: E1201 08:41:47.394033 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 08:41:47 crc kubenswrapper[4813]: E1201 08:41:47.394083 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-96mbr" podUID="0789e3f7-5b08-4c04-8596-9ba29f50cdd0" Dec 01 08:41:47 crc kubenswrapper[4813]: I1201 08:41:47.442693 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:47 crc kubenswrapper[4813]: I1201 08:41:47.442734 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:47 crc kubenswrapper[4813]: I1201 08:41:47.442743 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:47 crc kubenswrapper[4813]: I1201 08:41:47.442759 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:47 crc kubenswrapper[4813]: I1201 08:41:47.442768 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:47Z","lastTransitionTime":"2025-12-01T08:41:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:47 crc kubenswrapper[4813]: I1201 08:41:47.549087 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:47 crc kubenswrapper[4813]: I1201 08:41:47.549140 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:47 crc kubenswrapper[4813]: I1201 08:41:47.549152 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:47 crc kubenswrapper[4813]: I1201 08:41:47.549169 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:47 crc kubenswrapper[4813]: I1201 08:41:47.549183 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:47Z","lastTransitionTime":"2025-12-01T08:41:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:47 crc kubenswrapper[4813]: I1201 08:41:47.651122 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:47 crc kubenswrapper[4813]: I1201 08:41:47.651176 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:47 crc kubenswrapper[4813]: I1201 08:41:47.651185 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:47 crc kubenswrapper[4813]: I1201 08:41:47.651200 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:47 crc kubenswrapper[4813]: I1201 08:41:47.651209 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:47Z","lastTransitionTime":"2025-12-01T08:41:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:47 crc kubenswrapper[4813]: I1201 08:41:47.754025 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:47 crc kubenswrapper[4813]: I1201 08:41:47.754057 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:47 crc kubenswrapper[4813]: I1201 08:41:47.754064 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:47 crc kubenswrapper[4813]: I1201 08:41:47.754077 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:47 crc kubenswrapper[4813]: I1201 08:41:47.754086 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:47Z","lastTransitionTime":"2025-12-01T08:41:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:47 crc kubenswrapper[4813]: I1201 08:41:47.800477 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-j4pcc_b7505e14-3560-4d0d-be2c-d6a7a5953865/ovnkube-controller/2.log" Dec 01 08:41:47 crc kubenswrapper[4813]: I1201 08:41:47.803613 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-j4pcc" event={"ID":"b7505e14-3560-4d0d-be2c-d6a7a5953865","Type":"ContainerStarted","Data":"6aa74106386539915851c80789dac37ef865908f24cb6a67ffa406c2599cb5d7"} Dec 01 08:41:47 crc kubenswrapper[4813]: I1201 08:41:47.804027 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-j4pcc" Dec 01 08:41:47 crc kubenswrapper[4813]: I1201 08:41:47.856370 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=1.856348444 podStartE2EDuration="1.856348444s" podCreationTimestamp="2025-12-01 08:41:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:41:47.855327265 +0000 UTC m=+92.478248851" watchObservedRunningTime="2025-12-01 08:41:47.856348444 +0000 UTC m=+92.479270040" Dec 01 08:41:47 crc kubenswrapper[4813]: I1201 08:41:47.857635 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:47 crc kubenswrapper[4813]: I1201 08:41:47.857671 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:47 crc kubenswrapper[4813]: I1201 08:41:47.857680 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:47 crc kubenswrapper[4813]: I1201 08:41:47.857693 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:47 crc kubenswrapper[4813]: I1201 08:41:47.857703 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:47Z","lastTransitionTime":"2025-12-01T08:41:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:47 crc kubenswrapper[4813]: I1201 08:41:47.891681 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-j4pcc" podStartSLOduration=67.891661703 podStartE2EDuration="1m7.891661703s" podCreationTimestamp="2025-12-01 08:40:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:41:47.891359179 +0000 UTC m=+92.514280765" watchObservedRunningTime="2025-12-01 08:41:47.891661703 +0000 UTC m=+92.514583289" Dec 01 08:41:47 crc kubenswrapper[4813]: I1201 08:41:47.908019 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=43.90799661 podStartE2EDuration="43.90799661s" podCreationTimestamp="2025-12-01 08:41:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:41:47.906617615 +0000 UTC m=+92.529539201" watchObservedRunningTime="2025-12-01 08:41:47.90799661 +0000 UTC m=+92.530918196" Dec 01 08:41:47 crc kubenswrapper[4813]: I1201 08:41:47.959627 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:47 crc kubenswrapper[4813]: I1201 08:41:47.959661 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:47 crc kubenswrapper[4813]: I1201 08:41:47.959670 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:47 crc kubenswrapper[4813]: I1201 08:41:47.959684 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:47 crc kubenswrapper[4813]: I1201 08:41:47.959692 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:47Z","lastTransitionTime":"2025-12-01T08:41:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:48 crc kubenswrapper[4813]: I1201 08:41:48.062414 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:48 crc kubenswrapper[4813]: I1201 08:41:48.062455 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:48 crc kubenswrapper[4813]: I1201 08:41:48.062470 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:48 crc kubenswrapper[4813]: I1201 08:41:48.062488 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:48 crc kubenswrapper[4813]: I1201 08:41:48.062501 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:48Z","lastTransitionTime":"2025-12-01T08:41:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:48 crc kubenswrapper[4813]: I1201 08:41:48.165082 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:48 crc kubenswrapper[4813]: I1201 08:41:48.165132 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:48 crc kubenswrapper[4813]: I1201 08:41:48.165148 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:48 crc kubenswrapper[4813]: I1201 08:41:48.165171 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:48 crc kubenswrapper[4813]: I1201 08:41:48.165188 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:48Z","lastTransitionTime":"2025-12-01T08:41:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:48 crc kubenswrapper[4813]: I1201 08:41:48.267205 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:48 crc kubenswrapper[4813]: I1201 08:41:48.267243 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:48 crc kubenswrapper[4813]: I1201 08:41:48.267251 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:48 crc kubenswrapper[4813]: I1201 08:41:48.267270 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:48 crc kubenswrapper[4813]: I1201 08:41:48.267279 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:48Z","lastTransitionTime":"2025-12-01T08:41:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:48 crc kubenswrapper[4813]: I1201 08:41:48.369541 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:48 crc kubenswrapper[4813]: I1201 08:41:48.369593 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:48 crc kubenswrapper[4813]: I1201 08:41:48.369602 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:48 crc kubenswrapper[4813]: I1201 08:41:48.369616 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:48 crc kubenswrapper[4813]: I1201 08:41:48.369625 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:48Z","lastTransitionTime":"2025-12-01T08:41:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:48 crc kubenswrapper[4813]: I1201 08:41:48.393066 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 08:41:48 crc kubenswrapper[4813]: I1201 08:41:48.393113 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 08:41:48 crc kubenswrapper[4813]: E1201 08:41:48.393255 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 08:41:48 crc kubenswrapper[4813]: E1201 08:41:48.393416 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 08:41:48 crc kubenswrapper[4813]: I1201 08:41:48.472373 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:48 crc kubenswrapper[4813]: I1201 08:41:48.472411 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:48 crc kubenswrapper[4813]: I1201 08:41:48.472419 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:48 crc kubenswrapper[4813]: I1201 08:41:48.472435 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:48 crc kubenswrapper[4813]: I1201 08:41:48.472445 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:48Z","lastTransitionTime":"2025-12-01T08:41:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:48 crc kubenswrapper[4813]: I1201 08:41:48.574605 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:48 crc kubenswrapper[4813]: I1201 08:41:48.574664 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:48 crc kubenswrapper[4813]: I1201 08:41:48.574676 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:48 crc kubenswrapper[4813]: I1201 08:41:48.574694 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:48 crc kubenswrapper[4813]: I1201 08:41:48.574708 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:48Z","lastTransitionTime":"2025-12-01T08:41:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:48 crc kubenswrapper[4813]: I1201 08:41:48.664636 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-c576v" podStartSLOduration=68.664609834 podStartE2EDuration="1m8.664609834s" podCreationTimestamp="2025-12-01 08:40:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:41:47.935157462 +0000 UTC m=+92.558079048" watchObservedRunningTime="2025-12-01 08:41:48.664609834 +0000 UTC m=+93.287531440" Dec 01 08:41:48 crc kubenswrapper[4813]: I1201 08:41:48.665309 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-96mbr"] Dec 01 08:41:48 crc kubenswrapper[4813]: I1201 08:41:48.665438 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-96mbr" Dec 01 08:41:48 crc kubenswrapper[4813]: E1201 08:41:48.665553 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-96mbr" podUID="0789e3f7-5b08-4c04-8596-9ba29f50cdd0" Dec 01 08:41:48 crc kubenswrapper[4813]: I1201 08:41:48.678124 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:48 crc kubenswrapper[4813]: I1201 08:41:48.678178 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:48 crc kubenswrapper[4813]: I1201 08:41:48.678191 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:48 crc kubenswrapper[4813]: I1201 08:41:48.678210 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:48 crc kubenswrapper[4813]: I1201 08:41:48.678223 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:48Z","lastTransitionTime":"2025-12-01T08:41:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:48 crc kubenswrapper[4813]: I1201 08:41:48.780465 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:48 crc kubenswrapper[4813]: I1201 08:41:48.780501 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:48 crc kubenswrapper[4813]: I1201 08:41:48.780510 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:48 crc kubenswrapper[4813]: I1201 08:41:48.780523 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:48 crc kubenswrapper[4813]: I1201 08:41:48.780532 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:48Z","lastTransitionTime":"2025-12-01T08:41:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:48 crc kubenswrapper[4813]: I1201 08:41:48.882621 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:48 crc kubenswrapper[4813]: I1201 08:41:48.882649 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:48 crc kubenswrapper[4813]: I1201 08:41:48.882657 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:48 crc kubenswrapper[4813]: I1201 08:41:48.882670 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:48 crc kubenswrapper[4813]: I1201 08:41:48.882678 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:48Z","lastTransitionTime":"2025-12-01T08:41:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:48 crc kubenswrapper[4813]: I1201 08:41:48.984896 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:48 crc kubenswrapper[4813]: I1201 08:41:48.984941 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:48 crc kubenswrapper[4813]: I1201 08:41:48.984954 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:48 crc kubenswrapper[4813]: I1201 08:41:48.984984 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:48 crc kubenswrapper[4813]: I1201 08:41:48.984997 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:48Z","lastTransitionTime":"2025-12-01T08:41:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:49 crc kubenswrapper[4813]: I1201 08:41:49.087608 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:49 crc kubenswrapper[4813]: I1201 08:41:49.087655 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:49 crc kubenswrapper[4813]: I1201 08:41:49.087667 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:49 crc kubenswrapper[4813]: I1201 08:41:49.087685 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:49 crc kubenswrapper[4813]: I1201 08:41:49.087698 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:49Z","lastTransitionTime":"2025-12-01T08:41:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:49 crc kubenswrapper[4813]: I1201 08:41:49.198527 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:49 crc kubenswrapper[4813]: I1201 08:41:49.198580 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:49 crc kubenswrapper[4813]: I1201 08:41:49.198591 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:49 crc kubenswrapper[4813]: I1201 08:41:49.198610 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:49 crc kubenswrapper[4813]: I1201 08:41:49.198621 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:49Z","lastTransitionTime":"2025-12-01T08:41:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:49 crc kubenswrapper[4813]: I1201 08:41:49.302134 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:49 crc kubenswrapper[4813]: I1201 08:41:49.302196 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:49 crc kubenswrapper[4813]: I1201 08:41:49.302209 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:49 crc kubenswrapper[4813]: I1201 08:41:49.302228 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:49 crc kubenswrapper[4813]: I1201 08:41:49.302241 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:49Z","lastTransitionTime":"2025-12-01T08:41:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:49 crc kubenswrapper[4813]: I1201 08:41:49.392639 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 08:41:49 crc kubenswrapper[4813]: E1201 08:41:49.392793 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 08:41:49 crc kubenswrapper[4813]: I1201 08:41:49.405396 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:49 crc kubenswrapper[4813]: I1201 08:41:49.405429 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:49 crc kubenswrapper[4813]: I1201 08:41:49.405440 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:49 crc kubenswrapper[4813]: I1201 08:41:49.405482 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:49 crc kubenswrapper[4813]: I1201 08:41:49.405492 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:49Z","lastTransitionTime":"2025-12-01T08:41:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:49 crc kubenswrapper[4813]: I1201 08:41:49.508152 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:49 crc kubenswrapper[4813]: I1201 08:41:49.508208 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:49 crc kubenswrapper[4813]: I1201 08:41:49.508221 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:49 crc kubenswrapper[4813]: I1201 08:41:49.508240 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:49 crc kubenswrapper[4813]: I1201 08:41:49.508251 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:49Z","lastTransitionTime":"2025-12-01T08:41:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:49 crc kubenswrapper[4813]: I1201 08:41:49.610551 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:49 crc kubenswrapper[4813]: I1201 08:41:49.610583 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:49 crc kubenswrapper[4813]: I1201 08:41:49.610591 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:49 crc kubenswrapper[4813]: I1201 08:41:49.610607 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:49 crc kubenswrapper[4813]: I1201 08:41:49.610617 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:49Z","lastTransitionTime":"2025-12-01T08:41:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:49 crc kubenswrapper[4813]: I1201 08:41:49.713601 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:49 crc kubenswrapper[4813]: I1201 08:41:49.713653 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:49 crc kubenswrapper[4813]: I1201 08:41:49.713669 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:49 crc kubenswrapper[4813]: I1201 08:41:49.713693 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:49 crc kubenswrapper[4813]: I1201 08:41:49.713711 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:49Z","lastTransitionTime":"2025-12-01T08:41:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:49 crc kubenswrapper[4813]: I1201 08:41:49.817282 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:49 crc kubenswrapper[4813]: I1201 08:41:49.817394 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:49 crc kubenswrapper[4813]: I1201 08:41:49.817428 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:49 crc kubenswrapper[4813]: I1201 08:41:49.817457 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:49 crc kubenswrapper[4813]: I1201 08:41:49.817478 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:49Z","lastTransitionTime":"2025-12-01T08:41:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:49 crc kubenswrapper[4813]: I1201 08:41:49.920064 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:49 crc kubenswrapper[4813]: I1201 08:41:49.920124 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:49 crc kubenswrapper[4813]: I1201 08:41:49.920137 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:49 crc kubenswrapper[4813]: I1201 08:41:49.920154 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:49 crc kubenswrapper[4813]: I1201 08:41:49.920167 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:49Z","lastTransitionTime":"2025-12-01T08:41:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:50 crc kubenswrapper[4813]: I1201 08:41:50.023151 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:50 crc kubenswrapper[4813]: I1201 08:41:50.023212 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:50 crc kubenswrapper[4813]: I1201 08:41:50.023234 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:50 crc kubenswrapper[4813]: I1201 08:41:50.023263 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:50 crc kubenswrapper[4813]: I1201 08:41:50.023284 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:50Z","lastTransitionTime":"2025-12-01T08:41:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:50 crc kubenswrapper[4813]: I1201 08:41:50.126144 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:50 crc kubenswrapper[4813]: I1201 08:41:50.126218 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:50 crc kubenswrapper[4813]: I1201 08:41:50.126236 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:50 crc kubenswrapper[4813]: I1201 08:41:50.126639 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:50 crc kubenswrapper[4813]: I1201 08:41:50.126688 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:50Z","lastTransitionTime":"2025-12-01T08:41:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:50 crc kubenswrapper[4813]: I1201 08:41:50.230427 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:50 crc kubenswrapper[4813]: I1201 08:41:50.230486 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:50 crc kubenswrapper[4813]: I1201 08:41:50.230494 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:50 crc kubenswrapper[4813]: I1201 08:41:50.230509 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:50 crc kubenswrapper[4813]: I1201 08:41:50.230518 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:50Z","lastTransitionTime":"2025-12-01T08:41:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:50 crc kubenswrapper[4813]: I1201 08:41:50.334170 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:50 crc kubenswrapper[4813]: I1201 08:41:50.334223 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:50 crc kubenswrapper[4813]: I1201 08:41:50.334239 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:50 crc kubenswrapper[4813]: I1201 08:41:50.334266 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:50 crc kubenswrapper[4813]: I1201 08:41:50.334282 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:50Z","lastTransitionTime":"2025-12-01T08:41:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:50 crc kubenswrapper[4813]: I1201 08:41:50.393560 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 08:41:50 crc kubenswrapper[4813]: I1201 08:41:50.393657 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 08:41:50 crc kubenswrapper[4813]: E1201 08:41:50.393856 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 08:41:50 crc kubenswrapper[4813]: I1201 08:41:50.394003 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-96mbr" Dec 01 08:41:50 crc kubenswrapper[4813]: E1201 08:41:50.394227 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 08:41:50 crc kubenswrapper[4813]: E1201 08:41:50.394502 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-96mbr" podUID="0789e3f7-5b08-4c04-8596-9ba29f50cdd0" Dec 01 08:41:50 crc kubenswrapper[4813]: I1201 08:41:50.437228 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:50 crc kubenswrapper[4813]: I1201 08:41:50.437273 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:50 crc kubenswrapper[4813]: I1201 08:41:50.437283 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:50 crc kubenswrapper[4813]: I1201 08:41:50.437299 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:50 crc kubenswrapper[4813]: I1201 08:41:50.437309 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:50Z","lastTransitionTime":"2025-12-01T08:41:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:50 crc kubenswrapper[4813]: I1201 08:41:50.543420 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:50 crc kubenswrapper[4813]: I1201 08:41:50.543466 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:50 crc kubenswrapper[4813]: I1201 08:41:50.543484 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:50 crc kubenswrapper[4813]: I1201 08:41:50.543507 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:50 crc kubenswrapper[4813]: I1201 08:41:50.543566 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:50Z","lastTransitionTime":"2025-12-01T08:41:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:50 crc kubenswrapper[4813]: I1201 08:41:50.646570 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:50 crc kubenswrapper[4813]: I1201 08:41:50.646621 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:50 crc kubenswrapper[4813]: I1201 08:41:50.646640 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:50 crc kubenswrapper[4813]: I1201 08:41:50.646665 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:50 crc kubenswrapper[4813]: I1201 08:41:50.646689 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:50Z","lastTransitionTime":"2025-12-01T08:41:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:50 crc kubenswrapper[4813]: I1201 08:41:50.749678 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:50 crc kubenswrapper[4813]: I1201 08:41:50.749730 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:50 crc kubenswrapper[4813]: I1201 08:41:50.749746 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:50 crc kubenswrapper[4813]: I1201 08:41:50.749772 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:50 crc kubenswrapper[4813]: I1201 08:41:50.749790 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:50Z","lastTransitionTime":"2025-12-01T08:41:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:50 crc kubenswrapper[4813]: I1201 08:41:50.852411 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:50 crc kubenswrapper[4813]: I1201 08:41:50.852474 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:50 crc kubenswrapper[4813]: I1201 08:41:50.852492 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:50 crc kubenswrapper[4813]: I1201 08:41:50.852517 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:50 crc kubenswrapper[4813]: I1201 08:41:50.852534 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:50Z","lastTransitionTime":"2025-12-01T08:41:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:50 crc kubenswrapper[4813]: I1201 08:41:50.956598 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:50 crc kubenswrapper[4813]: I1201 08:41:50.956646 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:50 crc kubenswrapper[4813]: I1201 08:41:50.956655 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:50 crc kubenswrapper[4813]: I1201 08:41:50.956676 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:50 crc kubenswrapper[4813]: I1201 08:41:50.956685 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:50Z","lastTransitionTime":"2025-12-01T08:41:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:51 crc kubenswrapper[4813]: I1201 08:41:51.060503 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:51 crc kubenswrapper[4813]: I1201 08:41:51.060572 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:51 crc kubenswrapper[4813]: I1201 08:41:51.060595 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:51 crc kubenswrapper[4813]: I1201 08:41:51.060627 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:51 crc kubenswrapper[4813]: I1201 08:41:51.060652 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:51Z","lastTransitionTime":"2025-12-01T08:41:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:51 crc kubenswrapper[4813]: I1201 08:41:51.163879 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:51 crc kubenswrapper[4813]: I1201 08:41:51.163993 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:51 crc kubenswrapper[4813]: I1201 08:41:51.164009 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:51 crc kubenswrapper[4813]: I1201 08:41:51.164029 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:51 crc kubenswrapper[4813]: I1201 08:41:51.164041 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:51Z","lastTransitionTime":"2025-12-01T08:41:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:51 crc kubenswrapper[4813]: I1201 08:41:51.267406 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:51 crc kubenswrapper[4813]: I1201 08:41:51.267484 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:51 crc kubenswrapper[4813]: I1201 08:41:51.267508 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:51 crc kubenswrapper[4813]: I1201 08:41:51.267533 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:51 crc kubenswrapper[4813]: I1201 08:41:51.267553 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:51Z","lastTransitionTime":"2025-12-01T08:41:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:51 crc kubenswrapper[4813]: I1201 08:41:51.370807 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:51 crc kubenswrapper[4813]: I1201 08:41:51.370874 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:51 crc kubenswrapper[4813]: I1201 08:41:51.370891 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:51 crc kubenswrapper[4813]: I1201 08:41:51.370923 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:51 crc kubenswrapper[4813]: I1201 08:41:51.370946 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:51Z","lastTransitionTime":"2025-12-01T08:41:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:51 crc kubenswrapper[4813]: I1201 08:41:51.393481 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 08:41:51 crc kubenswrapper[4813]: E1201 08:41:51.393691 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 08:41:51 crc kubenswrapper[4813]: I1201 08:41:51.474268 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:51 crc kubenswrapper[4813]: I1201 08:41:51.474349 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:51 crc kubenswrapper[4813]: I1201 08:41:51.474372 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:51 crc kubenswrapper[4813]: I1201 08:41:51.474403 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:51 crc kubenswrapper[4813]: I1201 08:41:51.474434 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:51Z","lastTransitionTime":"2025-12-01T08:41:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:51 crc kubenswrapper[4813]: I1201 08:41:51.578827 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:51 crc kubenswrapper[4813]: I1201 08:41:51.578916 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:51 crc kubenswrapper[4813]: I1201 08:41:51.578940 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:51 crc kubenswrapper[4813]: I1201 08:41:51.579001 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:51 crc kubenswrapper[4813]: I1201 08:41:51.579025 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:51Z","lastTransitionTime":"2025-12-01T08:41:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:51 crc kubenswrapper[4813]: I1201 08:41:51.675726 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 08:41:51 crc kubenswrapper[4813]: I1201 08:41:51.676374 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 08:41:51 crc kubenswrapper[4813]: I1201 08:41:51.676418 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 08:41:51 crc kubenswrapper[4813]: I1201 08:41:51.676445 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 08:41:51 crc kubenswrapper[4813]: I1201 08:41:51.676463 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T08:41:51Z","lastTransitionTime":"2025-12-01T08:41:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 08:41:51 crc kubenswrapper[4813]: I1201 08:41:51.738160 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-xs4nq"] Dec 01 08:41:51 crc kubenswrapper[4813]: I1201 08:41:51.738729 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xs4nq" Dec 01 08:41:51 crc kubenswrapper[4813]: I1201 08:41:51.740853 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Dec 01 08:41:51 crc kubenswrapper[4813]: I1201 08:41:51.745988 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Dec 01 08:41:51 crc kubenswrapper[4813]: I1201 08:41:51.746160 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Dec 01 08:41:51 crc kubenswrapper[4813]: I1201 08:41:51.746299 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Dec 01 08:41:51 crc kubenswrapper[4813]: I1201 08:41:51.782820 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e70d6d2c-f0f3-4c8f-8213-efaa11de813b-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-xs4nq\" (UID: \"e70d6d2c-f0f3-4c8f-8213-efaa11de813b\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xs4nq" Dec 01 08:41:51 crc kubenswrapper[4813]: I1201 08:41:51.782888 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/e70d6d2c-f0f3-4c8f-8213-efaa11de813b-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-xs4nq\" (UID: \"e70d6d2c-f0f3-4c8f-8213-efaa11de813b\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xs4nq" Dec 01 08:41:51 crc kubenswrapper[4813]: I1201 08:41:51.782922 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/e70d6d2c-f0f3-4c8f-8213-efaa11de813b-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-xs4nq\" (UID: \"e70d6d2c-f0f3-4c8f-8213-efaa11de813b\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xs4nq" Dec 01 08:41:51 crc kubenswrapper[4813]: I1201 08:41:51.782990 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/e70d6d2c-f0f3-4c8f-8213-efaa11de813b-service-ca\") pod \"cluster-version-operator-5c965bbfc6-xs4nq\" (UID: \"e70d6d2c-f0f3-4c8f-8213-efaa11de813b\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xs4nq" Dec 01 08:41:51 crc kubenswrapper[4813]: I1201 08:41:51.783051 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e70d6d2c-f0f3-4c8f-8213-efaa11de813b-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-xs4nq\" (UID: \"e70d6d2c-f0f3-4c8f-8213-efaa11de813b\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xs4nq" Dec 01 08:41:51 crc kubenswrapper[4813]: I1201 08:41:51.884539 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e70d6d2c-f0f3-4c8f-8213-efaa11de813b-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-xs4nq\" (UID: \"e70d6d2c-f0f3-4c8f-8213-efaa11de813b\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xs4nq" Dec 01 08:41:51 crc kubenswrapper[4813]: I1201 08:41:51.884639 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/e70d6d2c-f0f3-4c8f-8213-efaa11de813b-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-xs4nq\" (UID: \"e70d6d2c-f0f3-4c8f-8213-efaa11de813b\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xs4nq" Dec 01 08:41:51 crc kubenswrapper[4813]: I1201 08:41:51.884674 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/e70d6d2c-f0f3-4c8f-8213-efaa11de813b-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-xs4nq\" (UID: \"e70d6d2c-f0f3-4c8f-8213-efaa11de813b\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xs4nq" Dec 01 08:41:51 crc kubenswrapper[4813]: I1201 08:41:51.884718 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/e70d6d2c-f0f3-4c8f-8213-efaa11de813b-service-ca\") pod \"cluster-version-operator-5c965bbfc6-xs4nq\" (UID: \"e70d6d2c-f0f3-4c8f-8213-efaa11de813b\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xs4nq" Dec 01 08:41:51 crc kubenswrapper[4813]: I1201 08:41:51.884814 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e70d6d2c-f0f3-4c8f-8213-efaa11de813b-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-xs4nq\" (UID: \"e70d6d2c-f0f3-4c8f-8213-efaa11de813b\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xs4nq" Dec 01 08:41:51 crc kubenswrapper[4813]: I1201 08:41:51.884841 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/e70d6d2c-f0f3-4c8f-8213-efaa11de813b-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-xs4nq\" (UID: \"e70d6d2c-f0f3-4c8f-8213-efaa11de813b\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xs4nq" Dec 01 08:41:51 crc kubenswrapper[4813]: I1201 08:41:51.884841 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/e70d6d2c-f0f3-4c8f-8213-efaa11de813b-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-xs4nq\" (UID: \"e70d6d2c-f0f3-4c8f-8213-efaa11de813b\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xs4nq" Dec 01 08:41:51 crc kubenswrapper[4813]: I1201 08:41:51.885605 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/e70d6d2c-f0f3-4c8f-8213-efaa11de813b-service-ca\") pod \"cluster-version-operator-5c965bbfc6-xs4nq\" (UID: \"e70d6d2c-f0f3-4c8f-8213-efaa11de813b\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xs4nq" Dec 01 08:41:51 crc kubenswrapper[4813]: I1201 08:41:51.906051 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e70d6d2c-f0f3-4c8f-8213-efaa11de813b-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-xs4nq\" (UID: \"e70d6d2c-f0f3-4c8f-8213-efaa11de813b\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xs4nq" Dec 01 08:41:51 crc kubenswrapper[4813]: I1201 08:41:51.910306 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e70d6d2c-f0f3-4c8f-8213-efaa11de813b-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-xs4nq\" (UID: \"e70d6d2c-f0f3-4c8f-8213-efaa11de813b\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xs4nq" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.054757 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xs4nq" Dec 01 08:41:52 crc kubenswrapper[4813]: W1201 08:41:52.080020 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode70d6d2c_f0f3_4c8f_8213_efaa11de813b.slice/crio-4e9c469b364cac6887e1816cc40a325002f071e55a8af011c1cdf3b510a3a532 WatchSource:0}: Error finding container 4e9c469b364cac6887e1816cc40a325002f071e55a8af011c1cdf3b510a3a532: Status 404 returned error can't find the container with id 4e9c469b364cac6887e1816cc40a325002f071e55a8af011c1cdf3b510a3a532 Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.122145 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.122324 4813 kubelet_node_status.go:538] "Fast updating node status as it just became ready" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.311677 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-dvhpg"] Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.312169 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-gnd5s"] Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.312410 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-gnd5s" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.312713 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-dvhpg" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.316957 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-n72r5"] Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.317345 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-l4zh5"] Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.317534 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-lt4dx"] Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.317931 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lt4dx" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.318563 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.318786 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.319020 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.319024 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.320049 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-9ql8j"] Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.324487 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-n72r5" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.325230 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-wtwtl"] Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.325333 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-l4zh5" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.325458 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-5ptsh"] Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.326073 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-gpd6h"] Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.326591 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-vkbsg"] Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.326903 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-vkbsg" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.327098 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.327280 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.327396 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.327570 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.327797 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.328077 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.328214 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.328343 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.328468 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.328797 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-wtwtl" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.328988 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-9ql8j" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.329295 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-gpd6h" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.329307 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-5ptsh" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.331908 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-s4vx5"] Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.332533 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-ct7w8"] Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.332830 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-7q5lc"] Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.333328 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-7q5lc" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.333946 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-s4vx5" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.334020 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-ct7w8" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.348253 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.349021 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.349249 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.349411 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.349766 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.350217 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.350494 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.350514 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.350716 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-9cxt7"] Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.351832 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-wqkxc"] Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.353498 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-9cxt7" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.354909 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.355256 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.381520 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-wqkxc" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.382447 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-k9t5k"] Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.383070 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.383121 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.383881 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.384096 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.384209 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.384359 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.385147 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.385527 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-hmh5v"] Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.385566 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.385997 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-k9t5k" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.385621 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.385640 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.385641 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.385660 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.385685 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.386620 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-7w2kz"] Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.386705 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-hmh5v" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.387181 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-7w2kz" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.385694 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.385718 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.385738 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.385749 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.385774 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.385798 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.387530 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.385832 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.385860 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.385883 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.388085 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.388123 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.388145 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.388214 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.388319 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.388344 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.389035 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-5fgn9"] Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.389111 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.389501 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-5fgn9" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.390156 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-2mb7d"] Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.391573 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.391716 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.391918 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.392031 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.392129 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.392313 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.392428 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.392577 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.392720 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.392734 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.392807 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.392846 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.392888 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.392984 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.393077 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.393130 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.393164 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.393250 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.393346 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.393434 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.393487 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.393510 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.393574 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.393601 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.393671 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.393681 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.392989 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.393945 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.393439 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.394415 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/b35b7145-810b-4945-9819-6d9bb650ec73-images\") pod \"machine-api-operator-5694c8668f-n72r5\" (UID: \"b35b7145-810b-4945-9819-6d9bb650ec73\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-n72r5" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.394435 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/159a2fad-3042-43e6-8df3-16f1157db0b1-etcd-service-ca\") pod \"etcd-operator-b45778765-9cxt7\" (UID: \"159a2fad-3042-43e6-8df3-16f1157db0b1\") " pod="openshift-etcd-operator/etcd-operator-b45778765-9cxt7" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.394455 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/705b2456-4ba7-4775-84ca-4dcea64b6755-serving-cert\") pod \"controller-manager-879f6c89f-l4zh5\" (UID: \"705b2456-4ba7-4775-84ca-4dcea64b6755\") " pod="openshift-controller-manager/controller-manager-879f6c89f-l4zh5" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.394472 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3c957bbc-fa1d-4188-9f9f-0b67be4c56c7-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-vkbsg\" (UID: \"3c957bbc-fa1d-4188-9f9f-0b67be4c56c7\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-vkbsg" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.394488 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d2aac47a-b117-42c7-afa3-2c989d9536a5-trusted-ca-bundle\") pod \"apiserver-76f77b778f-5ptsh\" (UID: \"d2aac47a-b117-42c7-afa3-2c989d9536a5\") " pod="openshift-apiserver/apiserver-76f77b778f-5ptsh" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.394502 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4828f51e-d209-4d31-a608-7d0951d311f4-config\") pod \"machine-approver-56656f9798-gpd6h\" (UID: \"4828f51e-d209-4d31-a608-7d0951d311f4\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-gpd6h" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.394516 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/65bbc5b9-f6f0-40ad-9e47-4122bb5cabab-console-oauth-config\") pod \"console-f9d7485db-9ql8j\" (UID: \"65bbc5b9-f6f0-40ad-9e47-4122bb5cabab\") " pod="openshift-console/console-f9d7485db-9ql8j" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.394556 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/d2aac47a-b117-42c7-afa3-2c989d9536a5-etcd-serving-ca\") pod \"apiserver-76f77b778f-5ptsh\" (UID: \"d2aac47a-b117-42c7-afa3-2c989d9536a5\") " pod="openshift-apiserver/apiserver-76f77b778f-5ptsh" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.394570 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/d2aac47a-b117-42c7-afa3-2c989d9536a5-image-import-ca\") pod \"apiserver-76f77b778f-5ptsh\" (UID: \"d2aac47a-b117-42c7-afa3-2c989d9536a5\") " pod="openshift-apiserver/apiserver-76f77b778f-5ptsh" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.394586 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/0e3d53a2-6105-4c2f-a74b-0b8dcd3b0706-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-wtwtl\" (UID: \"0e3d53a2-6105-4c2f-a74b-0b8dcd3b0706\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-wtwtl" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.394604 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j8cxb\" (UniqueName: \"kubernetes.io/projected/df338eea-faa9-4dca-9c75-f12a2bbf87a4-kube-api-access-j8cxb\") pod \"cluster-samples-operator-665b6dd947-dvhpg\" (UID: \"df338eea-faa9-4dca-9c75-f12a2bbf87a4\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-dvhpg" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.394628 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cdxmn\" (UniqueName: \"kubernetes.io/projected/b35b7145-810b-4945-9819-6d9bb650ec73-kube-api-access-cdxmn\") pod \"machine-api-operator-5694c8668f-n72r5\" (UID: \"b35b7145-810b-4945-9819-6d9bb650ec73\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-n72r5" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.394645 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qmxx7\" (UniqueName: \"kubernetes.io/projected/7f38918f-c551-4e3f-bb5e-d5dcac6ac826-kube-api-access-qmxx7\") pod \"apiserver-7bbb656c7d-lt4dx\" (UID: \"7f38918f-c551-4e3f-bb5e-d5dcac6ac826\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lt4dx" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.394663 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pndx7\" (UniqueName: \"kubernetes.io/projected/0e3d53a2-6105-4c2f-a74b-0b8dcd3b0706-kube-api-access-pndx7\") pod \"cluster-image-registry-operator-dc59b4c8b-wtwtl\" (UID: \"0e3d53a2-6105-4c2f-a74b-0b8dcd3b0706\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-wtwtl" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.394680 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/7f38918f-c551-4e3f-bb5e-d5dcac6ac826-audit-dir\") pod \"apiserver-7bbb656c7d-lt4dx\" (UID: \"7f38918f-c551-4e3f-bb5e-d5dcac6ac826\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lt4dx" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.394727 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/b35b7145-810b-4945-9819-6d9bb650ec73-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-n72r5\" (UID: \"b35b7145-810b-4945-9819-6d9bb650ec73\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-n72r5" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.394745 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/705b2456-4ba7-4775-84ca-4dcea64b6755-client-ca\") pod \"controller-manager-879f6c89f-l4zh5\" (UID: \"705b2456-4ba7-4775-84ca-4dcea64b6755\") " pod="openshift-controller-manager/controller-manager-879f6c89f-l4zh5" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.394761 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/d2aac47a-b117-42c7-afa3-2c989d9536a5-audit-dir\") pod \"apiserver-76f77b778f-5ptsh\" (UID: \"d2aac47a-b117-42c7-afa3-2c989d9536a5\") " pod="openshift-apiserver/apiserver-76f77b778f-5ptsh" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.394777 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/59a1ff0e-6c38-4e97-8423-1641e5620570-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-7q5lc\" (UID: \"59a1ff0e-6c38-4e97-8423-1641e5620570\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-7q5lc" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.394611 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.394823 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/7f38918f-c551-4e3f-bb5e-d5dcac6ac826-encryption-config\") pod \"apiserver-7bbb656c7d-lt4dx\" (UID: \"7f38918f-c551-4e3f-bb5e-d5dcac6ac826\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lt4dx" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.394873 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/156d5b06-7b82-4444-817f-c7bf13cb3c44-metrics-tls\") pod \"ingress-operator-5b745b69d9-wqkxc\" (UID: \"156d5b06-7b82-4444-817f-c7bf13cb3c44\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-wqkxc" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.394922 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7f38918f-c551-4e3f-bb5e-d5dcac6ac826-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-lt4dx\" (UID: \"7f38918f-c551-4e3f-bb5e-d5dcac6ac826\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lt4dx" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.394944 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gdrs4\" (UniqueName: \"kubernetes.io/projected/159a2fad-3042-43e6-8df3-16f1157db0b1-kube-api-access-gdrs4\") pod \"etcd-operator-b45778765-9cxt7\" (UID: \"159a2fad-3042-43e6-8df3-16f1157db0b1\") " pod="openshift-etcd-operator/etcd-operator-b45778765-9cxt7" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.394974 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/4828f51e-d209-4d31-a608-7d0951d311f4-auth-proxy-config\") pod \"machine-approver-56656f9798-gpd6h\" (UID: \"4828f51e-d209-4d31-a608-7d0951d311f4\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-gpd6h" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.395002 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tm5gl\" (UniqueName: \"kubernetes.io/projected/e17c11d6-eb41-4391-b513-a817eeffcdb4-kube-api-access-tm5gl\") pod \"downloads-7954f5f757-ct7w8\" (UID: \"e17c11d6-eb41-4391-b513-a817eeffcdb4\") " pod="openshift-console/downloads-7954f5f757-ct7w8" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.395022 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/159a2fad-3042-43e6-8df3-16f1157db0b1-etcd-client\") pod \"etcd-operator-b45778765-9cxt7\" (UID: \"159a2fad-3042-43e6-8df3-16f1157db0b1\") " pod="openshift-etcd-operator/etcd-operator-b45778765-9cxt7" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.395038 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/7f38918f-c551-4e3f-bb5e-d5dcac6ac826-audit-policies\") pod \"apiserver-7bbb656c7d-lt4dx\" (UID: \"7f38918f-c551-4e3f-bb5e-d5dcac6ac826\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lt4dx" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.395052 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/d2aac47a-b117-42c7-afa3-2c989d9536a5-audit\") pod \"apiserver-76f77b778f-5ptsh\" (UID: \"d2aac47a-b117-42c7-afa3-2c989d9536a5\") " pod="openshift-apiserver/apiserver-76f77b778f-5ptsh" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.395067 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/df338eea-faa9-4dca-9c75-f12a2bbf87a4-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-dvhpg\" (UID: \"df338eea-faa9-4dca-9c75-f12a2bbf87a4\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-dvhpg" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.395084 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/65bbc5b9-f6f0-40ad-9e47-4122bb5cabab-oauth-serving-cert\") pod \"console-f9d7485db-9ql8j\" (UID: \"65bbc5b9-f6f0-40ad-9e47-4122bb5cabab\") " pod="openshift-console/console-f9d7485db-9ql8j" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.395103 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/54725907-0303-4912-875b-7c9a533486fb-serving-cert\") pod \"authentication-operator-69f744f599-k9t5k\" (UID: \"54725907-0303-4912-875b-7c9a533486fb\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-k9t5k" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.395127 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9kdgg\" (UniqueName: \"kubernetes.io/projected/54725907-0303-4912-875b-7c9a533486fb-kube-api-access-9kdgg\") pod \"authentication-operator-69f744f599-k9t5k\" (UID: \"54725907-0303-4912-875b-7c9a533486fb\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-k9t5k" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.395181 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vgfgc\" (UniqueName: \"kubernetes.io/projected/adfb805f-2f2e-40a1-91eb-59e6f23df192-kube-api-access-vgfgc\") pod \"openshift-apiserver-operator-796bbdcf4f-gnd5s\" (UID: \"adfb805f-2f2e-40a1-91eb-59e6f23df192\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-gnd5s" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.395234 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zpbjs\" (UniqueName: \"kubernetes.io/projected/65bbc5b9-f6f0-40ad-9e47-4122bb5cabab-kube-api-access-zpbjs\") pod \"console-f9d7485db-9ql8j\" (UID: \"65bbc5b9-f6f0-40ad-9e47-4122bb5cabab\") " pod="openshift-console/console-f9d7485db-9ql8j" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.395271 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/54725907-0303-4912-875b-7c9a533486fb-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-k9t5k\" (UID: \"54725907-0303-4912-875b-7c9a533486fb\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-k9t5k" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.395314 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3c957bbc-fa1d-4188-9f9f-0b67be4c56c7-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-vkbsg\" (UID: \"3c957bbc-fa1d-4188-9f9f-0b67be4c56c7\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-vkbsg" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.395331 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n6sss\" (UniqueName: \"kubernetes.io/projected/3c957bbc-fa1d-4188-9f9f-0b67be4c56c7-kube-api-access-n6sss\") pod \"openshift-controller-manager-operator-756b6f6bc6-vkbsg\" (UID: \"3c957bbc-fa1d-4188-9f9f-0b67be4c56c7\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-vkbsg" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.395348 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/65bbc5b9-f6f0-40ad-9e47-4122bb5cabab-service-ca\") pod \"console-f9d7485db-9ql8j\" (UID: \"65bbc5b9-f6f0-40ad-9e47-4122bb5cabab\") " pod="openshift-console/console-f9d7485db-9ql8j" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.395378 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/0e3d53a2-6105-4c2f-a74b-0b8dcd3b0706-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-wtwtl\" (UID: \"0e3d53a2-6105-4c2f-a74b-0b8dcd3b0706\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-wtwtl" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.395412 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/d2aac47a-b117-42c7-afa3-2c989d9536a5-node-pullsecrets\") pod \"apiserver-76f77b778f-5ptsh\" (UID: \"d2aac47a-b117-42c7-afa3-2c989d9536a5\") " pod="openshift-apiserver/apiserver-76f77b778f-5ptsh" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.395431 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7s96q\" (UniqueName: \"kubernetes.io/projected/4828f51e-d209-4d31-a608-7d0951d311f4-kube-api-access-7s96q\") pod \"machine-approver-56656f9798-gpd6h\" (UID: \"4828f51e-d209-4d31-a608-7d0951d311f4\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-gpd6h" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.395445 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kwqbn\" (UniqueName: \"kubernetes.io/projected/156d5b06-7b82-4444-817f-c7bf13cb3c44-kube-api-access-kwqbn\") pod \"ingress-operator-5b745b69d9-wqkxc\" (UID: \"156d5b06-7b82-4444-817f-c7bf13cb3c44\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-wqkxc" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.395464 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7f38918f-c551-4e3f-bb5e-d5dcac6ac826-serving-cert\") pod \"apiserver-7bbb656c7d-lt4dx\" (UID: \"7f38918f-c551-4e3f-bb5e-d5dcac6ac826\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lt4dx" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.395481 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/159a2fad-3042-43e6-8df3-16f1157db0b1-config\") pod \"etcd-operator-b45778765-9cxt7\" (UID: \"159a2fad-3042-43e6-8df3-16f1157db0b1\") " pod="openshift-etcd-operator/etcd-operator-b45778765-9cxt7" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.395498 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/156d5b06-7b82-4444-817f-c7bf13cb3c44-bound-sa-token\") pod \"ingress-operator-5b745b69d9-wqkxc\" (UID: \"156d5b06-7b82-4444-817f-c7bf13cb3c44\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-wqkxc" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.395527 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n4h95\" (UniqueName: \"kubernetes.io/projected/d2aac47a-b117-42c7-afa3-2c989d9536a5-kube-api-access-n4h95\") pod \"apiserver-76f77b778f-5ptsh\" (UID: \"d2aac47a-b117-42c7-afa3-2c989d9536a5\") " pod="openshift-apiserver/apiserver-76f77b778f-5ptsh" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.395561 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/d2aac47a-b117-42c7-afa3-2c989d9536a5-etcd-client\") pod \"apiserver-76f77b778f-5ptsh\" (UID: \"d2aac47a-b117-42c7-afa3-2c989d9536a5\") " pod="openshift-apiserver/apiserver-76f77b778f-5ptsh" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.395577 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/cf79ef17-ec75-4d0b-b665-ce24618a92df-metrics-tls\") pod \"dns-operator-744455d44c-s4vx5\" (UID: \"cf79ef17-ec75-4d0b-b665-ce24618a92df\") " pod="openshift-dns-operator/dns-operator-744455d44c-s4vx5" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.395610 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/7f38918f-c551-4e3f-bb5e-d5dcac6ac826-etcd-client\") pod \"apiserver-7bbb656c7d-lt4dx\" (UID: \"7f38918f-c551-4e3f-bb5e-d5dcac6ac826\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lt4dx" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.395628 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/0e3d53a2-6105-4c2f-a74b-0b8dcd3b0706-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-wtwtl\" (UID: \"0e3d53a2-6105-4c2f-a74b-0b8dcd3b0706\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-wtwtl" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.395643 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/54725907-0303-4912-875b-7c9a533486fb-service-ca-bundle\") pod \"authentication-operator-69f744f599-k9t5k\" (UID: \"54725907-0303-4912-875b-7c9a533486fb\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-k9t5k" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.395661 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/159a2fad-3042-43e6-8df3-16f1157db0b1-serving-cert\") pod \"etcd-operator-b45778765-9cxt7\" (UID: \"159a2fad-3042-43e6-8df3-16f1157db0b1\") " pod="openshift-etcd-operator/etcd-operator-b45778765-9cxt7" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.395675 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/adfb805f-2f2e-40a1-91eb-59e6f23df192-config\") pod \"openshift-apiserver-operator-796bbdcf4f-gnd5s\" (UID: \"adfb805f-2f2e-40a1-91eb-59e6f23df192\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-gnd5s" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.395699 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/4828f51e-d209-4d31-a608-7d0951d311f4-machine-approver-tls\") pod \"machine-approver-56656f9798-gpd6h\" (UID: \"4828f51e-d209-4d31-a608-7d0951d311f4\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-gpd6h" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.395720 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/59a1ff0e-6c38-4e97-8423-1641e5620570-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-7q5lc\" (UID: \"59a1ff0e-6c38-4e97-8423-1641e5620570\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-7q5lc" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.395733 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/adfb805f-2f2e-40a1-91eb-59e6f23df192-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-gnd5s\" (UID: \"adfb805f-2f2e-40a1-91eb-59e6f23df192\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-gnd5s" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.395749 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pjxkr\" (UniqueName: \"kubernetes.io/projected/705b2456-4ba7-4775-84ca-4dcea64b6755-kube-api-access-pjxkr\") pod \"controller-manager-879f6c89f-l4zh5\" (UID: \"705b2456-4ba7-4775-84ca-4dcea64b6755\") " pod="openshift-controller-manager/controller-manager-879f6c89f-l4zh5" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.395773 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d2aac47a-b117-42c7-afa3-2c989d9536a5-config\") pod \"apiserver-76f77b778f-5ptsh\" (UID: \"d2aac47a-b117-42c7-afa3-2c989d9536a5\") " pod="openshift-apiserver/apiserver-76f77b778f-5ptsh" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.395787 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/54725907-0303-4912-875b-7c9a533486fb-config\") pod \"authentication-operator-69f744f599-k9t5k\" (UID: \"54725907-0303-4912-875b-7c9a533486fb\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-k9t5k" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.395811 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/705b2456-4ba7-4775-84ca-4dcea64b6755-config\") pod \"controller-manager-879f6c89f-l4zh5\" (UID: \"705b2456-4ba7-4775-84ca-4dcea64b6755\") " pod="openshift-controller-manager/controller-manager-879f6c89f-l4zh5" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.395849 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8ljvz\" (UniqueName: \"kubernetes.io/projected/cf79ef17-ec75-4d0b-b665-ce24618a92df-kube-api-access-8ljvz\") pod \"dns-operator-744455d44c-s4vx5\" (UID: \"cf79ef17-ec75-4d0b-b665-ce24618a92df\") " pod="openshift-dns-operator/dns-operator-744455d44c-s4vx5" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.395890 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/705b2456-4ba7-4775-84ca-4dcea64b6755-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-l4zh5\" (UID: \"705b2456-4ba7-4775-84ca-4dcea64b6755\") " pod="openshift-controller-manager/controller-manager-879f6c89f-l4zh5" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.395918 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d2aac47a-b117-42c7-afa3-2c989d9536a5-serving-cert\") pod \"apiserver-76f77b778f-5ptsh\" (UID: \"d2aac47a-b117-42c7-afa3-2c989d9536a5\") " pod="openshift-apiserver/apiserver-76f77b778f-5ptsh" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.395955 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zlhdb\" (UniqueName: \"kubernetes.io/projected/59a1ff0e-6c38-4e97-8423-1641e5620570-kube-api-access-zlhdb\") pod \"kube-storage-version-migrator-operator-b67b599dd-7q5lc\" (UID: \"59a1ff0e-6c38-4e97-8423-1641e5620570\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-7q5lc" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.395996 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/159a2fad-3042-43e6-8df3-16f1157db0b1-etcd-ca\") pod \"etcd-operator-b45778765-9cxt7\" (UID: \"159a2fad-3042-43e6-8df3-16f1157db0b1\") " pod="openshift-etcd-operator/etcd-operator-b45778765-9cxt7" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.396017 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/65bbc5b9-f6f0-40ad-9e47-4122bb5cabab-trusted-ca-bundle\") pod \"console-f9d7485db-9ql8j\" (UID: \"65bbc5b9-f6f0-40ad-9e47-4122bb5cabab\") " pod="openshift-console/console-f9d7485db-9ql8j" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.396052 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/7f38918f-c551-4e3f-bb5e-d5dcac6ac826-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-lt4dx\" (UID: \"7f38918f-c551-4e3f-bb5e-d5dcac6ac826\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lt4dx" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.396081 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/156d5b06-7b82-4444-817f-c7bf13cb3c44-trusted-ca\") pod \"ingress-operator-5b745b69d9-wqkxc\" (UID: \"156d5b06-7b82-4444-817f-c7bf13cb3c44\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-wqkxc" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.396100 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/65bbc5b9-f6f0-40ad-9e47-4122bb5cabab-console-config\") pod \"console-f9d7485db-9ql8j\" (UID: \"65bbc5b9-f6f0-40ad-9e47-4122bb5cabab\") " pod="openshift-console/console-f9d7485db-9ql8j" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.396123 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/d2aac47a-b117-42c7-afa3-2c989d9536a5-encryption-config\") pod \"apiserver-76f77b778f-5ptsh\" (UID: \"d2aac47a-b117-42c7-afa3-2c989d9536a5\") " pod="openshift-apiserver/apiserver-76f77b778f-5ptsh" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.396145 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b35b7145-810b-4945-9819-6d9bb650ec73-config\") pod \"machine-api-operator-5694c8668f-n72r5\" (UID: \"b35b7145-810b-4945-9819-6d9bb650ec73\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-n72r5" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.396171 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/65bbc5b9-f6f0-40ad-9e47-4122bb5cabab-console-serving-cert\") pod \"console-f9d7485db-9ql8j\" (UID: \"65bbc5b9-f6f0-40ad-9e47-4122bb5cabab\") " pod="openshift-console/console-f9d7485db-9ql8j" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.396338 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.396517 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-2mb7d" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.396791 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.397490 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.397721 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.397873 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.398002 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.398712 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.402165 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.451884 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.452890 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.455048 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.455594 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.455646 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-96mbr" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.456231 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.457051 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.459546 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.462894 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-mj64s"] Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.464522 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-8274k"] Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.467930 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-mj64s" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.489695 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-fxgbc"] Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.490177 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.490199 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-lfm7c"] Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.490475 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-sjjm7"] Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.490607 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.490882 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-dwkm7"] Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.490895 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.491288 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-sp4wh"] Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.491596 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-qsskc"] Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.491716 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-sp4wh" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.491861 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-sjjm7" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.491711 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-lfm7c" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.492881 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-qsskc" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.492914 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.493174 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-dwkm7" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.493452 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-fxgbc" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.493769 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.494397 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-8274k" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.495455 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.496407 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-jb992"] Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.497436 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-jb992" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.497569 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/705b2456-4ba7-4775-84ca-4dcea64b6755-serving-cert\") pod \"controller-manager-879f6c89f-l4zh5\" (UID: \"705b2456-4ba7-4775-84ca-4dcea64b6755\") " pod="openshift-controller-manager/controller-manager-879f6c89f-l4zh5" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.497599 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3c957bbc-fa1d-4188-9f9f-0b67be4c56c7-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-vkbsg\" (UID: \"3c957bbc-fa1d-4188-9f9f-0b67be4c56c7\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-vkbsg" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.497780 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d2aac47a-b117-42c7-afa3-2c989d9536a5-trusted-ca-bundle\") pod \"apiserver-76f77b778f-5ptsh\" (UID: \"d2aac47a-b117-42c7-afa3-2c989d9536a5\") " pod="openshift-apiserver/apiserver-76f77b778f-5ptsh" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.497820 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4828f51e-d209-4d31-a608-7d0951d311f4-config\") pod \"machine-approver-56656f9798-gpd6h\" (UID: \"4828f51e-d209-4d31-a608-7d0951d311f4\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-gpd6h" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.497835 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/65bbc5b9-f6f0-40ad-9e47-4122bb5cabab-console-oauth-config\") pod \"console-f9d7485db-9ql8j\" (UID: \"65bbc5b9-f6f0-40ad-9e47-4122bb5cabab\") " pod="openshift-console/console-f9d7485db-9ql8j" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.497851 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/d2aac47a-b117-42c7-afa3-2c989d9536a5-etcd-serving-ca\") pod \"apiserver-76f77b778f-5ptsh\" (UID: \"d2aac47a-b117-42c7-afa3-2c989d9536a5\") " pod="openshift-apiserver/apiserver-76f77b778f-5ptsh" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.497865 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/d2aac47a-b117-42c7-afa3-2c989d9536a5-image-import-ca\") pod \"apiserver-76f77b778f-5ptsh\" (UID: \"d2aac47a-b117-42c7-afa3-2c989d9536a5\") " pod="openshift-apiserver/apiserver-76f77b778f-5ptsh" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.497883 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/0e3d53a2-6105-4c2f-a74b-0b8dcd3b0706-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-wtwtl\" (UID: \"0e3d53a2-6105-4c2f-a74b-0b8dcd3b0706\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-wtwtl" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.497901 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j8cxb\" (UniqueName: \"kubernetes.io/projected/df338eea-faa9-4dca-9c75-f12a2bbf87a4-kube-api-access-j8cxb\") pod \"cluster-samples-operator-665b6dd947-dvhpg\" (UID: \"df338eea-faa9-4dca-9c75-f12a2bbf87a4\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-dvhpg" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.497924 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cdxmn\" (UniqueName: \"kubernetes.io/projected/b35b7145-810b-4945-9819-6d9bb650ec73-kube-api-access-cdxmn\") pod \"machine-api-operator-5694c8668f-n72r5\" (UID: \"b35b7145-810b-4945-9819-6d9bb650ec73\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-n72r5" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.497940 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qmxx7\" (UniqueName: \"kubernetes.io/projected/7f38918f-c551-4e3f-bb5e-d5dcac6ac826-kube-api-access-qmxx7\") pod \"apiserver-7bbb656c7d-lt4dx\" (UID: \"7f38918f-c551-4e3f-bb5e-d5dcac6ac826\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lt4dx" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.497955 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pndx7\" (UniqueName: \"kubernetes.io/projected/0e3d53a2-6105-4c2f-a74b-0b8dcd3b0706-kube-api-access-pndx7\") pod \"cluster-image-registry-operator-dc59b4c8b-wtwtl\" (UID: \"0e3d53a2-6105-4c2f-a74b-0b8dcd3b0706\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-wtwtl" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.497983 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/7f38918f-c551-4e3f-bb5e-d5dcac6ac826-audit-dir\") pod \"apiserver-7bbb656c7d-lt4dx\" (UID: \"7f38918f-c551-4e3f-bb5e-d5dcac6ac826\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lt4dx" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.497999 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/b35b7145-810b-4945-9819-6d9bb650ec73-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-n72r5\" (UID: \"b35b7145-810b-4945-9819-6d9bb650ec73\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-n72r5" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.498050 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/705b2456-4ba7-4775-84ca-4dcea64b6755-client-ca\") pod \"controller-manager-879f6c89f-l4zh5\" (UID: \"705b2456-4ba7-4775-84ca-4dcea64b6755\") " pod="openshift-controller-manager/controller-manager-879f6c89f-l4zh5" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.498089 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/d2aac47a-b117-42c7-afa3-2c989d9536a5-audit-dir\") pod \"apiserver-76f77b778f-5ptsh\" (UID: \"d2aac47a-b117-42c7-afa3-2c989d9536a5\") " pod="openshift-apiserver/apiserver-76f77b778f-5ptsh" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.498119 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/59a1ff0e-6c38-4e97-8423-1641e5620570-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-7q5lc\" (UID: \"59a1ff0e-6c38-4e97-8423-1641e5620570\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-7q5lc" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.498145 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ab324c2b-a49d-4874-b9c0-a869b226ce98-serving-cert\") pod \"openshift-config-operator-7777fb866f-hmh5v\" (UID: \"ab324c2b-a49d-4874-b9c0-a869b226ce98\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-hmh5v" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.498166 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/156d5b06-7b82-4444-817f-c7bf13cb3c44-metrics-tls\") pod \"ingress-operator-5b745b69d9-wqkxc\" (UID: \"156d5b06-7b82-4444-817f-c7bf13cb3c44\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-wqkxc" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.498198 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/7f38918f-c551-4e3f-bb5e-d5dcac6ac826-encryption-config\") pod \"apiserver-7bbb656c7d-lt4dx\" (UID: \"7f38918f-c551-4e3f-bb5e-d5dcac6ac826\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lt4dx" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.498246 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7f38918f-c551-4e3f-bb5e-d5dcac6ac826-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-lt4dx\" (UID: \"7f38918f-c551-4e3f-bb5e-d5dcac6ac826\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lt4dx" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.498265 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gdrs4\" (UniqueName: \"kubernetes.io/projected/159a2fad-3042-43e6-8df3-16f1157db0b1-kube-api-access-gdrs4\") pod \"etcd-operator-b45778765-9cxt7\" (UID: \"159a2fad-3042-43e6-8df3-16f1157db0b1\") " pod="openshift-etcd-operator/etcd-operator-b45778765-9cxt7" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.498283 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/4828f51e-d209-4d31-a608-7d0951d311f4-auth-proxy-config\") pod \"machine-approver-56656f9798-gpd6h\" (UID: \"4828f51e-d209-4d31-a608-7d0951d311f4\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-gpd6h" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.498303 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tm5gl\" (UniqueName: \"kubernetes.io/projected/e17c11d6-eb41-4391-b513-a817eeffcdb4-kube-api-access-tm5gl\") pod \"downloads-7954f5f757-ct7w8\" (UID: \"e17c11d6-eb41-4391-b513-a817eeffcdb4\") " pod="openshift-console/downloads-7954f5f757-ct7w8" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.498321 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/159a2fad-3042-43e6-8df3-16f1157db0b1-etcd-client\") pod \"etcd-operator-b45778765-9cxt7\" (UID: \"159a2fad-3042-43e6-8df3-16f1157db0b1\") " pod="openshift-etcd-operator/etcd-operator-b45778765-9cxt7" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.498345 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vv2sl\" (UniqueName: \"kubernetes.io/projected/ab324c2b-a49d-4874-b9c0-a869b226ce98-kube-api-access-vv2sl\") pod \"openshift-config-operator-7777fb866f-hmh5v\" (UID: \"ab324c2b-a49d-4874-b9c0-a869b226ce98\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-hmh5v" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.498371 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/7f38918f-c551-4e3f-bb5e-d5dcac6ac826-audit-policies\") pod \"apiserver-7bbb656c7d-lt4dx\" (UID: \"7f38918f-c551-4e3f-bb5e-d5dcac6ac826\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lt4dx" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.498391 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/d2aac47a-b117-42c7-afa3-2c989d9536a5-audit\") pod \"apiserver-76f77b778f-5ptsh\" (UID: \"d2aac47a-b117-42c7-afa3-2c989d9536a5\") " pod="openshift-apiserver/apiserver-76f77b778f-5ptsh" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.498406 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/df338eea-faa9-4dca-9c75-f12a2bbf87a4-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-dvhpg\" (UID: \"df338eea-faa9-4dca-9c75-f12a2bbf87a4\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-dvhpg" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.498430 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/65bbc5b9-f6f0-40ad-9e47-4122bb5cabab-oauth-serving-cert\") pod \"console-f9d7485db-9ql8j\" (UID: \"65bbc5b9-f6f0-40ad-9e47-4122bb5cabab\") " pod="openshift-console/console-f9d7485db-9ql8j" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.498448 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/54725907-0303-4912-875b-7c9a533486fb-serving-cert\") pod \"authentication-operator-69f744f599-k9t5k\" (UID: \"54725907-0303-4912-875b-7c9a533486fb\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-k9t5k" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.498469 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9kdgg\" (UniqueName: \"kubernetes.io/projected/54725907-0303-4912-875b-7c9a533486fb-kube-api-access-9kdgg\") pod \"authentication-operator-69f744f599-k9t5k\" (UID: \"54725907-0303-4912-875b-7c9a533486fb\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-k9t5k" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.498495 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vgfgc\" (UniqueName: \"kubernetes.io/projected/adfb805f-2f2e-40a1-91eb-59e6f23df192-kube-api-access-vgfgc\") pod \"openshift-apiserver-operator-796bbdcf4f-gnd5s\" (UID: \"adfb805f-2f2e-40a1-91eb-59e6f23df192\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-gnd5s" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.498513 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zpbjs\" (UniqueName: \"kubernetes.io/projected/65bbc5b9-f6f0-40ad-9e47-4122bb5cabab-kube-api-access-zpbjs\") pod \"console-f9d7485db-9ql8j\" (UID: \"65bbc5b9-f6f0-40ad-9e47-4122bb5cabab\") " pod="openshift-console/console-f9d7485db-9ql8j" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.498533 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/54725907-0303-4912-875b-7c9a533486fb-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-k9t5k\" (UID: \"54725907-0303-4912-875b-7c9a533486fb\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-k9t5k" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.498552 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3c957bbc-fa1d-4188-9f9f-0b67be4c56c7-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-vkbsg\" (UID: \"3c957bbc-fa1d-4188-9f9f-0b67be4c56c7\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-vkbsg" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.498569 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n6sss\" (UniqueName: \"kubernetes.io/projected/3c957bbc-fa1d-4188-9f9f-0b67be4c56c7-kube-api-access-n6sss\") pod \"openshift-controller-manager-operator-756b6f6bc6-vkbsg\" (UID: \"3c957bbc-fa1d-4188-9f9f-0b67be4c56c7\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-vkbsg" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.498586 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/65bbc5b9-f6f0-40ad-9e47-4122bb5cabab-service-ca\") pod \"console-f9d7485db-9ql8j\" (UID: \"65bbc5b9-f6f0-40ad-9e47-4122bb5cabab\") " pod="openshift-console/console-f9d7485db-9ql8j" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.498609 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/0e3d53a2-6105-4c2f-a74b-0b8dcd3b0706-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-wtwtl\" (UID: \"0e3d53a2-6105-4c2f-a74b-0b8dcd3b0706\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-wtwtl" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.498624 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kwqbn\" (UniqueName: \"kubernetes.io/projected/156d5b06-7b82-4444-817f-c7bf13cb3c44-kube-api-access-kwqbn\") pod \"ingress-operator-5b745b69d9-wqkxc\" (UID: \"156d5b06-7b82-4444-817f-c7bf13cb3c44\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-wqkxc" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.498640 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/d2aac47a-b117-42c7-afa3-2c989d9536a5-node-pullsecrets\") pod \"apiserver-76f77b778f-5ptsh\" (UID: \"d2aac47a-b117-42c7-afa3-2c989d9536a5\") " pod="openshift-apiserver/apiserver-76f77b778f-5ptsh" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.498656 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7s96q\" (UniqueName: \"kubernetes.io/projected/4828f51e-d209-4d31-a608-7d0951d311f4-kube-api-access-7s96q\") pod \"machine-approver-56656f9798-gpd6h\" (UID: \"4828f51e-d209-4d31-a608-7d0951d311f4\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-gpd6h" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.498671 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7f38918f-c551-4e3f-bb5e-d5dcac6ac826-serving-cert\") pod \"apiserver-7bbb656c7d-lt4dx\" (UID: \"7f38918f-c551-4e3f-bb5e-d5dcac6ac826\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lt4dx" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.498686 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/159a2fad-3042-43e6-8df3-16f1157db0b1-config\") pod \"etcd-operator-b45778765-9cxt7\" (UID: \"159a2fad-3042-43e6-8df3-16f1157db0b1\") " pod="openshift-etcd-operator/etcd-operator-b45778765-9cxt7" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.498701 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/156d5b06-7b82-4444-817f-c7bf13cb3c44-bound-sa-token\") pod \"ingress-operator-5b745b69d9-wqkxc\" (UID: \"156d5b06-7b82-4444-817f-c7bf13cb3c44\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-wqkxc" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.498718 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n4h95\" (UniqueName: \"kubernetes.io/projected/d2aac47a-b117-42c7-afa3-2c989d9536a5-kube-api-access-n4h95\") pod \"apiserver-76f77b778f-5ptsh\" (UID: \"d2aac47a-b117-42c7-afa3-2c989d9536a5\") " pod="openshift-apiserver/apiserver-76f77b778f-5ptsh" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.498733 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/d2aac47a-b117-42c7-afa3-2c989d9536a5-etcd-client\") pod \"apiserver-76f77b778f-5ptsh\" (UID: \"d2aac47a-b117-42c7-afa3-2c989d9536a5\") " pod="openshift-apiserver/apiserver-76f77b778f-5ptsh" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.498749 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/cf79ef17-ec75-4d0b-b665-ce24618a92df-metrics-tls\") pod \"dns-operator-744455d44c-s4vx5\" (UID: \"cf79ef17-ec75-4d0b-b665-ce24618a92df\") " pod="openshift-dns-operator/dns-operator-744455d44c-s4vx5" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.498764 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/adfb805f-2f2e-40a1-91eb-59e6f23df192-config\") pod \"openshift-apiserver-operator-796bbdcf4f-gnd5s\" (UID: \"adfb805f-2f2e-40a1-91eb-59e6f23df192\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-gnd5s" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.498785 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/7f38918f-c551-4e3f-bb5e-d5dcac6ac826-etcd-client\") pod \"apiserver-7bbb656c7d-lt4dx\" (UID: \"7f38918f-c551-4e3f-bb5e-d5dcac6ac826\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lt4dx" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.498799 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/0e3d53a2-6105-4c2f-a74b-0b8dcd3b0706-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-wtwtl\" (UID: \"0e3d53a2-6105-4c2f-a74b-0b8dcd3b0706\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-wtwtl" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.498815 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/54725907-0303-4912-875b-7c9a533486fb-service-ca-bundle\") pod \"authentication-operator-69f744f599-k9t5k\" (UID: \"54725907-0303-4912-875b-7c9a533486fb\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-k9t5k" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.498854 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/159a2fad-3042-43e6-8df3-16f1157db0b1-serving-cert\") pod \"etcd-operator-b45778765-9cxt7\" (UID: \"159a2fad-3042-43e6-8df3-16f1157db0b1\") " pod="openshift-etcd-operator/etcd-operator-b45778765-9cxt7" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.498869 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/4828f51e-d209-4d31-a608-7d0951d311f4-machine-approver-tls\") pod \"machine-approver-56656f9798-gpd6h\" (UID: \"4828f51e-d209-4d31-a608-7d0951d311f4\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-gpd6h" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.498898 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/59a1ff0e-6c38-4e97-8423-1641e5620570-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-7q5lc\" (UID: \"59a1ff0e-6c38-4e97-8423-1641e5620570\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-7q5lc" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.498915 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/adfb805f-2f2e-40a1-91eb-59e6f23df192-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-gnd5s\" (UID: \"adfb805f-2f2e-40a1-91eb-59e6f23df192\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-gnd5s" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.498932 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pjxkr\" (UniqueName: \"kubernetes.io/projected/705b2456-4ba7-4775-84ca-4dcea64b6755-kube-api-access-pjxkr\") pod \"controller-manager-879f6c89f-l4zh5\" (UID: \"705b2456-4ba7-4775-84ca-4dcea64b6755\") " pod="openshift-controller-manager/controller-manager-879f6c89f-l4zh5" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.498947 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d2aac47a-b117-42c7-afa3-2c989d9536a5-config\") pod \"apiserver-76f77b778f-5ptsh\" (UID: \"d2aac47a-b117-42c7-afa3-2c989d9536a5\") " pod="openshift-apiserver/apiserver-76f77b778f-5ptsh" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.498975 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/54725907-0303-4912-875b-7c9a533486fb-config\") pod \"authentication-operator-69f744f599-k9t5k\" (UID: \"54725907-0303-4912-875b-7c9a533486fb\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-k9t5k" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.498999 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/705b2456-4ba7-4775-84ca-4dcea64b6755-config\") pod \"controller-manager-879f6c89f-l4zh5\" (UID: \"705b2456-4ba7-4775-84ca-4dcea64b6755\") " pod="openshift-controller-manager/controller-manager-879f6c89f-l4zh5" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.499020 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8ljvz\" (UniqueName: \"kubernetes.io/projected/cf79ef17-ec75-4d0b-b665-ce24618a92df-kube-api-access-8ljvz\") pod \"dns-operator-744455d44c-s4vx5\" (UID: \"cf79ef17-ec75-4d0b-b665-ce24618a92df\") " pod="openshift-dns-operator/dns-operator-744455d44c-s4vx5" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.499040 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/159a2fad-3042-43e6-8df3-16f1157db0b1-etcd-ca\") pod \"etcd-operator-b45778765-9cxt7\" (UID: \"159a2fad-3042-43e6-8df3-16f1157db0b1\") " pod="openshift-etcd-operator/etcd-operator-b45778765-9cxt7" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.499063 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/705b2456-4ba7-4775-84ca-4dcea64b6755-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-l4zh5\" (UID: \"705b2456-4ba7-4775-84ca-4dcea64b6755\") " pod="openshift-controller-manager/controller-manager-879f6c89f-l4zh5" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.499090 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d2aac47a-b117-42c7-afa3-2c989d9536a5-serving-cert\") pod \"apiserver-76f77b778f-5ptsh\" (UID: \"d2aac47a-b117-42c7-afa3-2c989d9536a5\") " pod="openshift-apiserver/apiserver-76f77b778f-5ptsh" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.499106 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zlhdb\" (UniqueName: \"kubernetes.io/projected/59a1ff0e-6c38-4e97-8423-1641e5620570-kube-api-access-zlhdb\") pod \"kube-storage-version-migrator-operator-b67b599dd-7q5lc\" (UID: \"59a1ff0e-6c38-4e97-8423-1641e5620570\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-7q5lc" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.499121 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/65bbc5b9-f6f0-40ad-9e47-4122bb5cabab-trusted-ca-bundle\") pod \"console-f9d7485db-9ql8j\" (UID: \"65bbc5b9-f6f0-40ad-9e47-4122bb5cabab\") " pod="openshift-console/console-f9d7485db-9ql8j" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.499137 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/ab324c2b-a49d-4874-b9c0-a869b226ce98-available-featuregates\") pod \"openshift-config-operator-7777fb866f-hmh5v\" (UID: \"ab324c2b-a49d-4874-b9c0-a869b226ce98\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-hmh5v" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.499154 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/7f38918f-c551-4e3f-bb5e-d5dcac6ac826-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-lt4dx\" (UID: \"7f38918f-c551-4e3f-bb5e-d5dcac6ac826\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lt4dx" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.499174 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/156d5b06-7b82-4444-817f-c7bf13cb3c44-trusted-ca\") pod \"ingress-operator-5b745b69d9-wqkxc\" (UID: \"156d5b06-7b82-4444-817f-c7bf13cb3c44\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-wqkxc" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.499188 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/65bbc5b9-f6f0-40ad-9e47-4122bb5cabab-console-config\") pod \"console-f9d7485db-9ql8j\" (UID: \"65bbc5b9-f6f0-40ad-9e47-4122bb5cabab\") " pod="openshift-console/console-f9d7485db-9ql8j" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.499211 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/d2aac47a-b117-42c7-afa3-2c989d9536a5-encryption-config\") pod \"apiserver-76f77b778f-5ptsh\" (UID: \"d2aac47a-b117-42c7-afa3-2c989d9536a5\") " pod="openshift-apiserver/apiserver-76f77b778f-5ptsh" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.499228 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b35b7145-810b-4945-9819-6d9bb650ec73-config\") pod \"machine-api-operator-5694c8668f-n72r5\" (UID: \"b35b7145-810b-4945-9819-6d9bb650ec73\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-n72r5" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.499243 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/65bbc5b9-f6f0-40ad-9e47-4122bb5cabab-console-serving-cert\") pod \"console-f9d7485db-9ql8j\" (UID: \"65bbc5b9-f6f0-40ad-9e47-4122bb5cabab\") " pod="openshift-console/console-f9d7485db-9ql8j" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.499262 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/b35b7145-810b-4945-9819-6d9bb650ec73-images\") pod \"machine-api-operator-5694c8668f-n72r5\" (UID: \"b35b7145-810b-4945-9819-6d9bb650ec73\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-n72r5" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.499277 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/159a2fad-3042-43e6-8df3-16f1157db0b1-etcd-service-ca\") pod \"etcd-operator-b45778765-9cxt7\" (UID: \"159a2fad-3042-43e6-8df3-16f1157db0b1\") " pod="openshift-etcd-operator/etcd-operator-b45778765-9cxt7" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.500019 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/159a2fad-3042-43e6-8df3-16f1157db0b1-etcd-service-ca\") pod \"etcd-operator-b45778765-9cxt7\" (UID: \"159a2fad-3042-43e6-8df3-16f1157db0b1\") " pod="openshift-etcd-operator/etcd-operator-b45778765-9cxt7" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.500999 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d2aac47a-b117-42c7-afa3-2c989d9536a5-trusted-ca-bundle\") pod \"apiserver-76f77b778f-5ptsh\" (UID: \"d2aac47a-b117-42c7-afa3-2c989d9536a5\") " pod="openshift-apiserver/apiserver-76f77b778f-5ptsh" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.503182 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/705b2456-4ba7-4775-84ca-4dcea64b6755-serving-cert\") pod \"controller-manager-879f6c89f-l4zh5\" (UID: \"705b2456-4ba7-4775-84ca-4dcea64b6755\") " pod="openshift-controller-manager/controller-manager-879f6c89f-l4zh5" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.503844 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/65bbc5b9-f6f0-40ad-9e47-4122bb5cabab-trusted-ca-bundle\") pod \"console-f9d7485db-9ql8j\" (UID: \"65bbc5b9-f6f0-40ad-9e47-4122bb5cabab\") " pod="openshift-console/console-f9d7485db-9ql8j" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.504276 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d2aac47a-b117-42c7-afa3-2c989d9536a5-config\") pod \"apiserver-76f77b778f-5ptsh\" (UID: \"d2aac47a-b117-42c7-afa3-2c989d9536a5\") " pod="openshift-apiserver/apiserver-76f77b778f-5ptsh" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.504357 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3c957bbc-fa1d-4188-9f9f-0b67be4c56c7-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-vkbsg\" (UID: \"3c957bbc-fa1d-4188-9f9f-0b67be4c56c7\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-vkbsg" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.504936 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/54725907-0303-4912-875b-7c9a533486fb-config\") pod \"authentication-operator-69f744f599-k9t5k\" (UID: \"54725907-0303-4912-875b-7c9a533486fb\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-k9t5k" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.505395 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/54725907-0303-4912-875b-7c9a533486fb-serving-cert\") pod \"authentication-operator-69f744f599-k9t5k\" (UID: \"54725907-0303-4912-875b-7c9a533486fb\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-k9t5k" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.505595 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/d2aac47a-b117-42c7-afa3-2c989d9536a5-node-pullsecrets\") pod \"apiserver-76f77b778f-5ptsh\" (UID: \"d2aac47a-b117-42c7-afa3-2c989d9536a5\") " pod="openshift-apiserver/apiserver-76f77b778f-5ptsh" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.508032 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/d2aac47a-b117-42c7-afa3-2c989d9536a5-audit-dir\") pod \"apiserver-76f77b778f-5ptsh\" (UID: \"d2aac47a-b117-42c7-afa3-2c989d9536a5\") " pod="openshift-apiserver/apiserver-76f77b778f-5ptsh" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.508075 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/705b2456-4ba7-4775-84ca-4dcea64b6755-config\") pod \"controller-manager-879f6c89f-l4zh5\" (UID: \"705b2456-4ba7-4775-84ca-4dcea64b6755\") " pod="openshift-controller-manager/controller-manager-879f6c89f-l4zh5" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.508190 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/7f38918f-c551-4e3f-bb5e-d5dcac6ac826-audit-dir\") pod \"apiserver-7bbb656c7d-lt4dx\" (UID: \"7f38918f-c551-4e3f-bb5e-d5dcac6ac826\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lt4dx" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.509056 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4828f51e-d209-4d31-a608-7d0951d311f4-config\") pod \"machine-approver-56656f9798-gpd6h\" (UID: \"4828f51e-d209-4d31-a608-7d0951d311f4\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-gpd6h" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.509863 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/7f38918f-c551-4e3f-bb5e-d5dcac6ac826-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-lt4dx\" (UID: \"7f38918f-c551-4e3f-bb5e-d5dcac6ac826\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lt4dx" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.510948 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/156d5b06-7b82-4444-817f-c7bf13cb3c44-trusted-ca\") pod \"ingress-operator-5b745b69d9-wqkxc\" (UID: \"156d5b06-7b82-4444-817f-c7bf13cb3c44\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-wqkxc" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.511728 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.511869 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/7f38918f-c551-4e3f-bb5e-d5dcac6ac826-audit-policies\") pod \"apiserver-7bbb656c7d-lt4dx\" (UID: \"7f38918f-c551-4e3f-bb5e-d5dcac6ac826\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lt4dx" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.512285 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/159a2fad-3042-43e6-8df3-16f1157db0b1-etcd-ca\") pod \"etcd-operator-b45778765-9cxt7\" (UID: \"159a2fad-3042-43e6-8df3-16f1157db0b1\") " pod="openshift-etcd-operator/etcd-operator-b45778765-9cxt7" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.512353 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.512453 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/4828f51e-d209-4d31-a608-7d0951d311f4-auth-proxy-config\") pod \"machine-approver-56656f9798-gpd6h\" (UID: \"4828f51e-d209-4d31-a608-7d0951d311f4\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-gpd6h" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.512587 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/159a2fad-3042-43e6-8df3-16f1157db0b1-config\") pod \"etcd-operator-b45778765-9cxt7\" (UID: \"159a2fad-3042-43e6-8df3-16f1157db0b1\") " pod="openshift-etcd-operator/etcd-operator-b45778765-9cxt7" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.512954 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/65bbc5b9-f6f0-40ad-9e47-4122bb5cabab-oauth-serving-cert\") pod \"console-f9d7485db-9ql8j\" (UID: \"65bbc5b9-f6f0-40ad-9e47-4122bb5cabab\") " pod="openshift-console/console-f9d7485db-9ql8j" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.513384 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/705b2456-4ba7-4775-84ca-4dcea64b6755-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-l4zh5\" (UID: \"705b2456-4ba7-4775-84ca-4dcea64b6755\") " pod="openshift-controller-manager/controller-manager-879f6c89f-l4zh5" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.514430 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/d2aac47a-b117-42c7-afa3-2c989d9536a5-audit\") pod \"apiserver-76f77b778f-5ptsh\" (UID: \"d2aac47a-b117-42c7-afa3-2c989d9536a5\") " pod="openshift-apiserver/apiserver-76f77b778f-5ptsh" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.515497 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b35b7145-810b-4945-9819-6d9bb650ec73-config\") pod \"machine-api-operator-5694c8668f-n72r5\" (UID: \"b35b7145-810b-4945-9819-6d9bb650ec73\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-n72r5" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.516649 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7f38918f-c551-4e3f-bb5e-d5dcac6ac826-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-lt4dx\" (UID: \"7f38918f-c551-4e3f-bb5e-d5dcac6ac826\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lt4dx" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.517429 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-5z87f"] Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.517952 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-f8kzc"] Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.518368 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-f8kzc" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.518614 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-5z87f" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.519261 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/59a1ff0e-6c38-4e97-8423-1641e5620570-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-7q5lc\" (UID: \"59a1ff0e-6c38-4e97-8423-1641e5620570\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-7q5lc" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.519736 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3c957bbc-fa1d-4188-9f9f-0b67be4c56c7-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-vkbsg\" (UID: \"3c957bbc-fa1d-4188-9f9f-0b67be4c56c7\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-vkbsg" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.520807 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d2aac47a-b117-42c7-afa3-2c989d9536a5-serving-cert\") pod \"apiserver-76f77b778f-5ptsh\" (UID: \"d2aac47a-b117-42c7-afa3-2c989d9536a5\") " pod="openshift-apiserver/apiserver-76f77b778f-5ptsh" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.521829 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/156d5b06-7b82-4444-817f-c7bf13cb3c44-metrics-tls\") pod \"ingress-operator-5b745b69d9-wqkxc\" (UID: \"156d5b06-7b82-4444-817f-c7bf13cb3c44\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-wqkxc" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.521259 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/0e3d53a2-6105-4c2f-a74b-0b8dcd3b0706-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-wtwtl\" (UID: \"0e3d53a2-6105-4c2f-a74b-0b8dcd3b0706\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-wtwtl" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.522185 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/65bbc5b9-f6f0-40ad-9e47-4122bb5cabab-service-ca\") pod \"console-f9d7485db-9ql8j\" (UID: \"65bbc5b9-f6f0-40ad-9e47-4122bb5cabab\") " pod="openshift-console/console-f9d7485db-9ql8j" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.522531 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/b35b7145-810b-4945-9819-6d9bb650ec73-images\") pod \"machine-api-operator-5694c8668f-n72r5\" (UID: \"b35b7145-810b-4945-9819-6d9bb650ec73\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-n72r5" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.522854 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/d2aac47a-b117-42c7-afa3-2c989d9536a5-etcd-serving-ca\") pod \"apiserver-76f77b778f-5ptsh\" (UID: \"d2aac47a-b117-42c7-afa3-2c989d9536a5\") " pod="openshift-apiserver/apiserver-76f77b778f-5ptsh" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.523043 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/54725907-0303-4912-875b-7c9a533486fb-service-ca-bundle\") pod \"authentication-operator-69f744f599-k9t5k\" (UID: \"54725907-0303-4912-875b-7c9a533486fb\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-k9t5k" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.523335 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/0e3d53a2-6105-4c2f-a74b-0b8dcd3b0706-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-wtwtl\" (UID: \"0e3d53a2-6105-4c2f-a74b-0b8dcd3b0706\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-wtwtl" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.523586 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/d2aac47a-b117-42c7-afa3-2c989d9536a5-image-import-ca\") pod \"apiserver-76f77b778f-5ptsh\" (UID: \"d2aac47a-b117-42c7-afa3-2c989d9536a5\") " pod="openshift-apiserver/apiserver-76f77b778f-5ptsh" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.524228 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/adfb805f-2f2e-40a1-91eb-59e6f23df192-config\") pod \"openshift-apiserver-operator-796bbdcf4f-gnd5s\" (UID: \"adfb805f-2f2e-40a1-91eb-59e6f23df192\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-gnd5s" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.524362 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/705b2456-4ba7-4775-84ca-4dcea64b6755-client-ca\") pod \"controller-manager-879f6c89f-l4zh5\" (UID: \"705b2456-4ba7-4775-84ca-4dcea64b6755\") " pod="openshift-controller-manager/controller-manager-879f6c89f-l4zh5" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.524884 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/59a1ff0e-6c38-4e97-8423-1641e5620570-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-7q5lc\" (UID: \"59a1ff0e-6c38-4e97-8423-1641e5620570\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-7q5lc" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.526039 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/65bbc5b9-f6f0-40ad-9e47-4122bb5cabab-console-config\") pod \"console-f9d7485db-9ql8j\" (UID: \"65bbc5b9-f6f0-40ad-9e47-4122bb5cabab\") " pod="openshift-console/console-f9d7485db-9ql8j" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.526877 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/df338eea-faa9-4dca-9c75-f12a2bbf87a4-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-dvhpg\" (UID: \"df338eea-faa9-4dca-9c75-f12a2bbf87a4\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-dvhpg" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.527460 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/159a2fad-3042-43e6-8df3-16f1157db0b1-serving-cert\") pod \"etcd-operator-b45778765-9cxt7\" (UID: \"159a2fad-3042-43e6-8df3-16f1157db0b1\") " pod="openshift-etcd-operator/etcd-operator-b45778765-9cxt7" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.531451 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/159a2fad-3042-43e6-8df3-16f1157db0b1-etcd-client\") pod \"etcd-operator-b45778765-9cxt7\" (UID: \"159a2fad-3042-43e6-8df3-16f1157db0b1\") " pod="openshift-etcd-operator/etcd-operator-b45778765-9cxt7" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.531714 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7f38918f-c551-4e3f-bb5e-d5dcac6ac826-serving-cert\") pod \"apiserver-7bbb656c7d-lt4dx\" (UID: \"7f38918f-c551-4e3f-bb5e-d5dcac6ac826\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lt4dx" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.531781 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/adfb805f-2f2e-40a1-91eb-59e6f23df192-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-gnd5s\" (UID: \"adfb805f-2f2e-40a1-91eb-59e6f23df192\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-gnd5s" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.531907 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/d2aac47a-b117-42c7-afa3-2c989d9536a5-encryption-config\") pod \"apiserver-76f77b778f-5ptsh\" (UID: \"d2aac47a-b117-42c7-afa3-2c989d9536a5\") " pod="openshift-apiserver/apiserver-76f77b778f-5ptsh" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.532258 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.533491 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/4828f51e-d209-4d31-a608-7d0951d311f4-machine-approver-tls\") pod \"machine-approver-56656f9798-gpd6h\" (UID: \"4828f51e-d209-4d31-a608-7d0951d311f4\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-gpd6h" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.536066 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/65bbc5b9-f6f0-40ad-9e47-4122bb5cabab-console-serving-cert\") pod \"console-f9d7485db-9ql8j\" (UID: \"65bbc5b9-f6f0-40ad-9e47-4122bb5cabab\") " pod="openshift-console/console-f9d7485db-9ql8j" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.536515 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.537171 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.538580 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/65bbc5b9-f6f0-40ad-9e47-4122bb5cabab-console-oauth-config\") pod \"console-f9d7485db-9ql8j\" (UID: \"65bbc5b9-f6f0-40ad-9e47-4122bb5cabab\") " pod="openshift-console/console-f9d7485db-9ql8j" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.538585 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/cf79ef17-ec75-4d0b-b665-ce24618a92df-metrics-tls\") pod \"dns-operator-744455d44c-s4vx5\" (UID: \"cf79ef17-ec75-4d0b-b665-ce24618a92df\") " pod="openshift-dns-operator/dns-operator-744455d44c-s4vx5" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.540655 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/d2aac47a-b117-42c7-afa3-2c989d9536a5-etcd-client\") pod \"apiserver-76f77b778f-5ptsh\" (UID: \"d2aac47a-b117-42c7-afa3-2c989d9536a5\") " pod="openshift-apiserver/apiserver-76f77b778f-5ptsh" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.541025 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/54725907-0303-4912-875b-7c9a533486fb-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-k9t5k\" (UID: \"54725907-0303-4912-875b-7c9a533486fb\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-k9t5k" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.541729 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/b35b7145-810b-4945-9819-6d9bb650ec73-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-n72r5\" (UID: \"b35b7145-810b-4945-9819-6d9bb650ec73\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-n72r5" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.542421 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.543008 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-hf8fp"] Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.543659 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-hf8fp" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.557632 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/7f38918f-c551-4e3f-bb5e-d5dcac6ac826-etcd-client\") pod \"apiserver-7bbb656c7d-lt4dx\" (UID: \"7f38918f-c551-4e3f-bb5e-d5dcac6ac826\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lt4dx" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.558144 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/7f38918f-c551-4e3f-bb5e-d5dcac6ac826-encryption-config\") pod \"apiserver-7bbb656c7d-lt4dx\" (UID: \"7f38918f-c551-4e3f-bb5e-d5dcac6ac826\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lt4dx" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.558490 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-dvhpg"] Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.561393 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.568403 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-l5xp9"] Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.569746 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-l5xp9" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.570635 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-vcwrd"] Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.571769 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-vcwrd" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.572085 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-n4dp7"] Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.575688 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-n4dp7" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.575940 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.576185 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-6sclt"] Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.587351 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29409630-45pn5"] Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.587519 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-6sclt" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.588446 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-dh427"] Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.589042 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29409630-45pn5" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.589491 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-l4zh5"] Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.589520 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-n72r5"] Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.589534 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-9ql8j"] Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.589549 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-lt4dx"] Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.589561 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-5ptsh"] Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.589573 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-j7v65"] Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.589664 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-dh427" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.590345 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-j7v65" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.592748 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-k9t5k"] Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.594161 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.594293 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-s4vx5"] Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.597009 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-5fgn9"] Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.598550 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-gnd5s"] Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.600021 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-2mb7d"] Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.600668 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ab324c2b-a49d-4874-b9c0-a869b226ce98-serving-cert\") pod \"openshift-config-operator-7777fb866f-hmh5v\" (UID: \"ab324c2b-a49d-4874-b9c0-a869b226ce98\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-hmh5v" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.600799 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vv2sl\" (UniqueName: \"kubernetes.io/projected/ab324c2b-a49d-4874-b9c0-a869b226ce98-kube-api-access-vv2sl\") pod \"openshift-config-operator-7777fb866f-hmh5v\" (UID: \"ab324c2b-a49d-4874-b9c0-a869b226ce98\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-hmh5v" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.601006 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/ab324c2b-a49d-4874-b9c0-a869b226ce98-available-featuregates\") pod \"openshift-config-operator-7777fb866f-hmh5v\" (UID: \"ab324c2b-a49d-4874-b9c0-a869b226ce98\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-hmh5v" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.601417 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/ab324c2b-a49d-4874-b9c0-a869b226ce98-available-featuregates\") pod \"openshift-config-operator-7777fb866f-hmh5v\" (UID: \"ab324c2b-a49d-4874-b9c0-a869b226ce98\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-hmh5v" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.601824 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-wqkxc"] Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.603357 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ab324c2b-a49d-4874-b9c0-a869b226ce98-serving-cert\") pod \"openshift-config-operator-7777fb866f-hmh5v\" (UID: \"ab324c2b-a49d-4874-b9c0-a869b226ce98\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-hmh5v" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.603737 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-vkbsg"] Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.605621 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-ct7w8"] Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.607474 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-wtwtl"] Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.608700 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-fxgbc"] Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.610113 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-49bkd"] Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.611724 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-f8kzc"] Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.611838 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-49bkd" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.612240 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-sjjm7"] Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.613236 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-qsskc"] Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.614227 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-lfm7c"] Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.614287 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.615409 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-7w2kz"] Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.616492 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-m8jlq"] Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.618247 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-ddnwl"] Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.618249 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-m8jlq" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.619186 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-ddnwl" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.619578 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-hmh5v"] Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.620989 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-hf8fp"] Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.622006 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-9cxt7"] Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.623152 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-mj64s"] Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.624290 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-vcwrd"] Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.625330 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-l5xp9"] Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.633270 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-7q5lc"] Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.634941 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.639307 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-dwkm7"] Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.643693 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-6sclt"] Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.646090 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-5z87f"] Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.647213 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-8274k"] Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.648479 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-m8jlq"] Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.649776 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-n4dp7"] Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.651354 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-jb992"] Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.652389 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29409630-45pn5"] Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.653671 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-j7v65"] Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.655422 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-dh427"] Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.656848 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-49bkd"] Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.730399 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.730414 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.734841 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.735498 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.738308 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.755308 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.779354 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.794766 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.814662 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.834515 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.840208 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xs4nq" event={"ID":"e70d6d2c-f0f3-4c8f-8213-efaa11de813b","Type":"ContainerStarted","Data":"9cf39601ba5a70ddec03314281aaf6a34282b76e02bd61bff4742667a8d942e2"} Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.840264 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xs4nq" event={"ID":"e70d6d2c-f0f3-4c8f-8213-efaa11de813b","Type":"ContainerStarted","Data":"4e9c469b364cac6887e1816cc40a325002f071e55a8af011c1cdf3b510a3a532"} Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.855274 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.874915 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.894807 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.914341 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.942232 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.954628 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.975173 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Dec 01 08:41:52 crc kubenswrapper[4813]: I1201 08:41:52.995727 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Dec 01 08:41:53 crc kubenswrapper[4813]: I1201 08:41:53.014464 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Dec 01 08:41:53 crc kubenswrapper[4813]: I1201 08:41:53.056076 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Dec 01 08:41:53 crc kubenswrapper[4813]: I1201 08:41:53.076275 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Dec 01 08:41:53 crc kubenswrapper[4813]: I1201 08:41:53.095290 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Dec 01 08:41:53 crc kubenswrapper[4813]: I1201 08:41:53.116096 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Dec 01 08:41:53 crc kubenswrapper[4813]: I1201 08:41:53.136459 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Dec 01 08:41:53 crc kubenswrapper[4813]: I1201 08:41:53.155647 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Dec 01 08:41:53 crc kubenswrapper[4813]: I1201 08:41:53.174928 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Dec 01 08:41:53 crc kubenswrapper[4813]: I1201 08:41:53.196038 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Dec 01 08:41:53 crc kubenswrapper[4813]: I1201 08:41:53.216365 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Dec 01 08:41:53 crc kubenswrapper[4813]: I1201 08:41:53.235770 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Dec 01 08:41:53 crc kubenswrapper[4813]: I1201 08:41:53.255028 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Dec 01 08:41:53 crc kubenswrapper[4813]: I1201 08:41:53.275212 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Dec 01 08:41:53 crc kubenswrapper[4813]: I1201 08:41:53.295557 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Dec 01 08:41:53 crc kubenswrapper[4813]: I1201 08:41:53.315701 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Dec 01 08:41:53 crc kubenswrapper[4813]: I1201 08:41:53.335102 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Dec 01 08:41:53 crc kubenswrapper[4813]: I1201 08:41:53.355582 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Dec 01 08:41:53 crc kubenswrapper[4813]: I1201 08:41:53.376654 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Dec 01 08:41:53 crc kubenswrapper[4813]: I1201 08:41:53.393165 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 08:41:53 crc kubenswrapper[4813]: I1201 08:41:53.395086 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Dec 01 08:41:53 crc kubenswrapper[4813]: I1201 08:41:53.426366 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Dec 01 08:41:53 crc kubenswrapper[4813]: I1201 08:41:53.434497 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Dec 01 08:41:53 crc kubenswrapper[4813]: I1201 08:41:53.466822 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Dec 01 08:41:53 crc kubenswrapper[4813]: I1201 08:41:53.475100 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Dec 01 08:41:53 crc kubenswrapper[4813]: I1201 08:41:53.505103 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Dec 01 08:41:53 crc kubenswrapper[4813]: I1201 08:41:53.513642 4813 request.go:700] Waited for 1.019982496s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-machine-api/secrets?fieldSelector=metadata.name%3Dcontrol-plane-machine-set-operator-tls&limit=500&resourceVersion=0 Dec 01 08:41:53 crc kubenswrapper[4813]: I1201 08:41:53.515554 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Dec 01 08:41:53 crc kubenswrapper[4813]: I1201 08:41:53.535259 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Dec 01 08:41:53 crc kubenswrapper[4813]: I1201 08:41:53.554780 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Dec 01 08:41:53 crc kubenswrapper[4813]: I1201 08:41:53.576256 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Dec 01 08:41:53 crc kubenswrapper[4813]: I1201 08:41:53.596303 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Dec 01 08:41:53 crc kubenswrapper[4813]: I1201 08:41:53.615325 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Dec 01 08:41:53 crc kubenswrapper[4813]: I1201 08:41:53.636684 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Dec 01 08:41:53 crc kubenswrapper[4813]: I1201 08:41:53.655362 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Dec 01 08:41:53 crc kubenswrapper[4813]: I1201 08:41:53.675835 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Dec 01 08:41:53 crc kubenswrapper[4813]: I1201 08:41:53.694916 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Dec 01 08:41:53 crc kubenswrapper[4813]: I1201 08:41:53.733792 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pjxkr\" (UniqueName: \"kubernetes.io/projected/705b2456-4ba7-4775-84ca-4dcea64b6755-kube-api-access-pjxkr\") pod \"controller-manager-879f6c89f-l4zh5\" (UID: \"705b2456-4ba7-4775-84ca-4dcea64b6755\") " pod="openshift-controller-manager/controller-manager-879f6c89f-l4zh5" Dec 01 08:41:53 crc kubenswrapper[4813]: I1201 08:41:53.758111 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kwqbn\" (UniqueName: \"kubernetes.io/projected/156d5b06-7b82-4444-817f-c7bf13cb3c44-kube-api-access-kwqbn\") pod \"ingress-operator-5b745b69d9-wqkxc\" (UID: \"156d5b06-7b82-4444-817f-c7bf13cb3c44\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-wqkxc" Dec 01 08:41:53 crc kubenswrapper[4813]: I1201 08:41:53.782570 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zpbjs\" (UniqueName: \"kubernetes.io/projected/65bbc5b9-f6f0-40ad-9e47-4122bb5cabab-kube-api-access-zpbjs\") pod \"console-f9d7485db-9ql8j\" (UID: \"65bbc5b9-f6f0-40ad-9e47-4122bb5cabab\") " pod="openshift-console/console-f9d7485db-9ql8j" Dec 01 08:41:53 crc kubenswrapper[4813]: I1201 08:41:53.782745 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-9ql8j" Dec 01 08:41:53 crc kubenswrapper[4813]: I1201 08:41:53.796446 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9kdgg\" (UniqueName: \"kubernetes.io/projected/54725907-0303-4912-875b-7c9a533486fb-kube-api-access-9kdgg\") pod \"authentication-operator-69f744f599-k9t5k\" (UID: \"54725907-0303-4912-875b-7c9a533486fb\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-k9t5k" Dec 01 08:41:53 crc kubenswrapper[4813]: I1201 08:41:53.811743 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vgfgc\" (UniqueName: \"kubernetes.io/projected/adfb805f-2f2e-40a1-91eb-59e6f23df192-kube-api-access-vgfgc\") pod \"openshift-apiserver-operator-796bbdcf4f-gnd5s\" (UID: \"adfb805f-2f2e-40a1-91eb-59e6f23df192\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-gnd5s" Dec 01 08:41:53 crc kubenswrapper[4813]: I1201 08:41:53.829585 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7s96q\" (UniqueName: \"kubernetes.io/projected/4828f51e-d209-4d31-a608-7d0951d311f4-kube-api-access-7s96q\") pod \"machine-approver-56656f9798-gpd6h\" (UID: \"4828f51e-d209-4d31-a608-7d0951d311f4\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-gpd6h" Dec 01 08:41:53 crc kubenswrapper[4813]: I1201 08:41:53.849073 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pndx7\" (UniqueName: \"kubernetes.io/projected/0e3d53a2-6105-4c2f-a74b-0b8dcd3b0706-kube-api-access-pndx7\") pod \"cluster-image-registry-operator-dc59b4c8b-wtwtl\" (UID: \"0e3d53a2-6105-4c2f-a74b-0b8dcd3b0706\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-wtwtl" Dec 01 08:41:53 crc kubenswrapper[4813]: I1201 08:41:53.849286 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-k9t5k" Dec 01 08:41:53 crc kubenswrapper[4813]: I1201 08:41:53.853015 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-gnd5s" Dec 01 08:41:53 crc kubenswrapper[4813]: I1201 08:41:53.868883 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8ljvz\" (UniqueName: \"kubernetes.io/projected/cf79ef17-ec75-4d0b-b665-ce24618a92df-kube-api-access-8ljvz\") pod \"dns-operator-744455d44c-s4vx5\" (UID: \"cf79ef17-ec75-4d0b-b665-ce24618a92df\") " pod="openshift-dns-operator/dns-operator-744455d44c-s4vx5" Dec 01 08:41:53 crc kubenswrapper[4813]: I1201 08:41:53.893111 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gdrs4\" (UniqueName: \"kubernetes.io/projected/159a2fad-3042-43e6-8df3-16f1157db0b1-kube-api-access-gdrs4\") pod \"etcd-operator-b45778765-9cxt7\" (UID: \"159a2fad-3042-43e6-8df3-16f1157db0b1\") " pod="openshift-etcd-operator/etcd-operator-b45778765-9cxt7" Dec 01 08:41:53 crc kubenswrapper[4813]: I1201 08:41:53.907889 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tm5gl\" (UniqueName: \"kubernetes.io/projected/e17c11d6-eb41-4391-b513-a817eeffcdb4-kube-api-access-tm5gl\") pod \"downloads-7954f5f757-ct7w8\" (UID: \"e17c11d6-eb41-4391-b513-a817eeffcdb4\") " pod="openshift-console/downloads-7954f5f757-ct7w8" Dec 01 08:41:53 crc kubenswrapper[4813]: I1201 08:41:53.933448 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/156d5b06-7b82-4444-817f-c7bf13cb3c44-bound-sa-token\") pod \"ingress-operator-5b745b69d9-wqkxc\" (UID: \"156d5b06-7b82-4444-817f-c7bf13cb3c44\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-wqkxc" Dec 01 08:41:53 crc kubenswrapper[4813]: I1201 08:41:53.952288 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n4h95\" (UniqueName: \"kubernetes.io/projected/d2aac47a-b117-42c7-afa3-2c989d9536a5-kube-api-access-n4h95\") pod \"apiserver-76f77b778f-5ptsh\" (UID: \"d2aac47a-b117-42c7-afa3-2c989d9536a5\") " pod="openshift-apiserver/apiserver-76f77b778f-5ptsh" Dec 01 08:41:53 crc kubenswrapper[4813]: I1201 08:41:53.954660 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Dec 01 08:41:53 crc kubenswrapper[4813]: I1201 08:41:53.975245 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Dec 01 08:41:53 crc kubenswrapper[4813]: I1201 08:41:53.995199 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Dec 01 08:41:54 crc kubenswrapper[4813]: I1201 08:41:54.006925 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-l4zh5" Dec 01 08:41:54 crc kubenswrapper[4813]: I1201 08:41:54.020508 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Dec 01 08:41:54 crc kubenswrapper[4813]: I1201 08:41:54.035079 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Dec 01 08:41:54 crc kubenswrapper[4813]: I1201 08:41:54.055813 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Dec 01 08:41:54 crc kubenswrapper[4813]: I1201 08:41:54.075824 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Dec 01 08:41:54 crc kubenswrapper[4813]: I1201 08:41:54.091558 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-gpd6h" Dec 01 08:41:54 crc kubenswrapper[4813]: I1201 08:41:54.095111 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Dec 01 08:41:54 crc kubenswrapper[4813]: I1201 08:41:54.097446 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-5ptsh" Dec 01 08:41:54 crc kubenswrapper[4813]: I1201 08:41:54.110751 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-s4vx5" Dec 01 08:41:54 crc kubenswrapper[4813]: W1201 08:41:54.112812 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4828f51e_d209_4d31_a608_7d0951d311f4.slice/crio-800dfc8eb89391bec38cba17bd70e401cf23876a6e76c80d7b30944739102ec5 WatchSource:0}: Error finding container 800dfc8eb89391bec38cba17bd70e401cf23876a6e76c80d7b30944739102ec5: Status 404 returned error can't find the container with id 800dfc8eb89391bec38cba17bd70e401cf23876a6e76c80d7b30944739102ec5 Dec 01 08:41:54 crc kubenswrapper[4813]: I1201 08:41:54.122018 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-9cxt7" Dec 01 08:41:54 crc kubenswrapper[4813]: I1201 08:41:54.134445 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-ct7w8" Dec 01 08:41:54 crc kubenswrapper[4813]: I1201 08:41:54.137924 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zlhdb\" (UniqueName: \"kubernetes.io/projected/59a1ff0e-6c38-4e97-8423-1641e5620570-kube-api-access-zlhdb\") pod \"kube-storage-version-migrator-operator-b67b599dd-7q5lc\" (UID: \"59a1ff0e-6c38-4e97-8423-1641e5620570\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-7q5lc" Dec 01 08:41:54 crc kubenswrapper[4813]: I1201 08:41:54.142143 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-wqkxc" Dec 01 08:41:54 crc kubenswrapper[4813]: I1201 08:41:54.156912 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n6sss\" (UniqueName: \"kubernetes.io/projected/3c957bbc-fa1d-4188-9f9f-0b67be4c56c7-kube-api-access-n6sss\") pod \"openshift-controller-manager-operator-756b6f6bc6-vkbsg\" (UID: \"3c957bbc-fa1d-4188-9f9f-0b67be4c56c7\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-vkbsg" Dec 01 08:41:54 crc kubenswrapper[4813]: I1201 08:41:54.163893 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-k9t5k"] Dec 01 08:41:54 crc kubenswrapper[4813]: I1201 08:41:54.174131 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cdxmn\" (UniqueName: \"kubernetes.io/projected/b35b7145-810b-4945-9819-6d9bb650ec73-kube-api-access-cdxmn\") pod \"machine-api-operator-5694c8668f-n72r5\" (UID: \"b35b7145-810b-4945-9819-6d9bb650ec73\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-n72r5" Dec 01 08:41:54 crc kubenswrapper[4813]: I1201 08:41:54.179752 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-gnd5s"] Dec 01 08:41:54 crc kubenswrapper[4813]: I1201 08:41:54.213536 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j8cxb\" (UniqueName: \"kubernetes.io/projected/df338eea-faa9-4dca-9c75-f12a2bbf87a4-kube-api-access-j8cxb\") pod \"cluster-samples-operator-665b6dd947-dvhpg\" (UID: \"df338eea-faa9-4dca-9c75-f12a2bbf87a4\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-dvhpg" Dec 01 08:41:54 crc kubenswrapper[4813]: I1201 08:41:54.228474 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qmxx7\" (UniqueName: \"kubernetes.io/projected/7f38918f-c551-4e3f-bb5e-d5dcac6ac826-kube-api-access-qmxx7\") pod \"apiserver-7bbb656c7d-lt4dx\" (UID: \"7f38918f-c551-4e3f-bb5e-d5dcac6ac826\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lt4dx" Dec 01 08:41:54 crc kubenswrapper[4813]: I1201 08:41:54.234361 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/0e3d53a2-6105-4c2f-a74b-0b8dcd3b0706-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-wtwtl\" (UID: \"0e3d53a2-6105-4c2f-a74b-0b8dcd3b0706\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-wtwtl" Dec 01 08:41:54 crc kubenswrapper[4813]: I1201 08:41:54.256336 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-dvhpg" Dec 01 08:41:54 crc kubenswrapper[4813]: I1201 08:41:54.258682 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Dec 01 08:41:54 crc kubenswrapper[4813]: I1201 08:41:54.275742 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Dec 01 08:41:54 crc kubenswrapper[4813]: I1201 08:41:54.293323 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-n72r5" Dec 01 08:41:54 crc kubenswrapper[4813]: I1201 08:41:54.294866 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Dec 01 08:41:54 crc kubenswrapper[4813]: I1201 08:41:54.317231 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-l4zh5"] Dec 01 08:41:54 crc kubenswrapper[4813]: I1201 08:41:54.325894 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Dec 01 08:41:54 crc kubenswrapper[4813]: I1201 08:41:54.337039 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-9ql8j"] Dec 01 08:41:54 crc kubenswrapper[4813]: I1201 08:41:54.341779 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Dec 01 08:41:54 crc kubenswrapper[4813]: I1201 08:41:54.355233 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Dec 01 08:41:54 crc kubenswrapper[4813]: I1201 08:41:54.355623 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-s4vx5"] Dec 01 08:41:54 crc kubenswrapper[4813]: I1201 08:41:54.367982 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-vkbsg" Dec 01 08:41:54 crc kubenswrapper[4813]: I1201 08:41:54.368542 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lt4dx" Dec 01 08:41:54 crc kubenswrapper[4813]: I1201 08:41:54.371510 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-5ptsh"] Dec 01 08:41:54 crc kubenswrapper[4813]: I1201 08:41:54.374331 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-wtwtl" Dec 01 08:41:54 crc kubenswrapper[4813]: I1201 08:41:54.375430 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Dec 01 08:41:54 crc kubenswrapper[4813]: I1201 08:41:54.407205 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-7q5lc" Dec 01 08:41:54 crc kubenswrapper[4813]: I1201 08:41:54.412152 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Dec 01 08:41:54 crc kubenswrapper[4813]: I1201 08:41:54.414373 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Dec 01 08:41:54 crc kubenswrapper[4813]: I1201 08:41:54.454661 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Dec 01 08:41:54 crc kubenswrapper[4813]: I1201 08:41:54.458400 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Dec 01 08:41:54 crc kubenswrapper[4813]: I1201 08:41:54.477272 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 01 08:41:54 crc kubenswrapper[4813]: I1201 08:41:54.497480 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 01 08:41:54 crc kubenswrapper[4813]: I1201 08:41:54.513634 4813 request.go:700] Waited for 1.92370868s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-service-ca-operator/configmaps?fieldSelector=metadata.name%3Dservice-ca-operator-config&limit=500&resourceVersion=0 Dec 01 08:41:54 crc kubenswrapper[4813]: I1201 08:41:54.514711 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Dec 01 08:41:54 crc kubenswrapper[4813]: I1201 08:41:54.537688 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Dec 01 08:41:54 crc kubenswrapper[4813]: I1201 08:41:54.594680 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Dec 01 08:41:54 crc kubenswrapper[4813]: I1201 08:41:54.594957 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Dec 01 08:41:54 crc kubenswrapper[4813]: I1201 08:41:54.600305 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Dec 01 08:41:54 crc kubenswrapper[4813]: I1201 08:41:54.615266 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Dec 01 08:41:54 crc kubenswrapper[4813]: I1201 08:41:54.636118 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Dec 01 08:41:54 crc kubenswrapper[4813]: I1201 08:41:54.657990 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Dec 01 08:41:54 crc kubenswrapper[4813]: I1201 08:41:54.688758 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-n72r5"] Dec 01 08:41:54 crc kubenswrapper[4813]: I1201 08:41:54.693010 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-ct7w8"] Dec 01 08:41:54 crc kubenswrapper[4813]: I1201 08:41:54.696550 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Dec 01 08:41:54 crc kubenswrapper[4813]: I1201 08:41:54.698493 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vv2sl\" (UniqueName: \"kubernetes.io/projected/ab324c2b-a49d-4874-b9c0-a869b226ce98-kube-api-access-vv2sl\") pod \"openshift-config-operator-7777fb866f-hmh5v\" (UID: \"ab324c2b-a49d-4874-b9c0-a869b226ce98\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-hmh5v" Dec 01 08:41:54 crc kubenswrapper[4813]: W1201 08:41:54.702219 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb35b7145_810b_4945_9819_6d9bb650ec73.slice/crio-da0c2028a04fb545c0ef29f7c66f7e86fb18cea2f94f26dc9a40698e5c4d1b08 WatchSource:0}: Error finding container da0c2028a04fb545c0ef29f7c66f7e86fb18cea2f94f26dc9a40698e5c4d1b08: Status 404 returned error can't find the container with id da0c2028a04fb545c0ef29f7c66f7e86fb18cea2f94f26dc9a40698e5c4d1b08 Dec 01 08:41:54 crc kubenswrapper[4813]: I1201 08:41:54.703617 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-9cxt7"] Dec 01 08:41:54 crc kubenswrapper[4813]: I1201 08:41:54.714129 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Dec 01 08:41:54 crc kubenswrapper[4813]: I1201 08:41:54.715889 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-dvhpg"] Dec 01 08:41:54 crc kubenswrapper[4813]: I1201 08:41:54.736748 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Dec 01 08:41:54 crc kubenswrapper[4813]: I1201 08:41:54.754951 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Dec 01 08:41:54 crc kubenswrapper[4813]: I1201 08:41:54.758278 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-hmh5v" Dec 01 08:41:54 crc kubenswrapper[4813]: I1201 08:41:54.776793 4813 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Dec 01 08:41:54 crc kubenswrapper[4813]: I1201 08:41:54.798459 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Dec 01 08:41:54 crc kubenswrapper[4813]: I1201 08:41:54.815659 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Dec 01 08:41:54 crc kubenswrapper[4813]: I1201 08:41:54.835758 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Dec 01 08:41:54 crc kubenswrapper[4813]: I1201 08:41:54.838474 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-wqkxc"] Dec 01 08:41:54 crc kubenswrapper[4813]: I1201 08:41:54.854614 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Dec 01 08:41:54 crc kubenswrapper[4813]: I1201 08:41:54.857293 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-n72r5" event={"ID":"b35b7145-810b-4945-9819-6d9bb650ec73","Type":"ContainerStarted","Data":"da0c2028a04fb545c0ef29f7c66f7e86fb18cea2f94f26dc9a40698e5c4d1b08"} Dec 01 08:41:54 crc kubenswrapper[4813]: I1201 08:41:54.862704 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-k9t5k" event={"ID":"54725907-0303-4912-875b-7c9a533486fb","Type":"ContainerStarted","Data":"2c3ab887b46abceada02569829c8a8b711fbd584ef5dcf330fe9c331c4171fc9"} Dec 01 08:41:54 crc kubenswrapper[4813]: I1201 08:41:54.862755 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-k9t5k" event={"ID":"54725907-0303-4912-875b-7c9a533486fb","Type":"ContainerStarted","Data":"49dde413f781ff23d9c1f23ea6b884d02c1652568d3caf169373cfd47a7c7065"} Dec 01 08:41:54 crc kubenswrapper[4813]: I1201 08:41:54.871457 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-gpd6h" event={"ID":"4828f51e-d209-4d31-a608-7d0951d311f4","Type":"ContainerStarted","Data":"30c0bb9af1fda6b84d6ae4449c8ab29af7ca389a7a141c3749ac958a92adf03d"} Dec 01 08:41:54 crc kubenswrapper[4813]: I1201 08:41:54.871508 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-gpd6h" event={"ID":"4828f51e-d209-4d31-a608-7d0951d311f4","Type":"ContainerStarted","Data":"800dfc8eb89391bec38cba17bd70e401cf23876a6e76c80d7b30944739102ec5"} Dec 01 08:41:54 crc kubenswrapper[4813]: I1201 08:41:54.874603 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-gnd5s" event={"ID":"adfb805f-2f2e-40a1-91eb-59e6f23df192","Type":"ContainerStarted","Data":"ab4fba3ebc92cc9dab0574aef456166f32a49d3b9c97eaeafbae703f9c444e45"} Dec 01 08:41:54 crc kubenswrapper[4813]: I1201 08:41:54.874626 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-gnd5s" event={"ID":"adfb805f-2f2e-40a1-91eb-59e6f23df192","Type":"ContainerStarted","Data":"fb507e52c81c1a70753b269d61a3ed428ce2be45d682ef2e5f5e563346c4a066"} Dec 01 08:41:54 crc kubenswrapper[4813]: I1201 08:41:54.875509 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Dec 01 08:41:54 crc kubenswrapper[4813]: I1201 08:41:54.880370 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-s4vx5" event={"ID":"cf79ef17-ec75-4d0b-b665-ce24618a92df","Type":"ContainerStarted","Data":"e0c61c8417e84e7cea9d6cae722b35e1507a69eaa4a85c617f3d909cf17d02c7"} Dec 01 08:41:54 crc kubenswrapper[4813]: W1201 08:41:54.898933 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod156d5b06_7b82_4444_817f_c7bf13cb3c44.slice/crio-99624914f2863f51592983e15a317b4fe132cf3ec663077b4534638f9320c7c9 WatchSource:0}: Error finding container 99624914f2863f51592983e15a317b4fe132cf3ec663077b4534638f9320c7c9: Status 404 returned error can't find the container with id 99624914f2863f51592983e15a317b4fe132cf3ec663077b4534638f9320c7c9 Dec 01 08:41:54 crc kubenswrapper[4813]: I1201 08:41:54.900581 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-l4zh5" event={"ID":"705b2456-4ba7-4775-84ca-4dcea64b6755","Type":"ContainerStarted","Data":"2e0e06d86b703d0484153ad01443faf32d8bb1a0bbbf5e6ecbecc5c8847cce3f"} Dec 01 08:41:54 crc kubenswrapper[4813]: I1201 08:41:54.900639 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-l4zh5" event={"ID":"705b2456-4ba7-4775-84ca-4dcea64b6755","Type":"ContainerStarted","Data":"7301dd69d74eda5bd74707420800383217ecf5e7c3028a1e77774f282ad69285"} Dec 01 08:41:54 crc kubenswrapper[4813]: I1201 08:41:54.900809 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-l4zh5" Dec 01 08:41:54 crc kubenswrapper[4813]: I1201 08:41:54.904602 4813 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-l4zh5 container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.8:8443/healthz\": dial tcp 10.217.0.8:8443: connect: connection refused" start-of-body= Dec 01 08:41:54 crc kubenswrapper[4813]: I1201 08:41:54.904665 4813 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-l4zh5" podUID="705b2456-4ba7-4775-84ca-4dcea64b6755" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.8:8443/healthz\": dial tcp 10.217.0.8:8443: connect: connection refused" Dec 01 08:41:54 crc kubenswrapper[4813]: I1201 08:41:54.905651 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-9ql8j" event={"ID":"65bbc5b9-f6f0-40ad-9e47-4122bb5cabab","Type":"ContainerStarted","Data":"493ae0d0e18a7b60abbe8ca0ced9a48482559aec2aa28cb34773df2a95cba479"} Dec 01 08:41:54 crc kubenswrapper[4813]: I1201 08:41:54.905694 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-9ql8j" event={"ID":"65bbc5b9-f6f0-40ad-9e47-4122bb5cabab","Type":"ContainerStarted","Data":"c30920a025b030952269dc79f0043cdf088400d852d382596157d322a3b24b17"} Dec 01 08:41:54 crc kubenswrapper[4813]: I1201 08:41:54.919305 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-ct7w8" event={"ID":"e17c11d6-eb41-4391-b513-a817eeffcdb4","Type":"ContainerStarted","Data":"21eba6106a69adf93c62b9718c50684e32230de0dc52b6b9975c3997b16b4dfb"} Dec 01 08:41:54 crc kubenswrapper[4813]: I1201 08:41:54.923009 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-9cxt7" event={"ID":"159a2fad-3042-43e6-8df3-16f1157db0b1","Type":"ContainerStarted","Data":"ad37127f043a6bfd371c627d46c8b14b97324378786b31d38381d7435258d51d"} Dec 01 08:41:54 crc kubenswrapper[4813]: I1201 08:41:54.940688 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-5ptsh" event={"ID":"d2aac47a-b117-42c7-afa3-2c989d9536a5","Type":"ContainerStarted","Data":"2ad1a5521a3b53bc19a04297e95ef7226ef71a33b0e3633f94e1c42ca0b90836"} Dec 01 08:41:54 crc kubenswrapper[4813]: I1201 08:41:54.965289 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a2b231fa-0ecc-44da-9259-a7dc7261127c-bound-sa-token\") pod \"image-registry-697d97f7c8-mj64s\" (UID: \"a2b231fa-0ecc-44da-9259-a7dc7261127c\") " pod="openshift-image-registry/image-registry-697d97f7c8-mj64s" Dec 01 08:41:54 crc kubenswrapper[4813]: I1201 08:41:54.965359 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lmpxp\" (UniqueName: \"kubernetes.io/projected/52120ed5-ff4f-4df5-8c60-3342df484be8-kube-api-access-lmpxp\") pod \"oauth-openshift-558db77b4-5fgn9\" (UID: \"52120ed5-ff4f-4df5-8c60-3342df484be8\") " pod="openshift-authentication/oauth-openshift-558db77b4-5fgn9" Dec 01 08:41:54 crc kubenswrapper[4813]: I1201 08:41:54.965422 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8b7d482d-6638-41f3-93b3-0b6aebf7401b-config\") pod \"console-operator-58897d9998-7w2kz\" (UID: \"8b7d482d-6638-41f3-93b3-0b6aebf7401b\") " pod="openshift-console-operator/console-operator-58897d9998-7w2kz" Dec 01 08:41:54 crc kubenswrapper[4813]: I1201 08:41:54.965440 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0f0340e8-faa5-447b-a5cd-a878e982d98f-serving-cert\") pod \"route-controller-manager-6576b87f9c-2mb7d\" (UID: \"0f0340e8-faa5-447b-a5cd-a878e982d98f\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-2mb7d" Dec 01 08:41:54 crc kubenswrapper[4813]: I1201 08:41:54.965458 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/52120ed5-ff4f-4df5-8c60-3342df484be8-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-5fgn9\" (UID: \"52120ed5-ff4f-4df5-8c60-3342df484be8\") " pod="openshift-authentication/oauth-openshift-558db77b4-5fgn9" Dec 01 08:41:54 crc kubenswrapper[4813]: I1201 08:41:54.965523 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a2b231fa-0ecc-44da-9259-a7dc7261127c-trusted-ca\") pod \"image-registry-697d97f7c8-mj64s\" (UID: \"a2b231fa-0ecc-44da-9259-a7dc7261127c\") " pod="openshift-image-registry/image-registry-697d97f7c8-mj64s" Dec 01 08:41:54 crc kubenswrapper[4813]: I1201 08:41:54.965573 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b8p9d\" (UniqueName: \"kubernetes.io/projected/8b7d482d-6638-41f3-93b3-0b6aebf7401b-kube-api-access-b8p9d\") pod \"console-operator-58897d9998-7w2kz\" (UID: \"8b7d482d-6638-41f3-93b3-0b6aebf7401b\") " pod="openshift-console-operator/console-operator-58897d9998-7w2kz" Dec 01 08:41:54 crc kubenswrapper[4813]: I1201 08:41:54.965610 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mj64s\" (UID: \"a2b231fa-0ecc-44da-9259-a7dc7261127c\") " pod="openshift-image-registry/image-registry-697d97f7c8-mj64s" Dec 01 08:41:54 crc kubenswrapper[4813]: I1201 08:41:54.965653 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/a2b231fa-0ecc-44da-9259-a7dc7261127c-registry-certificates\") pod \"image-registry-697d97f7c8-mj64s\" (UID: \"a2b231fa-0ecc-44da-9259-a7dc7261127c\") " pod="openshift-image-registry/image-registry-697d97f7c8-mj64s" Dec 01 08:41:54 crc kubenswrapper[4813]: I1201 08:41:54.965677 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/52120ed5-ff4f-4df5-8c60-3342df484be8-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-5fgn9\" (UID: \"52120ed5-ff4f-4df5-8c60-3342df484be8\") " pod="openshift-authentication/oauth-openshift-558db77b4-5fgn9" Dec 01 08:41:54 crc kubenswrapper[4813]: I1201 08:41:54.965692 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/52120ed5-ff4f-4df5-8c60-3342df484be8-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-5fgn9\" (UID: \"52120ed5-ff4f-4df5-8c60-3342df484be8\") " pod="openshift-authentication/oauth-openshift-558db77b4-5fgn9" Dec 01 08:41:54 crc kubenswrapper[4813]: I1201 08:41:54.965724 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8b7d482d-6638-41f3-93b3-0b6aebf7401b-serving-cert\") pod \"console-operator-58897d9998-7w2kz\" (UID: \"8b7d482d-6638-41f3-93b3-0b6aebf7401b\") " pod="openshift-console-operator/console-operator-58897d9998-7w2kz" Dec 01 08:41:54 crc kubenswrapper[4813]: I1201 08:41:54.965739 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/52120ed5-ff4f-4df5-8c60-3342df484be8-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-5fgn9\" (UID: \"52120ed5-ff4f-4df5-8c60-3342df484be8\") " pod="openshift-authentication/oauth-openshift-558db77b4-5fgn9" Dec 01 08:41:54 crc kubenswrapper[4813]: I1201 08:41:54.965770 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/52120ed5-ff4f-4df5-8c60-3342df484be8-audit-dir\") pod \"oauth-openshift-558db77b4-5fgn9\" (UID: \"52120ed5-ff4f-4df5-8c60-3342df484be8\") " pod="openshift-authentication/oauth-openshift-558db77b4-5fgn9" Dec 01 08:41:54 crc kubenswrapper[4813]: I1201 08:41:54.965800 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/52120ed5-ff4f-4df5-8c60-3342df484be8-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-5fgn9\" (UID: \"52120ed5-ff4f-4df5-8c60-3342df484be8\") " pod="openshift-authentication/oauth-openshift-558db77b4-5fgn9" Dec 01 08:41:54 crc kubenswrapper[4813]: I1201 08:41:54.965815 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/52120ed5-ff4f-4df5-8c60-3342df484be8-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-5fgn9\" (UID: \"52120ed5-ff4f-4df5-8c60-3342df484be8\") " pod="openshift-authentication/oauth-openshift-558db77b4-5fgn9" Dec 01 08:41:54 crc kubenswrapper[4813]: I1201 08:41:54.965847 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/57c3bc40-b325-4343-a01b-6705e9e8eb77-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-8274k\" (UID: \"57c3bc40-b325-4343-a01b-6705e9e8eb77\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-8274k" Dec 01 08:41:54 crc kubenswrapper[4813]: I1201 08:41:54.965886 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0f0340e8-faa5-447b-a5cd-a878e982d98f-config\") pod \"route-controller-manager-6576b87f9c-2mb7d\" (UID: \"0f0340e8-faa5-447b-a5cd-a878e982d98f\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-2mb7d" Dec 01 08:41:54 crc kubenswrapper[4813]: I1201 08:41:54.965910 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/52120ed5-ff4f-4df5-8c60-3342df484be8-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-5fgn9\" (UID: \"52120ed5-ff4f-4df5-8c60-3342df484be8\") " pod="openshift-authentication/oauth-openshift-558db77b4-5fgn9" Dec 01 08:41:54 crc kubenswrapper[4813]: I1201 08:41:54.965985 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/a2b231fa-0ecc-44da-9259-a7dc7261127c-registry-tls\") pod \"image-registry-697d97f7c8-mj64s\" (UID: \"a2b231fa-0ecc-44da-9259-a7dc7261127c\") " pod="openshift-image-registry/image-registry-697d97f7c8-mj64s" Dec 01 08:41:54 crc kubenswrapper[4813]: I1201 08:41:54.966001 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/52120ed5-ff4f-4df5-8c60-3342df484be8-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-5fgn9\" (UID: \"52120ed5-ff4f-4df5-8c60-3342df484be8\") " pod="openshift-authentication/oauth-openshift-558db77b4-5fgn9" Dec 01 08:41:54 crc kubenswrapper[4813]: I1201 08:41:54.966106 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/a2b231fa-0ecc-44da-9259-a7dc7261127c-installation-pull-secrets\") pod \"image-registry-697d97f7c8-mj64s\" (UID: \"a2b231fa-0ecc-44da-9259-a7dc7261127c\") " pod="openshift-image-registry/image-registry-697d97f7c8-mj64s" Dec 01 08:41:54 crc kubenswrapper[4813]: I1201 08:41:54.966122 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/52120ed5-ff4f-4df5-8c60-3342df484be8-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-5fgn9\" (UID: \"52120ed5-ff4f-4df5-8c60-3342df484be8\") " pod="openshift-authentication/oauth-openshift-558db77b4-5fgn9" Dec 01 08:41:54 crc kubenswrapper[4813]: I1201 08:41:54.966150 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/57c3bc40-b325-4343-a01b-6705e9e8eb77-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-8274k\" (UID: \"57c3bc40-b325-4343-a01b-6705e9e8eb77\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-8274k" Dec 01 08:41:54 crc kubenswrapper[4813]: I1201 08:41:54.966176 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/52120ed5-ff4f-4df5-8c60-3342df484be8-audit-policies\") pod \"oauth-openshift-558db77b4-5fgn9\" (UID: \"52120ed5-ff4f-4df5-8c60-3342df484be8\") " pod="openshift-authentication/oauth-openshift-558db77b4-5fgn9" Dec 01 08:41:54 crc kubenswrapper[4813]: I1201 08:41:54.966191 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-72d4v\" (UniqueName: \"kubernetes.io/projected/0f0340e8-faa5-447b-a5cd-a878e982d98f-kube-api-access-72d4v\") pod \"route-controller-manager-6576b87f9c-2mb7d\" (UID: \"0f0340e8-faa5-447b-a5cd-a878e982d98f\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-2mb7d" Dec 01 08:41:54 crc kubenswrapper[4813]: I1201 08:41:54.966238 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/a2b231fa-0ecc-44da-9259-a7dc7261127c-ca-trust-extracted\") pod \"image-registry-697d97f7c8-mj64s\" (UID: \"a2b231fa-0ecc-44da-9259-a7dc7261127c\") " pod="openshift-image-registry/image-registry-697d97f7c8-mj64s" Dec 01 08:41:54 crc kubenswrapper[4813]: I1201 08:41:54.966253 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9p9xx\" (UniqueName: \"kubernetes.io/projected/a2b231fa-0ecc-44da-9259-a7dc7261127c-kube-api-access-9p9xx\") pod \"image-registry-697d97f7c8-mj64s\" (UID: \"a2b231fa-0ecc-44da-9259-a7dc7261127c\") " pod="openshift-image-registry/image-registry-697d97f7c8-mj64s" Dec 01 08:41:54 crc kubenswrapper[4813]: I1201 08:41:54.966267 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8b7d482d-6638-41f3-93b3-0b6aebf7401b-trusted-ca\") pod \"console-operator-58897d9998-7w2kz\" (UID: \"8b7d482d-6638-41f3-93b3-0b6aebf7401b\") " pod="openshift-console-operator/console-operator-58897d9998-7w2kz" Dec 01 08:41:54 crc kubenswrapper[4813]: I1201 08:41:54.966439 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/52120ed5-ff4f-4df5-8c60-3342df484be8-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-5fgn9\" (UID: \"52120ed5-ff4f-4df5-8c60-3342df484be8\") " pod="openshift-authentication/oauth-openshift-558db77b4-5fgn9" Dec 01 08:41:54 crc kubenswrapper[4813]: I1201 08:41:54.966672 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/52120ed5-ff4f-4df5-8c60-3342df484be8-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-5fgn9\" (UID: \"52120ed5-ff4f-4df5-8c60-3342df484be8\") " pod="openshift-authentication/oauth-openshift-558db77b4-5fgn9" Dec 01 08:41:54 crc kubenswrapper[4813]: I1201 08:41:54.966719 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/0f0340e8-faa5-447b-a5cd-a878e982d98f-client-ca\") pod \"route-controller-manager-6576b87f9c-2mb7d\" (UID: \"0f0340e8-faa5-447b-a5cd-a878e982d98f\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-2mb7d" Dec 01 08:41:54 crc kubenswrapper[4813]: I1201 08:41:54.966738 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/57c3bc40-b325-4343-a01b-6705e9e8eb77-config\") pod \"kube-controller-manager-operator-78b949d7b-8274k\" (UID: \"57c3bc40-b325-4343-a01b-6705e9e8eb77\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-8274k" Dec 01 08:41:54 crc kubenswrapper[4813]: E1201 08:41:54.966784 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 08:41:55.466756027 +0000 UTC m=+100.089677683 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mj64s" (UID: "a2b231fa-0ecc-44da-9259-a7dc7261127c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.013871 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-lt4dx"] Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.037253 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-vkbsg"] Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.068050 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 08:41:55 crc kubenswrapper[4813]: E1201 08:41:55.068144 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 08:41:55.56812331 +0000 UTC m=+100.191044896 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.068430 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-66c2t\" (UniqueName: \"kubernetes.io/projected/489e9f98-0728-4a72-80d4-3d88a77823f5-kube-api-access-66c2t\") pod \"olm-operator-6b444d44fb-hf8fp\" (UID: \"489e9f98-0728-4a72-80d4-3d88a77823f5\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-hf8fp" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.068465 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/52120ed5-ff4f-4df5-8c60-3342df484be8-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-5fgn9\" (UID: \"52120ed5-ff4f-4df5-8c60-3342df484be8\") " pod="openshift-authentication/oauth-openshift-558db77b4-5fgn9" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.068496 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/7d654f9a-b190-47bc-89ce-0caae2ec1513-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-jb992\" (UID: \"7d654f9a-b190-47bc-89ce-0caae2ec1513\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-jb992" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.068538 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0f0340e8-faa5-447b-a5cd-a878e982d98f-config\") pod \"route-controller-manager-6576b87f9c-2mb7d\" (UID: \"0f0340e8-faa5-447b-a5cd-a878e982d98f\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-2mb7d" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.068558 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/57a1918e-f28a-4890-b88c-bc80df5cf704-serving-cert\") pod \"service-ca-operator-777779d784-dh427\" (UID: \"57a1918e-f28a-4890-b88c-bc80df5cf704\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-dh427" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.068574 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/52120ed5-ff4f-4df5-8c60-3342df484be8-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-5fgn9\" (UID: \"52120ed5-ff4f-4df5-8c60-3342df484be8\") " pod="openshift-authentication/oauth-openshift-558db77b4-5fgn9" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.068593 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/469840fe-2ac3-4431-9b2b-3b41451e6cbc-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-lfm7c\" (UID: \"469840fe-2ac3-4431-9b2b-3b41451e6cbc\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-lfm7c" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.068613 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/2e3f10d2-ac94-459c-9898-542e9f710747-webhook-cert\") pod \"packageserver-d55dfcdfc-f8kzc\" (UID: \"2e3f10d2-ac94-459c-9898-542e9f710747\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-f8kzc" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.068634 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/52120ed5-ff4f-4df5-8c60-3342df484be8-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-5fgn9\" (UID: \"52120ed5-ff4f-4df5-8c60-3342df484be8\") " pod="openshift-authentication/oauth-openshift-558db77b4-5fgn9" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.068652 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/489e9f98-0728-4a72-80d4-3d88a77823f5-profile-collector-cert\") pod \"olm-operator-6b444d44fb-hf8fp\" (UID: \"489e9f98-0728-4a72-80d4-3d88a77823f5\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-hf8fp" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.068687 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-njwpj\" (UniqueName: \"kubernetes.io/projected/a905919c-6c7e-40be-8f43-b938a60f14f6-kube-api-access-njwpj\") pod \"router-default-5444994796-sp4wh\" (UID: \"a905919c-6c7e-40be-8f43-b938a60f14f6\") " pod="openshift-ingress/router-default-5444994796-sp4wh" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.068705 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/95e0dc94-8aec-42d0-8b19-7cdf12fa7dee-config-volume\") pod \"collect-profiles-29409630-45pn5\" (UID: \"95e0dc94-8aec-42d0-8b19-7cdf12fa7dee\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29409630-45pn5" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.068735 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/a905919c-6c7e-40be-8f43-b938a60f14f6-stats-auth\") pod \"router-default-5444994796-sp4wh\" (UID: \"a905919c-6c7e-40be-8f43-b938a60f14f6\") " pod="openshift-ingress/router-default-5444994796-sp4wh" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.068757 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/52120ed5-ff4f-4df5-8c60-3342df484be8-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-5fgn9\" (UID: \"52120ed5-ff4f-4df5-8c60-3342df484be8\") " pod="openshift-authentication/oauth-openshift-558db77b4-5fgn9" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.068776 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/57c3bc40-b325-4343-a01b-6705e9e8eb77-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-8274k\" (UID: \"57c3bc40-b325-4343-a01b-6705e9e8eb77\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-8274k" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.068804 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/52120ed5-ff4f-4df5-8c60-3342df484be8-audit-policies\") pod \"oauth-openshift-558db77b4-5fgn9\" (UID: \"52120ed5-ff4f-4df5-8c60-3342df484be8\") " pod="openshift-authentication/oauth-openshift-558db77b4-5fgn9" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.068833 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-72d4v\" (UniqueName: \"kubernetes.io/projected/0f0340e8-faa5-447b-a5cd-a878e982d98f-kube-api-access-72d4v\") pod \"route-controller-manager-6576b87f9c-2mb7d\" (UID: \"0f0340e8-faa5-447b-a5cd-a878e982d98f\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-2mb7d" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.068853 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/c342090f-f331-484d-b287-929d8303f5d7-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-qsskc\" (UID: \"c342090f-f331-484d-b287-929d8303f5d7\") " pod="openshift-marketplace/marketplace-operator-79b997595-qsskc" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.068886 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/a2b231fa-0ecc-44da-9259-a7dc7261127c-ca-trust-extracted\") pod \"image-registry-697d97f7c8-mj64s\" (UID: \"a2b231fa-0ecc-44da-9259-a7dc7261127c\") " pod="openshift-image-registry/image-registry-697d97f7c8-mj64s" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.068902 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9p9xx\" (UniqueName: \"kubernetes.io/projected/a2b231fa-0ecc-44da-9259-a7dc7261127c-kube-api-access-9p9xx\") pod \"image-registry-697d97f7c8-mj64s\" (UID: \"a2b231fa-0ecc-44da-9259-a7dc7261127c\") " pod="openshift-image-registry/image-registry-697d97f7c8-mj64s" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.068933 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a905919c-6c7e-40be-8f43-b938a60f14f6-service-ca-bundle\") pod \"router-default-5444994796-sp4wh\" (UID: \"a905919c-6c7e-40be-8f43-b938a60f14f6\") " pod="openshift-ingress/router-default-5444994796-sp4wh" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.068973 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/52120ed5-ff4f-4df5-8c60-3342df484be8-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-5fgn9\" (UID: \"52120ed5-ff4f-4df5-8c60-3342df484be8\") " pod="openshift-authentication/oauth-openshift-558db77b4-5fgn9" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.068996 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xpwm2\" (UniqueName: \"kubernetes.io/projected/5aa447ba-69b3-4291-9db5-25f4aeee2370-kube-api-access-xpwm2\") pod \"catalog-operator-68c6474976-n4dp7\" (UID: \"5aa447ba-69b3-4291-9db5-25f4aeee2370\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-n4dp7" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.069065 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/363f808f-a265-464a-a3a7-4128e3734b14-mountpoint-dir\") pod \"csi-hostpathplugin-m8jlq\" (UID: \"363f808f-a265-464a-a3a7-4128e3734b14\") " pod="hostpath-provisioner/csi-hostpathplugin-m8jlq" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.069083 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c342090f-f331-484d-b287-929d8303f5d7-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-qsskc\" (UID: \"c342090f-f331-484d-b287-929d8303f5d7\") " pod="openshift-marketplace/marketplace-operator-79b997595-qsskc" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.069102 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h7wwf\" (UniqueName: \"kubernetes.io/projected/2fd313fd-c0b1-4aae-8b3b-2d078a7fc7d0-kube-api-access-h7wwf\") pod \"control-plane-machine-set-operator-78cbb6b69f-fxgbc\" (UID: \"2fd313fd-c0b1-4aae-8b3b-2d078a7fc7d0\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-fxgbc" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.069131 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-49f6c\" (UniqueName: \"kubernetes.io/projected/57a1918e-f28a-4890-b88c-bc80df5cf704-kube-api-access-49f6c\") pod \"service-ca-operator-777779d784-dh427\" (UID: \"57a1918e-f28a-4890-b88c-bc80df5cf704\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-dh427" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.069171 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5fvsv\" (UniqueName: \"kubernetes.io/projected/2e3f10d2-ac94-459c-9898-542e9f710747-kube-api-access-5fvsv\") pod \"packageserver-d55dfcdfc-f8kzc\" (UID: \"2e3f10d2-ac94-459c-9898-542e9f710747\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-f8kzc" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.069199 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a2b231fa-0ecc-44da-9259-a7dc7261127c-bound-sa-token\") pod \"image-registry-697d97f7c8-mj64s\" (UID: \"a2b231fa-0ecc-44da-9259-a7dc7261127c\") " pod="openshift-image-registry/image-registry-697d97f7c8-mj64s" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.069217 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t8jvf\" (UniqueName: \"kubernetes.io/projected/dadb7bca-672d-434b-897d-7eb83059f189-kube-api-access-t8jvf\") pod \"migrator-59844c95c7-sjjm7\" (UID: \"dadb7bca-672d-434b-897d-7eb83059f189\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-sjjm7" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.069235 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/95e0dc94-8aec-42d0-8b19-7cdf12fa7dee-secret-volume\") pod \"collect-profiles-29409630-45pn5\" (UID: \"95e0dc94-8aec-42d0-8b19-7cdf12fa7dee\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29409630-45pn5" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.069253 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nr7m5\" (UniqueName: \"kubernetes.io/projected/7d654f9a-b190-47bc-89ce-0caae2ec1513-kube-api-access-nr7m5\") pod \"multus-admission-controller-857f4d67dd-jb992\" (UID: \"7d654f9a-b190-47bc-89ce-0caae2ec1513\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-jb992" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.069271 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/61054467-59d0-40a9-a60c-09c6c6106acd-config-volume\") pod \"dns-default-j7v65\" (UID: \"61054467-59d0-40a9-a60c-09c6c6106acd\") " pod="openshift-dns/dns-default-j7v65" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.069304 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/52120ed5-ff4f-4df5-8c60-3342df484be8-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-5fgn9\" (UID: \"52120ed5-ff4f-4df5-8c60-3342df484be8\") " pod="openshift-authentication/oauth-openshift-558db77b4-5fgn9" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.069331 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/363f808f-a265-464a-a3a7-4128e3734b14-socket-dir\") pod \"csi-hostpathplugin-m8jlq\" (UID: \"363f808f-a265-464a-a3a7-4128e3734b14\") " pod="hostpath-provisioner/csi-hostpathplugin-m8jlq" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.069374 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a7588f88-9a71-4e13-9fc3-a34cbfa6a090-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-5z87f\" (UID: \"a7588f88-9a71-4e13-9fc3-a34cbfa6a090\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-5z87f" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.069393 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/1e597d43-ddb6-4cdb-8ce7-94b7427fb2f0-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-vcwrd\" (UID: \"1e597d43-ddb6-4cdb-8ce7-94b7427fb2f0\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-vcwrd" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.069431 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ppsvf\" (UniqueName: \"kubernetes.io/projected/363f808f-a265-464a-a3a7-4128e3734b14-kube-api-access-ppsvf\") pod \"csi-hostpathplugin-m8jlq\" (UID: \"363f808f-a265-464a-a3a7-4128e3734b14\") " pod="hostpath-provisioner/csi-hostpathplugin-m8jlq" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.069450 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/52120ed5-ff4f-4df5-8c60-3342df484be8-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-5fgn9\" (UID: \"52120ed5-ff4f-4df5-8c60-3342df484be8\") " pod="openshift-authentication/oauth-openshift-558db77b4-5fgn9" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.069468 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/52120ed5-ff4f-4df5-8c60-3342df484be8-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-5fgn9\" (UID: \"52120ed5-ff4f-4df5-8c60-3342df484be8\") " pod="openshift-authentication/oauth-openshift-558db77b4-5fgn9" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.069486 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/5aa447ba-69b3-4291-9db5-25f4aeee2370-srv-cert\") pod \"catalog-operator-68c6474976-n4dp7\" (UID: \"5aa447ba-69b3-4291-9db5-25f4aeee2370\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-n4dp7" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.069519 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mj64s\" (UID: \"a2b231fa-0ecc-44da-9259-a7dc7261127c\") " pod="openshift-image-registry/image-registry-697d97f7c8-mj64s" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.069542 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/a2b231fa-0ecc-44da-9259-a7dc7261127c-registry-certificates\") pod \"image-registry-697d97f7c8-mj64s\" (UID: \"a2b231fa-0ecc-44da-9259-a7dc7261127c\") " pod="openshift-image-registry/image-registry-697d97f7c8-mj64s" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.069563 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/aa7f6b23-f07d-44a9-8963-c2edbbb83918-node-bootstrap-token\") pod \"machine-config-server-ddnwl\" (UID: \"aa7f6b23-f07d-44a9-8963-c2edbbb83918\") " pod="openshift-machine-config-operator/machine-config-server-ddnwl" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.069582 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8b7d482d-6638-41f3-93b3-0b6aebf7401b-serving-cert\") pod \"console-operator-58897d9998-7w2kz\" (UID: \"8b7d482d-6638-41f3-93b3-0b6aebf7401b\") " pod="openshift-console-operator/console-operator-58897d9998-7w2kz" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.069600 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/52120ed5-ff4f-4df5-8c60-3342df484be8-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-5fgn9\" (UID: \"52120ed5-ff4f-4df5-8c60-3342df484be8\") " pod="openshift-authentication/oauth-openshift-558db77b4-5fgn9" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.069618 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/52120ed5-ff4f-4df5-8c60-3342df484be8-audit-dir\") pod \"oauth-openshift-558db77b4-5fgn9\" (UID: \"52120ed5-ff4f-4df5-8c60-3342df484be8\") " pod="openshift-authentication/oauth-openshift-558db77b4-5fgn9" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.069636 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/52120ed5-ff4f-4df5-8c60-3342df484be8-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-5fgn9\" (UID: \"52120ed5-ff4f-4df5-8c60-3342df484be8\") " pod="openshift-authentication/oauth-openshift-558db77b4-5fgn9" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.069674 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/57c3bc40-b325-4343-a01b-6705e9e8eb77-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-8274k\" (UID: \"57c3bc40-b325-4343-a01b-6705e9e8eb77\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-8274k" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.069708 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/f3561d3e-0e57-4243-83df-9a475da69fb6-proxy-tls\") pod \"machine-config-operator-74547568cd-dwkm7\" (UID: \"f3561d3e-0e57-4243-83df-9a475da69fb6\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-dwkm7" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.069725 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/7caa1280-3fcc-47b0-ab38-8aa5479fcfce-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-l5xp9\" (UID: \"7caa1280-3fcc-47b0-ab38-8aa5479fcfce\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-l5xp9" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.069745 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rdgwf\" (UniqueName: \"kubernetes.io/projected/525b7be7-bc47-4ff1-ada6-0521aa1c8620-kube-api-access-rdgwf\") pod \"service-ca-9c57cc56f-6sclt\" (UID: \"525b7be7-bc47-4ff1-ada6-0521aa1c8620\") " pod="openshift-service-ca/service-ca-9c57cc56f-6sclt" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.069811 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/a2b231fa-0ecc-44da-9259-a7dc7261127c-registry-tls\") pod \"image-registry-697d97f7c8-mj64s\" (UID: \"a2b231fa-0ecc-44da-9259-a7dc7261127c\") " pod="openshift-image-registry/image-registry-697d97f7c8-mj64s" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.069835 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/469840fe-2ac3-4431-9b2b-3b41451e6cbc-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-lfm7c\" (UID: \"469840fe-2ac3-4431-9b2b-3b41451e6cbc\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-lfm7c" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.069854 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/1e597d43-ddb6-4cdb-8ce7-94b7427fb2f0-proxy-tls\") pod \"machine-config-controller-84d6567774-vcwrd\" (UID: \"1e597d43-ddb6-4cdb-8ce7-94b7427fb2f0\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-vcwrd" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.069891 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/2e3f10d2-ac94-459c-9898-542e9f710747-tmpfs\") pod \"packageserver-d55dfcdfc-f8kzc\" (UID: \"2e3f10d2-ac94-459c-9898-542e9f710747\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-f8kzc" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.069915 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/a905919c-6c7e-40be-8f43-b938a60f14f6-default-certificate\") pod \"router-default-5444994796-sp4wh\" (UID: \"a905919c-6c7e-40be-8f43-b938a60f14f6\") " pod="openshift-ingress/router-default-5444994796-sp4wh" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.069952 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rdvtz\" (UniqueName: \"kubernetes.io/projected/7caa1280-3fcc-47b0-ab38-8aa5479fcfce-kube-api-access-rdvtz\") pod \"package-server-manager-789f6589d5-l5xp9\" (UID: \"7caa1280-3fcc-47b0-ab38-8aa5479fcfce\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-l5xp9" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.069990 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/adf5f998-8117-4362-868e-b7cdbbf91e0e-cert\") pod \"ingress-canary-49bkd\" (UID: \"adf5f998-8117-4362-868e-b7cdbbf91e0e\") " pod="openshift-ingress-canary/ingress-canary-49bkd" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.070033 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/363f808f-a265-464a-a3a7-4128e3734b14-csi-data-dir\") pod \"csi-hostpathplugin-m8jlq\" (UID: \"363f808f-a265-464a-a3a7-4128e3734b14\") " pod="hostpath-provisioner/csi-hostpathplugin-m8jlq" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.070077 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/f3561d3e-0e57-4243-83df-9a475da69fb6-auth-proxy-config\") pod \"machine-config-operator-74547568cd-dwkm7\" (UID: \"f3561d3e-0e57-4243-83df-9a475da69fb6\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-dwkm7" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.070099 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5krct\" (UniqueName: \"kubernetes.io/projected/f3561d3e-0e57-4243-83df-9a475da69fb6-kube-api-access-5krct\") pod \"machine-config-operator-74547568cd-dwkm7\" (UID: \"f3561d3e-0e57-4243-83df-9a475da69fb6\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-dwkm7" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.070124 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nvbp8\" (UniqueName: \"kubernetes.io/projected/1e597d43-ddb6-4cdb-8ce7-94b7427fb2f0-kube-api-access-nvbp8\") pod \"machine-config-controller-84d6567774-vcwrd\" (UID: \"1e597d43-ddb6-4cdb-8ce7-94b7427fb2f0\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-vcwrd" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.070166 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gjfw8\" (UniqueName: \"kubernetes.io/projected/95e0dc94-8aec-42d0-8b19-7cdf12fa7dee-kube-api-access-gjfw8\") pod \"collect-profiles-29409630-45pn5\" (UID: \"95e0dc94-8aec-42d0-8b19-7cdf12fa7dee\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29409630-45pn5" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.070200 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p66qw\" (UniqueName: \"kubernetes.io/projected/61054467-59d0-40a9-a60c-09c6c6106acd-kube-api-access-p66qw\") pod \"dns-default-j7v65\" (UID: \"61054467-59d0-40a9-a60c-09c6c6106acd\") " pod="openshift-dns/dns-default-j7v65" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.070240 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/a2b231fa-0ecc-44da-9259-a7dc7261127c-installation-pull-secrets\") pod \"image-registry-697d97f7c8-mj64s\" (UID: \"a2b231fa-0ecc-44da-9259-a7dc7261127c\") " pod="openshift-image-registry/image-registry-697d97f7c8-mj64s" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.070280 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/469840fe-2ac3-4431-9b2b-3b41451e6cbc-config\") pod \"kube-apiserver-operator-766d6c64bb-lfm7c\" (UID: \"469840fe-2ac3-4431-9b2b-3b41451e6cbc\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-lfm7c" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.070299 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/2fd313fd-c0b1-4aae-8b3b-2d078a7fc7d0-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-fxgbc\" (UID: \"2fd313fd-c0b1-4aae-8b3b-2d078a7fc7d0\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-fxgbc" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.070335 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8b7d482d-6638-41f3-93b3-0b6aebf7401b-trusted-ca\") pod \"console-operator-58897d9998-7w2kz\" (UID: \"8b7d482d-6638-41f3-93b3-0b6aebf7401b\") " pod="openshift-console-operator/console-operator-58897d9998-7w2kz" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.070356 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/52120ed5-ff4f-4df5-8c60-3342df484be8-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-5fgn9\" (UID: \"52120ed5-ff4f-4df5-8c60-3342df484be8\") " pod="openshift-authentication/oauth-openshift-558db77b4-5fgn9" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.070429 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/0f0340e8-faa5-447b-a5cd-a878e982d98f-client-ca\") pod \"route-controller-manager-6576b87f9c-2mb7d\" (UID: \"0f0340e8-faa5-447b-a5cd-a878e982d98f\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-2mb7d" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.070449 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/57c3bc40-b325-4343-a01b-6705e9e8eb77-config\") pod \"kube-controller-manager-operator-78b949d7b-8274k\" (UID: \"57c3bc40-b325-4343-a01b-6705e9e8eb77\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-8274k" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.070464 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/57a1918e-f28a-4890-b88c-bc80df5cf704-config\") pod \"service-ca-operator-777779d784-dh427\" (UID: \"57a1918e-f28a-4890-b88c-bc80df5cf704\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-dh427" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.070508 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/525b7be7-bc47-4ff1-ada6-0521aa1c8620-signing-cabundle\") pod \"service-ca-9c57cc56f-6sclt\" (UID: \"525b7be7-bc47-4ff1-ada6-0521aa1c8620\") " pod="openshift-service-ca/service-ca-9c57cc56f-6sclt" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.070528 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/f3561d3e-0e57-4243-83df-9a475da69fb6-images\") pod \"machine-config-operator-74547568cd-dwkm7\" (UID: \"f3561d3e-0e57-4243-83df-9a475da69fb6\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-dwkm7" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.070553 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/363f808f-a265-464a-a3a7-4128e3734b14-registration-dir\") pod \"csi-hostpathplugin-m8jlq\" (UID: \"363f808f-a265-464a-a3a7-4128e3734b14\") " pod="hostpath-provisioner/csi-hostpathplugin-m8jlq" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.070590 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/489e9f98-0728-4a72-80d4-3d88a77823f5-srv-cert\") pod \"olm-operator-6b444d44fb-hf8fp\" (UID: \"489e9f98-0728-4a72-80d4-3d88a77823f5\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-hf8fp" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.070608 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/61054467-59d0-40a9-a60c-09c6c6106acd-metrics-tls\") pod \"dns-default-j7v65\" (UID: \"61054467-59d0-40a9-a60c-09c6c6106acd\") " pod="openshift-dns/dns-default-j7v65" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.070630 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a7588f88-9a71-4e13-9fc3-a34cbfa6a090-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-5z87f\" (UID: \"a7588f88-9a71-4e13-9fc3-a34cbfa6a090\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-5z87f" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.070650 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jm2f5\" (UniqueName: \"kubernetes.io/projected/c342090f-f331-484d-b287-929d8303f5d7-kube-api-access-jm2f5\") pod \"marketplace-operator-79b997595-qsskc\" (UID: \"c342090f-f331-484d-b287-929d8303f5d7\") " pod="openshift-marketplace/marketplace-operator-79b997595-qsskc" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.070705 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/525b7be7-bc47-4ff1-ada6-0521aa1c8620-signing-key\") pod \"service-ca-9c57cc56f-6sclt\" (UID: \"525b7be7-bc47-4ff1-ada6-0521aa1c8620\") " pod="openshift-service-ca/service-ca-9c57cc56f-6sclt" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.070753 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/5aa447ba-69b3-4291-9db5-25f4aeee2370-profile-collector-cert\") pod \"catalog-operator-68c6474976-n4dp7\" (UID: \"5aa447ba-69b3-4291-9db5-25f4aeee2370\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-n4dp7" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.070771 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a7588f88-9a71-4e13-9fc3-a34cbfa6a090-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-5z87f\" (UID: \"a7588f88-9a71-4e13-9fc3-a34cbfa6a090\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-5z87f" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.070810 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lmpxp\" (UniqueName: \"kubernetes.io/projected/52120ed5-ff4f-4df5-8c60-3342df484be8-kube-api-access-lmpxp\") pod \"oauth-openshift-558db77b4-5fgn9\" (UID: \"52120ed5-ff4f-4df5-8c60-3342df484be8\") " pod="openshift-authentication/oauth-openshift-558db77b4-5fgn9" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.070831 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/aa7f6b23-f07d-44a9-8963-c2edbbb83918-certs\") pod \"machine-config-server-ddnwl\" (UID: \"aa7f6b23-f07d-44a9-8963-c2edbbb83918\") " pod="openshift-machine-config-operator/machine-config-server-ddnwl" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.070848 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/2e3f10d2-ac94-459c-9898-542e9f710747-apiservice-cert\") pod \"packageserver-d55dfcdfc-f8kzc\" (UID: \"2e3f10d2-ac94-459c-9898-542e9f710747\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-f8kzc" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.070872 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s9rtm\" (UniqueName: \"kubernetes.io/projected/aa7f6b23-f07d-44a9-8963-c2edbbb83918-kube-api-access-s9rtm\") pod \"machine-config-server-ddnwl\" (UID: \"aa7f6b23-f07d-44a9-8963-c2edbbb83918\") " pod="openshift-machine-config-operator/machine-config-server-ddnwl" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.070890 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8b7d482d-6638-41f3-93b3-0b6aebf7401b-config\") pod \"console-operator-58897d9998-7w2kz\" (UID: \"8b7d482d-6638-41f3-93b3-0b6aebf7401b\") " pod="openshift-console-operator/console-operator-58897d9998-7w2kz" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.070912 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0f0340e8-faa5-447b-a5cd-a878e982d98f-serving-cert\") pod \"route-controller-manager-6576b87f9c-2mb7d\" (UID: \"0f0340e8-faa5-447b-a5cd-a878e982d98f\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-2mb7d" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.070933 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/363f808f-a265-464a-a3a7-4128e3734b14-plugins-dir\") pod \"csi-hostpathplugin-m8jlq\" (UID: \"363f808f-a265-464a-a3a7-4128e3734b14\") " pod="hostpath-provisioner/csi-hostpathplugin-m8jlq" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.070953 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a2b231fa-0ecc-44da-9259-a7dc7261127c-trusted-ca\") pod \"image-registry-697d97f7c8-mj64s\" (UID: \"a2b231fa-0ecc-44da-9259-a7dc7261127c\") " pod="openshift-image-registry/image-registry-697d97f7c8-mj64s" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.080955 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6t6bj\" (UniqueName: \"kubernetes.io/projected/adf5f998-8117-4362-868e-b7cdbbf91e0e-kube-api-access-6t6bj\") pod \"ingress-canary-49bkd\" (UID: \"adf5f998-8117-4362-868e-b7cdbbf91e0e\") " pod="openshift-ingress-canary/ingress-canary-49bkd" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.081062 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b8p9d\" (UniqueName: \"kubernetes.io/projected/8b7d482d-6638-41f3-93b3-0b6aebf7401b-kube-api-access-b8p9d\") pod \"console-operator-58897d9998-7w2kz\" (UID: \"8b7d482d-6638-41f3-93b3-0b6aebf7401b\") " pod="openshift-console-operator/console-operator-58897d9998-7w2kz" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.081090 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/a905919c-6c7e-40be-8f43-b938a60f14f6-metrics-certs\") pod \"router-default-5444994796-sp4wh\" (UID: \"a905919c-6c7e-40be-8f43-b938a60f14f6\") " pod="openshift-ingress/router-default-5444994796-sp4wh" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.077147 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a2b231fa-0ecc-44da-9259-a7dc7261127c-trusted-ca\") pod \"image-registry-697d97f7c8-mj64s\" (UID: \"a2b231fa-0ecc-44da-9259-a7dc7261127c\") " pod="openshift-image-registry/image-registry-697d97f7c8-mj64s" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.080723 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-wtwtl"] Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.085305 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/52120ed5-ff4f-4df5-8c60-3342df484be8-audit-policies\") pod \"oauth-openshift-558db77b4-5fgn9\" (UID: \"52120ed5-ff4f-4df5-8c60-3342df484be8\") " pod="openshift-authentication/oauth-openshift-558db77b4-5fgn9" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.086894 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/a2b231fa-0ecc-44da-9259-a7dc7261127c-ca-trust-extracted\") pod \"image-registry-697d97f7c8-mj64s\" (UID: \"a2b231fa-0ecc-44da-9259-a7dc7261127c\") " pod="openshift-image-registry/image-registry-697d97f7c8-mj64s" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.089547 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/52120ed5-ff4f-4df5-8c60-3342df484be8-audit-dir\") pod \"oauth-openshift-558db77b4-5fgn9\" (UID: \"52120ed5-ff4f-4df5-8c60-3342df484be8\") " pod="openshift-authentication/oauth-openshift-558db77b4-5fgn9" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.089887 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/52120ed5-ff4f-4df5-8c60-3342df484be8-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-5fgn9\" (UID: \"52120ed5-ff4f-4df5-8c60-3342df484be8\") " pod="openshift-authentication/oauth-openshift-558db77b4-5fgn9" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.090111 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/52120ed5-ff4f-4df5-8c60-3342df484be8-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-5fgn9\" (UID: \"52120ed5-ff4f-4df5-8c60-3342df484be8\") " pod="openshift-authentication/oauth-openshift-558db77b4-5fgn9" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.090432 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/a2b231fa-0ecc-44da-9259-a7dc7261127c-registry-tls\") pod \"image-registry-697d97f7c8-mj64s\" (UID: \"a2b231fa-0ecc-44da-9259-a7dc7261127c\") " pod="openshift-image-registry/image-registry-697d97f7c8-mj64s" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.090459 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/52120ed5-ff4f-4df5-8c60-3342df484be8-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-5fgn9\" (UID: \"52120ed5-ff4f-4df5-8c60-3342df484be8\") " pod="openshift-authentication/oauth-openshift-558db77b4-5fgn9" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.091178 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/57c3bc40-b325-4343-a01b-6705e9e8eb77-config\") pod \"kube-controller-manager-operator-78b949d7b-8274k\" (UID: \"57c3bc40-b325-4343-a01b-6705e9e8eb77\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-8274k" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.091745 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/0f0340e8-faa5-447b-a5cd-a878e982d98f-client-ca\") pod \"route-controller-manager-6576b87f9c-2mb7d\" (UID: \"0f0340e8-faa5-447b-a5cd-a878e982d98f\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-2mb7d" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.091845 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8b7d482d-6638-41f3-93b3-0b6aebf7401b-trusted-ca\") pod \"console-operator-58897d9998-7w2kz\" (UID: \"8b7d482d-6638-41f3-93b3-0b6aebf7401b\") " pod="openshift-console-operator/console-operator-58897d9998-7w2kz" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.092480 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/52120ed5-ff4f-4df5-8c60-3342df484be8-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-5fgn9\" (UID: \"52120ed5-ff4f-4df5-8c60-3342df484be8\") " pod="openshift-authentication/oauth-openshift-558db77b4-5fgn9" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.092806 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8b7d482d-6638-41f3-93b3-0b6aebf7401b-config\") pod \"console-operator-58897d9998-7w2kz\" (UID: \"8b7d482d-6638-41f3-93b3-0b6aebf7401b\") " pod="openshift-console-operator/console-operator-58897d9998-7w2kz" Dec 01 08:41:55 crc kubenswrapper[4813]: E1201 08:41:55.092944 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 08:41:55.592904179 +0000 UTC m=+100.215825785 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mj64s" (UID: "a2b231fa-0ecc-44da-9259-a7dc7261127c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.093580 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0f0340e8-faa5-447b-a5cd-a878e982d98f-serving-cert\") pod \"route-controller-manager-6576b87f9c-2mb7d\" (UID: \"0f0340e8-faa5-447b-a5cd-a878e982d98f\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-2mb7d" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.094343 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/a2b231fa-0ecc-44da-9259-a7dc7261127c-installation-pull-secrets\") pod \"image-registry-697d97f7c8-mj64s\" (UID: \"a2b231fa-0ecc-44da-9259-a7dc7261127c\") " pod="openshift-image-registry/image-registry-697d97f7c8-mj64s" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.095180 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/52120ed5-ff4f-4df5-8c60-3342df484be8-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-5fgn9\" (UID: \"52120ed5-ff4f-4df5-8c60-3342df484be8\") " pod="openshift-authentication/oauth-openshift-558db77b4-5fgn9" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.096979 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0f0340e8-faa5-447b-a5cd-a878e982d98f-config\") pod \"route-controller-manager-6576b87f9c-2mb7d\" (UID: \"0f0340e8-faa5-447b-a5cd-a878e982d98f\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-2mb7d" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.097563 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/52120ed5-ff4f-4df5-8c60-3342df484be8-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-5fgn9\" (UID: \"52120ed5-ff4f-4df5-8c60-3342df484be8\") " pod="openshift-authentication/oauth-openshift-558db77b4-5fgn9" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.098234 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/a2b231fa-0ecc-44da-9259-a7dc7261127c-registry-certificates\") pod \"image-registry-697d97f7c8-mj64s\" (UID: \"a2b231fa-0ecc-44da-9259-a7dc7261127c\") " pod="openshift-image-registry/image-registry-697d97f7c8-mj64s" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.100278 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/52120ed5-ff4f-4df5-8c60-3342df484be8-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-5fgn9\" (UID: \"52120ed5-ff4f-4df5-8c60-3342df484be8\") " pod="openshift-authentication/oauth-openshift-558db77b4-5fgn9" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.102344 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/52120ed5-ff4f-4df5-8c60-3342df484be8-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-5fgn9\" (UID: \"52120ed5-ff4f-4df5-8c60-3342df484be8\") " pod="openshift-authentication/oauth-openshift-558db77b4-5fgn9" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.103530 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/52120ed5-ff4f-4df5-8c60-3342df484be8-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-5fgn9\" (UID: \"52120ed5-ff4f-4df5-8c60-3342df484be8\") " pod="openshift-authentication/oauth-openshift-558db77b4-5fgn9" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.111284 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/52120ed5-ff4f-4df5-8c60-3342df484be8-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-5fgn9\" (UID: \"52120ed5-ff4f-4df5-8c60-3342df484be8\") " pod="openshift-authentication/oauth-openshift-558db77b4-5fgn9" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.113012 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8b7d482d-6638-41f3-93b3-0b6aebf7401b-serving-cert\") pod \"console-operator-58897d9998-7w2kz\" (UID: \"8b7d482d-6638-41f3-93b3-0b6aebf7401b\") " pod="openshift-console-operator/console-operator-58897d9998-7w2kz" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.116390 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/57c3bc40-b325-4343-a01b-6705e9e8eb77-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-8274k\" (UID: \"57c3bc40-b325-4343-a01b-6705e9e8eb77\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-8274k" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.121166 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/52120ed5-ff4f-4df5-8c60-3342df484be8-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-5fgn9\" (UID: \"52120ed5-ff4f-4df5-8c60-3342df484be8\") " pod="openshift-authentication/oauth-openshift-558db77b4-5fgn9" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.137388 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b8p9d\" (UniqueName: \"kubernetes.io/projected/8b7d482d-6638-41f3-93b3-0b6aebf7401b-kube-api-access-b8p9d\") pod \"console-operator-58897d9998-7w2kz\" (UID: \"8b7d482d-6638-41f3-93b3-0b6aebf7401b\") " pod="openshift-console-operator/console-operator-58897d9998-7w2kz" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.149531 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/57c3bc40-b325-4343-a01b-6705e9e8eb77-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-8274k\" (UID: \"57c3bc40-b325-4343-a01b-6705e9e8eb77\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-8274k" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.160674 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-hmh5v"] Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.208307 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.208596 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a7588f88-9a71-4e13-9fc3-a34cbfa6a090-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-5z87f\" (UID: \"a7588f88-9a71-4e13-9fc3-a34cbfa6a090\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-5z87f" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.208622 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/1e597d43-ddb6-4cdb-8ce7-94b7427fb2f0-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-vcwrd\" (UID: \"1e597d43-ddb6-4cdb-8ce7-94b7427fb2f0\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-vcwrd" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.208643 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ppsvf\" (UniqueName: \"kubernetes.io/projected/363f808f-a265-464a-a3a7-4128e3734b14-kube-api-access-ppsvf\") pod \"csi-hostpathplugin-m8jlq\" (UID: \"363f808f-a265-464a-a3a7-4128e3734b14\") " pod="hostpath-provisioner/csi-hostpathplugin-m8jlq" Dec 01 08:41:55 crc kubenswrapper[4813]: E1201 08:41:55.209098 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 08:41:55.709064045 +0000 UTC m=+100.331985661 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.209388 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/5aa447ba-69b3-4291-9db5-25f4aeee2370-srv-cert\") pod \"catalog-operator-68c6474976-n4dp7\" (UID: \"5aa447ba-69b3-4291-9db5-25f4aeee2370\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-n4dp7" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.209436 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mj64s\" (UID: \"a2b231fa-0ecc-44da-9259-a7dc7261127c\") " pod="openshift-image-registry/image-registry-697d97f7c8-mj64s" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.209454 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/aa7f6b23-f07d-44a9-8963-c2edbbb83918-node-bootstrap-token\") pod \"machine-config-server-ddnwl\" (UID: \"aa7f6b23-f07d-44a9-8963-c2edbbb83918\") " pod="openshift-machine-config-operator/machine-config-server-ddnwl" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.209482 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/f3561d3e-0e57-4243-83df-9a475da69fb6-proxy-tls\") pod \"machine-config-operator-74547568cd-dwkm7\" (UID: \"f3561d3e-0e57-4243-83df-9a475da69fb6\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-dwkm7" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.209499 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/7caa1280-3fcc-47b0-ab38-8aa5479fcfce-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-l5xp9\" (UID: \"7caa1280-3fcc-47b0-ab38-8aa5479fcfce\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-l5xp9" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.209520 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdgwf\" (UniqueName: \"kubernetes.io/projected/525b7be7-bc47-4ff1-ada6-0521aa1c8620-kube-api-access-rdgwf\") pod \"service-ca-9c57cc56f-6sclt\" (UID: \"525b7be7-bc47-4ff1-ada6-0521aa1c8620\") " pod="openshift-service-ca/service-ca-9c57cc56f-6sclt" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.209548 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/469840fe-2ac3-4431-9b2b-3b41451e6cbc-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-lfm7c\" (UID: \"469840fe-2ac3-4431-9b2b-3b41451e6cbc\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-lfm7c" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.209564 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/1e597d43-ddb6-4cdb-8ce7-94b7427fb2f0-proxy-tls\") pod \"machine-config-controller-84d6567774-vcwrd\" (UID: \"1e597d43-ddb6-4cdb-8ce7-94b7427fb2f0\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-vcwrd" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.209582 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/2e3f10d2-ac94-459c-9898-542e9f710747-tmpfs\") pod \"packageserver-d55dfcdfc-f8kzc\" (UID: \"2e3f10d2-ac94-459c-9898-542e9f710747\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-f8kzc" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.209603 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/a905919c-6c7e-40be-8f43-b938a60f14f6-default-certificate\") pod \"router-default-5444994796-sp4wh\" (UID: \"a905919c-6c7e-40be-8f43-b938a60f14f6\") " pod="openshift-ingress/router-default-5444994796-sp4wh" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.209624 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdvtz\" (UniqueName: \"kubernetes.io/projected/7caa1280-3fcc-47b0-ab38-8aa5479fcfce-kube-api-access-rdvtz\") pod \"package-server-manager-789f6589d5-l5xp9\" (UID: \"7caa1280-3fcc-47b0-ab38-8aa5479fcfce\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-l5xp9" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.209640 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/adf5f998-8117-4362-868e-b7cdbbf91e0e-cert\") pod \"ingress-canary-49bkd\" (UID: \"adf5f998-8117-4362-868e-b7cdbbf91e0e\") " pod="openshift-ingress-canary/ingress-canary-49bkd" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.209656 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/363f808f-a265-464a-a3a7-4128e3734b14-csi-data-dir\") pod \"csi-hostpathplugin-m8jlq\" (UID: \"363f808f-a265-464a-a3a7-4128e3734b14\") " pod="hostpath-provisioner/csi-hostpathplugin-m8jlq" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.209679 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/f3561d3e-0e57-4243-83df-9a475da69fb6-auth-proxy-config\") pod \"machine-config-operator-74547568cd-dwkm7\" (UID: \"f3561d3e-0e57-4243-83df-9a475da69fb6\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-dwkm7" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.209694 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5krct\" (UniqueName: \"kubernetes.io/projected/f3561d3e-0e57-4243-83df-9a475da69fb6-kube-api-access-5krct\") pod \"machine-config-operator-74547568cd-dwkm7\" (UID: \"f3561d3e-0e57-4243-83df-9a475da69fb6\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-dwkm7" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.209709 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nvbp8\" (UniqueName: \"kubernetes.io/projected/1e597d43-ddb6-4cdb-8ce7-94b7427fb2f0-kube-api-access-nvbp8\") pod \"machine-config-controller-84d6567774-vcwrd\" (UID: \"1e597d43-ddb6-4cdb-8ce7-94b7427fb2f0\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-vcwrd" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.209726 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gjfw8\" (UniqueName: \"kubernetes.io/projected/95e0dc94-8aec-42d0-8b19-7cdf12fa7dee-kube-api-access-gjfw8\") pod \"collect-profiles-29409630-45pn5\" (UID: \"95e0dc94-8aec-42d0-8b19-7cdf12fa7dee\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29409630-45pn5" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.209759 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p66qw\" (UniqueName: \"kubernetes.io/projected/61054467-59d0-40a9-a60c-09c6c6106acd-kube-api-access-p66qw\") pod \"dns-default-j7v65\" (UID: \"61054467-59d0-40a9-a60c-09c6c6106acd\") " pod="openshift-dns/dns-default-j7v65" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.209788 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/469840fe-2ac3-4431-9b2b-3b41451e6cbc-config\") pod \"kube-apiserver-operator-766d6c64bb-lfm7c\" (UID: \"469840fe-2ac3-4431-9b2b-3b41451e6cbc\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-lfm7c" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.209804 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/2fd313fd-c0b1-4aae-8b3b-2d078a7fc7d0-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-fxgbc\" (UID: \"2fd313fd-c0b1-4aae-8b3b-2d078a7fc7d0\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-fxgbc" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.209832 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/57a1918e-f28a-4890-b88c-bc80df5cf704-config\") pod \"service-ca-operator-777779d784-dh427\" (UID: \"57a1918e-f28a-4890-b88c-bc80df5cf704\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-dh427" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.209853 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/525b7be7-bc47-4ff1-ada6-0521aa1c8620-signing-cabundle\") pod \"service-ca-9c57cc56f-6sclt\" (UID: \"525b7be7-bc47-4ff1-ada6-0521aa1c8620\") " pod="openshift-service-ca/service-ca-9c57cc56f-6sclt" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.209869 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/f3561d3e-0e57-4243-83df-9a475da69fb6-images\") pod \"machine-config-operator-74547568cd-dwkm7\" (UID: \"f3561d3e-0e57-4243-83df-9a475da69fb6\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-dwkm7" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.209885 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/363f808f-a265-464a-a3a7-4128e3734b14-registration-dir\") pod \"csi-hostpathplugin-m8jlq\" (UID: \"363f808f-a265-464a-a3a7-4128e3734b14\") " pod="hostpath-provisioner/csi-hostpathplugin-m8jlq" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.209905 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/489e9f98-0728-4a72-80d4-3d88a77823f5-srv-cert\") pod \"olm-operator-6b444d44fb-hf8fp\" (UID: \"489e9f98-0728-4a72-80d4-3d88a77823f5\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-hf8fp" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.209918 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/61054467-59d0-40a9-a60c-09c6c6106acd-metrics-tls\") pod \"dns-default-j7v65\" (UID: \"61054467-59d0-40a9-a60c-09c6c6106acd\") " pod="openshift-dns/dns-default-j7v65" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.209935 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a7588f88-9a71-4e13-9fc3-a34cbfa6a090-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-5z87f\" (UID: \"a7588f88-9a71-4e13-9fc3-a34cbfa6a090\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-5z87f" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.209949 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jm2f5\" (UniqueName: \"kubernetes.io/projected/c342090f-f331-484d-b287-929d8303f5d7-kube-api-access-jm2f5\") pod \"marketplace-operator-79b997595-qsskc\" (UID: \"c342090f-f331-484d-b287-929d8303f5d7\") " pod="openshift-marketplace/marketplace-operator-79b997595-qsskc" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.210011 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/525b7be7-bc47-4ff1-ada6-0521aa1c8620-signing-key\") pod \"service-ca-9c57cc56f-6sclt\" (UID: \"525b7be7-bc47-4ff1-ada6-0521aa1c8620\") " pod="openshift-service-ca/service-ca-9c57cc56f-6sclt" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.210031 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/5aa447ba-69b3-4291-9db5-25f4aeee2370-profile-collector-cert\") pod \"catalog-operator-68c6474976-n4dp7\" (UID: \"5aa447ba-69b3-4291-9db5-25f4aeee2370\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-n4dp7" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.210045 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a7588f88-9a71-4e13-9fc3-a34cbfa6a090-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-5z87f\" (UID: \"a7588f88-9a71-4e13-9fc3-a34cbfa6a090\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-5z87f" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.210066 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/aa7f6b23-f07d-44a9-8963-c2edbbb83918-certs\") pod \"machine-config-server-ddnwl\" (UID: \"aa7f6b23-f07d-44a9-8963-c2edbbb83918\") " pod="openshift-machine-config-operator/machine-config-server-ddnwl" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.210080 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/2e3f10d2-ac94-459c-9898-542e9f710747-apiservice-cert\") pod \"packageserver-d55dfcdfc-f8kzc\" (UID: \"2e3f10d2-ac94-459c-9898-542e9f710747\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-f8kzc" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.210094 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s9rtm\" (UniqueName: \"kubernetes.io/projected/aa7f6b23-f07d-44a9-8963-c2edbbb83918-kube-api-access-s9rtm\") pod \"machine-config-server-ddnwl\" (UID: \"aa7f6b23-f07d-44a9-8963-c2edbbb83918\") " pod="openshift-machine-config-operator/machine-config-server-ddnwl" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.210111 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/363f808f-a265-464a-a3a7-4128e3734b14-plugins-dir\") pod \"csi-hostpathplugin-m8jlq\" (UID: \"363f808f-a265-464a-a3a7-4128e3734b14\") " pod="hostpath-provisioner/csi-hostpathplugin-m8jlq" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.210125 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6t6bj\" (UniqueName: \"kubernetes.io/projected/adf5f998-8117-4362-868e-b7cdbbf91e0e-kube-api-access-6t6bj\") pod \"ingress-canary-49bkd\" (UID: \"adf5f998-8117-4362-868e-b7cdbbf91e0e\") " pod="openshift-ingress-canary/ingress-canary-49bkd" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.210145 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/a905919c-6c7e-40be-8f43-b938a60f14f6-metrics-certs\") pod \"router-default-5444994796-sp4wh\" (UID: \"a905919c-6c7e-40be-8f43-b938a60f14f6\") " pod="openshift-ingress/router-default-5444994796-sp4wh" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.210161 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-66c2t\" (UniqueName: \"kubernetes.io/projected/489e9f98-0728-4a72-80d4-3d88a77823f5-kube-api-access-66c2t\") pod \"olm-operator-6b444d44fb-hf8fp\" (UID: \"489e9f98-0728-4a72-80d4-3d88a77823f5\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-hf8fp" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.210180 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/7d654f9a-b190-47bc-89ce-0caae2ec1513-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-jb992\" (UID: \"7d654f9a-b190-47bc-89ce-0caae2ec1513\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-jb992" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.210197 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/57a1918e-f28a-4890-b88c-bc80df5cf704-serving-cert\") pod \"service-ca-operator-777779d784-dh427\" (UID: \"57a1918e-f28a-4890-b88c-bc80df5cf704\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-dh427" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.210211 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/469840fe-2ac3-4431-9b2b-3b41451e6cbc-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-lfm7c\" (UID: \"469840fe-2ac3-4431-9b2b-3b41451e6cbc\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-lfm7c" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.210227 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/2e3f10d2-ac94-459c-9898-542e9f710747-webhook-cert\") pod \"packageserver-d55dfcdfc-f8kzc\" (UID: \"2e3f10d2-ac94-459c-9898-542e9f710747\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-f8kzc" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.210245 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/489e9f98-0728-4a72-80d4-3d88a77823f5-profile-collector-cert\") pod \"olm-operator-6b444d44fb-hf8fp\" (UID: \"489e9f98-0728-4a72-80d4-3d88a77823f5\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-hf8fp" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.210267 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-njwpj\" (UniqueName: \"kubernetes.io/projected/a905919c-6c7e-40be-8f43-b938a60f14f6-kube-api-access-njwpj\") pod \"router-default-5444994796-sp4wh\" (UID: \"a905919c-6c7e-40be-8f43-b938a60f14f6\") " pod="openshift-ingress/router-default-5444994796-sp4wh" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.210282 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/95e0dc94-8aec-42d0-8b19-7cdf12fa7dee-config-volume\") pod \"collect-profiles-29409630-45pn5\" (UID: \"95e0dc94-8aec-42d0-8b19-7cdf12fa7dee\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29409630-45pn5" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.210300 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/a905919c-6c7e-40be-8f43-b938a60f14f6-stats-auth\") pod \"router-default-5444994796-sp4wh\" (UID: \"a905919c-6c7e-40be-8f43-b938a60f14f6\") " pod="openshift-ingress/router-default-5444994796-sp4wh" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.210323 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/c342090f-f331-484d-b287-929d8303f5d7-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-qsskc\" (UID: \"c342090f-f331-484d-b287-929d8303f5d7\") " pod="openshift-marketplace/marketplace-operator-79b997595-qsskc" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.210371 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a905919c-6c7e-40be-8f43-b938a60f14f6-service-ca-bundle\") pod \"router-default-5444994796-sp4wh\" (UID: \"a905919c-6c7e-40be-8f43-b938a60f14f6\") " pod="openshift-ingress/router-default-5444994796-sp4wh" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.210389 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xpwm2\" (UniqueName: \"kubernetes.io/projected/5aa447ba-69b3-4291-9db5-25f4aeee2370-kube-api-access-xpwm2\") pod \"catalog-operator-68c6474976-n4dp7\" (UID: \"5aa447ba-69b3-4291-9db5-25f4aeee2370\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-n4dp7" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.210409 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/363f808f-a265-464a-a3a7-4128e3734b14-mountpoint-dir\") pod \"csi-hostpathplugin-m8jlq\" (UID: \"363f808f-a265-464a-a3a7-4128e3734b14\") " pod="hostpath-provisioner/csi-hostpathplugin-m8jlq" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.210423 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c342090f-f331-484d-b287-929d8303f5d7-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-qsskc\" (UID: \"c342090f-f331-484d-b287-929d8303f5d7\") " pod="openshift-marketplace/marketplace-operator-79b997595-qsskc" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.210448 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h7wwf\" (UniqueName: \"kubernetes.io/projected/2fd313fd-c0b1-4aae-8b3b-2d078a7fc7d0-kube-api-access-h7wwf\") pod \"control-plane-machine-set-operator-78cbb6b69f-fxgbc\" (UID: \"2fd313fd-c0b1-4aae-8b3b-2d078a7fc7d0\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-fxgbc" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.210467 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-49f6c\" (UniqueName: \"kubernetes.io/projected/57a1918e-f28a-4890-b88c-bc80df5cf704-kube-api-access-49f6c\") pod \"service-ca-operator-777779d784-dh427\" (UID: \"57a1918e-f28a-4890-b88c-bc80df5cf704\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-dh427" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.210485 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5fvsv\" (UniqueName: \"kubernetes.io/projected/2e3f10d2-ac94-459c-9898-542e9f710747-kube-api-access-5fvsv\") pod \"packageserver-d55dfcdfc-f8kzc\" (UID: \"2e3f10d2-ac94-459c-9898-542e9f710747\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-f8kzc" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.210506 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t8jvf\" (UniqueName: \"kubernetes.io/projected/dadb7bca-672d-434b-897d-7eb83059f189-kube-api-access-t8jvf\") pod \"migrator-59844c95c7-sjjm7\" (UID: \"dadb7bca-672d-434b-897d-7eb83059f189\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-sjjm7" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.210523 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/95e0dc94-8aec-42d0-8b19-7cdf12fa7dee-secret-volume\") pod \"collect-profiles-29409630-45pn5\" (UID: \"95e0dc94-8aec-42d0-8b19-7cdf12fa7dee\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29409630-45pn5" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.210540 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nr7m5\" (UniqueName: \"kubernetes.io/projected/7d654f9a-b190-47bc-89ce-0caae2ec1513-kube-api-access-nr7m5\") pod \"multus-admission-controller-857f4d67dd-jb992\" (UID: \"7d654f9a-b190-47bc-89ce-0caae2ec1513\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-jb992" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.210555 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/61054467-59d0-40a9-a60c-09c6c6106acd-config-volume\") pod \"dns-default-j7v65\" (UID: \"61054467-59d0-40a9-a60c-09c6c6106acd\") " pod="openshift-dns/dns-default-j7v65" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.210571 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/363f808f-a265-464a-a3a7-4128e3734b14-socket-dir\") pod \"csi-hostpathplugin-m8jlq\" (UID: \"363f808f-a265-464a-a3a7-4128e3734b14\") " pod="hostpath-provisioner/csi-hostpathplugin-m8jlq" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.210749 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/363f808f-a265-464a-a3a7-4128e3734b14-socket-dir\") pod \"csi-hostpathplugin-m8jlq\" (UID: \"363f808f-a265-464a-a3a7-4128e3734b14\") " pod="hostpath-provisioner/csi-hostpathplugin-m8jlq" Dec 01 08:41:55 crc kubenswrapper[4813]: E1201 08:41:55.211342 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 08:41:55.711324973 +0000 UTC m=+100.334246559 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mj64s" (UID: "a2b231fa-0ecc-44da-9259-a7dc7261127c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.213270 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a7588f88-9a71-4e13-9fc3-a34cbfa6a090-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-5z87f\" (UID: \"a7588f88-9a71-4e13-9fc3-a34cbfa6a090\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-5z87f" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.214558 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/5aa447ba-69b3-4291-9db5-25f4aeee2370-srv-cert\") pod \"catalog-operator-68c6474976-n4dp7\" (UID: \"5aa447ba-69b3-4291-9db5-25f4aeee2370\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-n4dp7" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.217058 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/a905919c-6c7e-40be-8f43-b938a60f14f6-stats-auth\") pod \"router-default-5444994796-sp4wh\" (UID: \"a905919c-6c7e-40be-8f43-b938a60f14f6\") " pod="openshift-ingress/router-default-5444994796-sp4wh" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.217129 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/363f808f-a265-464a-a3a7-4128e3734b14-mountpoint-dir\") pod \"csi-hostpathplugin-m8jlq\" (UID: \"363f808f-a265-464a-a3a7-4128e3734b14\") " pod="hostpath-provisioner/csi-hostpathplugin-m8jlq" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.218361 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c342090f-f331-484d-b287-929d8303f5d7-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-qsskc\" (UID: \"c342090f-f331-484d-b287-929d8303f5d7\") " pod="openshift-marketplace/marketplace-operator-79b997595-qsskc" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.218943 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/1e597d43-ddb6-4cdb-8ce7-94b7427fb2f0-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-vcwrd\" (UID: \"1e597d43-ddb6-4cdb-8ce7-94b7427fb2f0\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-vcwrd" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.218952 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/363f808f-a265-464a-a3a7-4128e3734b14-plugins-dir\") pod \"csi-hostpathplugin-m8jlq\" (UID: \"363f808f-a265-464a-a3a7-4128e3734b14\") " pod="hostpath-provisioner/csi-hostpathplugin-m8jlq" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.219341 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a905919c-6c7e-40be-8f43-b938a60f14f6-service-ca-bundle\") pod \"router-default-5444994796-sp4wh\" (UID: \"a905919c-6c7e-40be-8f43-b938a60f14f6\") " pod="openshift-ingress/router-default-5444994796-sp4wh" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.219550 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/aa7f6b23-f07d-44a9-8963-c2edbbb83918-node-bootstrap-token\") pod \"machine-config-server-ddnwl\" (UID: \"aa7f6b23-f07d-44a9-8963-c2edbbb83918\") " pod="openshift-machine-config-operator/machine-config-server-ddnwl" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.220232 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/f3561d3e-0e57-4243-83df-9a475da69fb6-proxy-tls\") pod \"machine-config-operator-74547568cd-dwkm7\" (UID: \"f3561d3e-0e57-4243-83df-9a475da69fb6\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-dwkm7" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.220867 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/363f808f-a265-464a-a3a7-4128e3734b14-csi-data-dir\") pod \"csi-hostpathplugin-m8jlq\" (UID: \"363f808f-a265-464a-a3a7-4128e3734b14\") " pod="hostpath-provisioner/csi-hostpathplugin-m8jlq" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.221343 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/f3561d3e-0e57-4243-83df-9a475da69fb6-auth-proxy-config\") pod \"machine-config-operator-74547568cd-dwkm7\" (UID: \"f3561d3e-0e57-4243-83df-9a475da69fb6\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-dwkm7" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.221769 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/aa7f6b23-f07d-44a9-8963-c2edbbb83918-certs\") pod \"machine-config-server-ddnwl\" (UID: \"aa7f6b23-f07d-44a9-8963-c2edbbb83918\") " pod="openshift-machine-config-operator/machine-config-server-ddnwl" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.224720 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/c342090f-f331-484d-b287-929d8303f5d7-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-qsskc\" (UID: \"c342090f-f331-484d-b287-929d8303f5d7\") " pod="openshift-marketplace/marketplace-operator-79b997595-qsskc" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.224869 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/95e0dc94-8aec-42d0-8b19-7cdf12fa7dee-config-volume\") pod \"collect-profiles-29409630-45pn5\" (UID: \"95e0dc94-8aec-42d0-8b19-7cdf12fa7dee\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29409630-45pn5" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.226545 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/61054467-59d0-40a9-a60c-09c6c6106acd-config-volume\") pod \"dns-default-j7v65\" (UID: \"61054467-59d0-40a9-a60c-09c6c6106acd\") " pod="openshift-dns/dns-default-j7v65" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.227434 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-7q5lc"] Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.228051 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/57a1918e-f28a-4890-b88c-bc80df5cf704-config\") pod \"service-ca-operator-777779d784-dh427\" (UID: \"57a1918e-f28a-4890-b88c-bc80df5cf704\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-dh427" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.228068 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/525b7be7-bc47-4ff1-ada6-0521aa1c8620-signing-cabundle\") pod \"service-ca-9c57cc56f-6sclt\" (UID: \"525b7be7-bc47-4ff1-ada6-0521aa1c8620\") " pod="openshift-service-ca/service-ca-9c57cc56f-6sclt" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.228205 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/2e3f10d2-ac94-459c-9898-542e9f710747-tmpfs\") pod \"packageserver-d55dfcdfc-f8kzc\" (UID: \"2e3f10d2-ac94-459c-9898-542e9f710747\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-f8kzc" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.228223 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a7588f88-9a71-4e13-9fc3-a34cbfa6a090-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-5z87f\" (UID: \"a7588f88-9a71-4e13-9fc3-a34cbfa6a090\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-5z87f" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.228336 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/469840fe-2ac3-4431-9b2b-3b41451e6cbc-config\") pod \"kube-apiserver-operator-766d6c64bb-lfm7c\" (UID: \"469840fe-2ac3-4431-9b2b-3b41451e6cbc\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-lfm7c" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.229066 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/363f808f-a265-464a-a3a7-4128e3734b14-registration-dir\") pod \"csi-hostpathplugin-m8jlq\" (UID: \"363f808f-a265-464a-a3a7-4128e3734b14\") " pod="hostpath-provisioner/csi-hostpathplugin-m8jlq" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.229242 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/57a1918e-f28a-4890-b88c-bc80df5cf704-serving-cert\") pod \"service-ca-operator-777779d784-dh427\" (UID: \"57a1918e-f28a-4890-b88c-bc80df5cf704\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-dh427" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.229501 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/469840fe-2ac3-4431-9b2b-3b41451e6cbc-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-lfm7c\" (UID: \"469840fe-2ac3-4431-9b2b-3b41451e6cbc\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-lfm7c" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.229702 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/7caa1280-3fcc-47b0-ab38-8aa5479fcfce-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-l5xp9\" (UID: \"7caa1280-3fcc-47b0-ab38-8aa5479fcfce\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-l5xp9" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.230015 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-8274k" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.230105 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/489e9f98-0728-4a72-80d4-3d88a77823f5-profile-collector-cert\") pod \"olm-operator-6b444d44fb-hf8fp\" (UID: \"489e9f98-0728-4a72-80d4-3d88a77823f5\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-hf8fp" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.230399 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/f3561d3e-0e57-4243-83df-9a475da69fb6-images\") pod \"machine-config-operator-74547568cd-dwkm7\" (UID: \"f3561d3e-0e57-4243-83df-9a475da69fb6\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-dwkm7" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.231318 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/95e0dc94-8aec-42d0-8b19-7cdf12fa7dee-secret-volume\") pod \"collect-profiles-29409630-45pn5\" (UID: \"95e0dc94-8aec-42d0-8b19-7cdf12fa7dee\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29409630-45pn5" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.231847 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/489e9f98-0728-4a72-80d4-3d88a77823f5-srv-cert\") pod \"olm-operator-6b444d44fb-hf8fp\" (UID: \"489e9f98-0728-4a72-80d4-3d88a77823f5\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-hf8fp" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.232875 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-72d4v\" (UniqueName: \"kubernetes.io/projected/0f0340e8-faa5-447b-a5cd-a878e982d98f-kube-api-access-72d4v\") pod \"route-controller-manager-6576b87f9c-2mb7d\" (UID: \"0f0340e8-faa5-447b-a5cd-a878e982d98f\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-2mb7d" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.235632 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a2b231fa-0ecc-44da-9259-a7dc7261127c-bound-sa-token\") pod \"image-registry-697d97f7c8-mj64s\" (UID: \"a2b231fa-0ecc-44da-9259-a7dc7261127c\") " pod="openshift-image-registry/image-registry-697d97f7c8-mj64s" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.236569 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9p9xx\" (UniqueName: \"kubernetes.io/projected/a2b231fa-0ecc-44da-9259-a7dc7261127c-kube-api-access-9p9xx\") pod \"image-registry-697d97f7c8-mj64s\" (UID: \"a2b231fa-0ecc-44da-9259-a7dc7261127c\") " pod="openshift-image-registry/image-registry-697d97f7c8-mj64s" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.237640 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/61054467-59d0-40a9-a60c-09c6c6106acd-metrics-tls\") pod \"dns-default-j7v65\" (UID: \"61054467-59d0-40a9-a60c-09c6c6106acd\") " pod="openshift-dns/dns-default-j7v65" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.238078 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/a905919c-6c7e-40be-8f43-b938a60f14f6-default-certificate\") pod \"router-default-5444994796-sp4wh\" (UID: \"a905919c-6c7e-40be-8f43-b938a60f14f6\") " pod="openshift-ingress/router-default-5444994796-sp4wh" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.238147 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/2e3f10d2-ac94-459c-9898-542e9f710747-webhook-cert\") pod \"packageserver-d55dfcdfc-f8kzc\" (UID: \"2e3f10d2-ac94-459c-9898-542e9f710747\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-f8kzc" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.238295 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/5aa447ba-69b3-4291-9db5-25f4aeee2370-profile-collector-cert\") pod \"catalog-operator-68c6474976-n4dp7\" (UID: \"5aa447ba-69b3-4291-9db5-25f4aeee2370\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-n4dp7" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.238885 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/a905919c-6c7e-40be-8f43-b938a60f14f6-metrics-certs\") pod \"router-default-5444994796-sp4wh\" (UID: \"a905919c-6c7e-40be-8f43-b938a60f14f6\") " pod="openshift-ingress/router-default-5444994796-sp4wh" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.240194 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/2e3f10d2-ac94-459c-9898-542e9f710747-apiservice-cert\") pod \"packageserver-d55dfcdfc-f8kzc\" (UID: \"2e3f10d2-ac94-459c-9898-542e9f710747\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-f8kzc" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.244079 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/1e597d43-ddb6-4cdb-8ce7-94b7427fb2f0-proxy-tls\") pod \"machine-config-controller-84d6567774-vcwrd\" (UID: \"1e597d43-ddb6-4cdb-8ce7-94b7427fb2f0\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-vcwrd" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.244226 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/2fd313fd-c0b1-4aae-8b3b-2d078a7fc7d0-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-fxgbc\" (UID: \"2fd313fd-c0b1-4aae-8b3b-2d078a7fc7d0\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-fxgbc" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.244283 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/adf5f998-8117-4362-868e-b7cdbbf91e0e-cert\") pod \"ingress-canary-49bkd\" (UID: \"adf5f998-8117-4362-868e-b7cdbbf91e0e\") " pod="openshift-ingress-canary/ingress-canary-49bkd" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.244747 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/525b7be7-bc47-4ff1-ada6-0521aa1c8620-signing-key\") pod \"service-ca-9c57cc56f-6sclt\" (UID: \"525b7be7-bc47-4ff1-ada6-0521aa1c8620\") " pod="openshift-service-ca/service-ca-9c57cc56f-6sclt" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.244766 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lmpxp\" (UniqueName: \"kubernetes.io/projected/52120ed5-ff4f-4df5-8c60-3342df484be8-kube-api-access-lmpxp\") pod \"oauth-openshift-558db77b4-5fgn9\" (UID: \"52120ed5-ff4f-4df5-8c60-3342df484be8\") " pod="openshift-authentication/oauth-openshift-558db77b4-5fgn9" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.249350 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/7d654f9a-b190-47bc-89ce-0caae2ec1513-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-jb992\" (UID: \"7d654f9a-b190-47bc-89ce-0caae2ec1513\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-jb992" Dec 01 08:41:55 crc kubenswrapper[4813]: W1201 08:41:55.249758 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod59a1ff0e_6c38_4e97_8423_1641e5620570.slice/crio-36712eb5c923b301388efd9b953d1c4b62cbf79630482f5d43ff42d2f3fdce0c WatchSource:0}: Error finding container 36712eb5c923b301388efd9b953d1c4b62cbf79630482f5d43ff42d2f3fdce0c: Status 404 returned error can't find the container with id 36712eb5c923b301388efd9b953d1c4b62cbf79630482f5d43ff42d2f3fdce0c Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.269695 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ppsvf\" (UniqueName: \"kubernetes.io/projected/363f808f-a265-464a-a3a7-4128e3734b14-kube-api-access-ppsvf\") pod \"csi-hostpathplugin-m8jlq\" (UID: \"363f808f-a265-464a-a3a7-4128e3734b14\") " pod="hostpath-provisioner/csi-hostpathplugin-m8jlq" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.293762 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a7588f88-9a71-4e13-9fc3-a34cbfa6a090-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-5z87f\" (UID: \"a7588f88-9a71-4e13-9fc3-a34cbfa6a090\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-5z87f" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.342186 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 08:41:55 crc kubenswrapper[4813]: E1201 08:41:55.342733 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 08:41:55.842713193 +0000 UTC m=+100.465634789 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.385256 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h7wwf\" (UniqueName: \"kubernetes.io/projected/2fd313fd-c0b1-4aae-8b3b-2d078a7fc7d0-kube-api-access-h7wwf\") pod \"control-plane-machine-set-operator-78cbb6b69f-fxgbc\" (UID: \"2fd313fd-c0b1-4aae-8b3b-2d078a7fc7d0\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-fxgbc" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.395193 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t8jvf\" (UniqueName: \"kubernetes.io/projected/dadb7bca-672d-434b-897d-7eb83059f189-kube-api-access-t8jvf\") pod \"migrator-59844c95c7-sjjm7\" (UID: \"dadb7bca-672d-434b-897d-7eb83059f189\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-sjjm7" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.404937 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-49f6c\" (UniqueName: \"kubernetes.io/projected/57a1918e-f28a-4890-b88c-bc80df5cf704-kube-api-access-49f6c\") pod \"service-ca-operator-777779d784-dh427\" (UID: \"57a1918e-f28a-4890-b88c-bc80df5cf704\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-dh427" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.412505 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xpwm2\" (UniqueName: \"kubernetes.io/projected/5aa447ba-69b3-4291-9db5-25f4aeee2370-kube-api-access-xpwm2\") pod \"catalog-operator-68c6474976-n4dp7\" (UID: \"5aa447ba-69b3-4291-9db5-25f4aeee2370\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-n4dp7" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.412827 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-m8jlq" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.456481 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mj64s\" (UID: \"a2b231fa-0ecc-44da-9259-a7dc7261127c\") " pod="openshift-image-registry/image-registry-697d97f7c8-mj64s" Dec 01 08:41:55 crc kubenswrapper[4813]: E1201 08:41:55.457421 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 08:41:55.957398869 +0000 UTC m=+100.580320455 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mj64s" (UID: "a2b231fa-0ecc-44da-9259-a7dc7261127c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.458582 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-5fgn9" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.459237 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-7w2kz" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.460117 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-2mb7d" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.472436 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5fvsv\" (UniqueName: \"kubernetes.io/projected/2e3f10d2-ac94-459c-9898-542e9f710747-kube-api-access-5fvsv\") pod \"packageserver-d55dfcdfc-f8kzc\" (UID: \"2e3f10d2-ac94-459c-9898-542e9f710747\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-f8kzc" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.479185 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gjfw8\" (UniqueName: \"kubernetes.io/projected/95e0dc94-8aec-42d0-8b19-7cdf12fa7dee-kube-api-access-gjfw8\") pod \"collect-profiles-29409630-45pn5\" (UID: \"95e0dc94-8aec-42d0-8b19-7cdf12fa7dee\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29409630-45pn5" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.485685 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6t6bj\" (UniqueName: \"kubernetes.io/projected/adf5f998-8117-4362-868e-b7cdbbf91e0e-kube-api-access-6t6bj\") pod \"ingress-canary-49bkd\" (UID: \"adf5f998-8117-4362-868e-b7cdbbf91e0e\") " pod="openshift-ingress-canary/ingress-canary-49bkd" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.486167 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-66c2t\" (UniqueName: \"kubernetes.io/projected/489e9f98-0728-4a72-80d4-3d88a77823f5-kube-api-access-66c2t\") pod \"olm-operator-6b444d44fb-hf8fp\" (UID: \"489e9f98-0728-4a72-80d4-3d88a77823f5\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-hf8fp" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.490528 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s9rtm\" (UniqueName: \"kubernetes.io/projected/aa7f6b23-f07d-44a9-8963-c2edbbb83918-kube-api-access-s9rtm\") pod \"machine-config-server-ddnwl\" (UID: \"aa7f6b23-f07d-44a9-8963-c2edbbb83918\") " pod="openshift-machine-config-operator/machine-config-server-ddnwl" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.495952 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-sjjm7" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.499112 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nr7m5\" (UniqueName: \"kubernetes.io/projected/7d654f9a-b190-47bc-89ce-0caae2ec1513-kube-api-access-nr7m5\") pod \"multus-admission-controller-857f4d67dd-jb992\" (UID: \"7d654f9a-b190-47bc-89ce-0caae2ec1513\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-jb992" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.514841 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdvtz\" (UniqueName: \"kubernetes.io/projected/7caa1280-3fcc-47b0-ab38-8aa5479fcfce-kube-api-access-rdvtz\") pod \"package-server-manager-789f6589d5-l5xp9\" (UID: \"7caa1280-3fcc-47b0-ab38-8aa5479fcfce\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-l5xp9" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.522364 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-fxgbc" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.535621 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdgwf\" (UniqueName: \"kubernetes.io/projected/525b7be7-bc47-4ff1-ada6-0521aa1c8620-kube-api-access-rdgwf\") pod \"service-ca-9c57cc56f-6sclt\" (UID: \"525b7be7-bc47-4ff1-ada6-0521aa1c8620\") " pod="openshift-service-ca/service-ca-9c57cc56f-6sclt" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.539188 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-jb992" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.548464 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-f8kzc" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.554630 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-5z87f" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.561370 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-hf8fp" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.562103 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p66qw\" (UniqueName: \"kubernetes.io/projected/61054467-59d0-40a9-a60c-09c6c6106acd-kube-api-access-p66qw\") pod \"dns-default-j7v65\" (UID: \"61054467-59d0-40a9-a60c-09c6c6106acd\") " pod="openshift-dns/dns-default-j7v65" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.567791 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 08:41:55 crc kubenswrapper[4813]: E1201 08:41:55.568139 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 08:41:56.068121376 +0000 UTC m=+100.691042962 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.577486 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-l5xp9" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.595671 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5krct\" (UniqueName: \"kubernetes.io/projected/f3561d3e-0e57-4243-83df-9a475da69fb6-kube-api-access-5krct\") pod \"machine-config-operator-74547568cd-dwkm7\" (UID: \"f3561d3e-0e57-4243-83df-9a475da69fb6\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-dwkm7" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.595761 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-n4dp7" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.601598 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nvbp8\" (UniqueName: \"kubernetes.io/projected/1e597d43-ddb6-4cdb-8ce7-94b7427fb2f0-kube-api-access-nvbp8\") pod \"machine-config-controller-84d6567774-vcwrd\" (UID: \"1e597d43-ddb6-4cdb-8ce7-94b7427fb2f0\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-vcwrd" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.613116 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-njwpj\" (UniqueName: \"kubernetes.io/projected/a905919c-6c7e-40be-8f43-b938a60f14f6-kube-api-access-njwpj\") pod \"router-default-5444994796-sp4wh\" (UID: \"a905919c-6c7e-40be-8f43-b938a60f14f6\") " pod="openshift-ingress/router-default-5444994796-sp4wh" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.694726 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-6sclt" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.694779 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-49bkd" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.695250 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-dh427" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.696097 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mj64s\" (UID: \"a2b231fa-0ecc-44da-9259-a7dc7261127c\") " pod="openshift-image-registry/image-registry-697d97f7c8-mj64s" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.696243 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-j7v65" Dec 01 08:41:55 crc kubenswrapper[4813]: E1201 08:41:55.696438 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 08:41:56.196423241 +0000 UTC m=+100.819344827 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mj64s" (UID: "a2b231fa-0ecc-44da-9259-a7dc7261127c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.696598 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29409630-45pn5" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.697270 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/469840fe-2ac3-4431-9b2b-3b41451e6cbc-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-lfm7c\" (UID: \"469840fe-2ac3-4431-9b2b-3b41451e6cbc\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-lfm7c" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.746579 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-ddnwl" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.801193 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-lfm7c" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.804341 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 08:41:55 crc kubenswrapper[4813]: E1201 08:41:55.804925 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 08:41:56.304907381 +0000 UTC m=+100.927828987 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.805756 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jm2f5\" (UniqueName: \"kubernetes.io/projected/c342090f-f331-484d-b287-929d8303f5d7-kube-api-access-jm2f5\") pod \"marketplace-operator-79b997595-qsskc\" (UID: \"c342090f-f331-484d-b287-929d8303f5d7\") " pod="openshift-marketplace/marketplace-operator-79b997595-qsskc" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.808178 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-sp4wh" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.819528 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-qsskc" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.833515 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-dwkm7" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.888199 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-vcwrd" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.906569 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mj64s\" (UID: \"a2b231fa-0ecc-44da-9259-a7dc7261127c\") " pod="openshift-image-registry/image-registry-697d97f7c8-mj64s" Dec 01 08:41:55 crc kubenswrapper[4813]: E1201 08:41:55.907009 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 08:41:56.406992628 +0000 UTC m=+101.029914224 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mj64s" (UID: "a2b231fa-0ecc-44da-9259-a7dc7261127c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.954220 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xs4nq" podStartSLOduration=75.954197554 podStartE2EDuration="1m15.954197554s" podCreationTimestamp="2025-12-01 08:40:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:41:55.950728339 +0000 UTC m=+100.573649935" watchObservedRunningTime="2025-12-01 08:41:55.954197554 +0000 UTC m=+100.577119140" Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.974610 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-wqkxc" event={"ID":"156d5b06-7b82-4444-817f-c7bf13cb3c44","Type":"ContainerStarted","Data":"c64da6e1a414bd8ab8ea1fa6c93edba9726a289e9fd9f39925d142960addf860"} Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.974668 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-wqkxc" event={"ID":"156d5b06-7b82-4444-817f-c7bf13cb3c44","Type":"ContainerStarted","Data":"99624914f2863f51592983e15a317b4fe132cf3ec663077b4534638f9320c7c9"} Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.976729 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-dvhpg" event={"ID":"df338eea-faa9-4dca-9c75-f12a2bbf87a4","Type":"ContainerStarted","Data":"183fb9f303ee6f0231762e444be8acb40bbf7fa7e5ea71db5707c69a7898a3de"} Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.976760 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-dvhpg" event={"ID":"df338eea-faa9-4dca-9c75-f12a2bbf87a4","Type":"ContainerStarted","Data":"06eac8954cc9eb28eb5b566e40d3d56b28a6be87225cc1d5cfae01e1debba715"} Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.980125 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-8274k"] Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.984747 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-hmh5v" event={"ID":"ab324c2b-a49d-4874-b9c0-a869b226ce98","Type":"ContainerStarted","Data":"89dccaf22e4249e3f80a392da053bae578eb943649fb2740e85dae3b9a25c8c5"} Dec 01 08:41:55 crc kubenswrapper[4813]: I1201 08:41:55.984787 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-hmh5v" event={"ID":"ab324c2b-a49d-4874-b9c0-a869b226ce98","Type":"ContainerStarted","Data":"ae6286f1e76f3ebe423aa5d6d8490fe33845a7a9830557ca274afa738c26223b"} Dec 01 08:41:56 crc kubenswrapper[4813]: I1201 08:41:55.997041 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-7q5lc" event={"ID":"59a1ff0e-6c38-4e97-8423-1641e5620570","Type":"ContainerStarted","Data":"3888bb11e948b6771975ed2f401e3bde33cd73cfeaf3d7ad81112830dbb6a02c"} Dec 01 08:41:56 crc kubenswrapper[4813]: I1201 08:41:55.997083 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-7q5lc" event={"ID":"59a1ff0e-6c38-4e97-8423-1641e5620570","Type":"ContainerStarted","Data":"36712eb5c923b301388efd9b953d1c4b62cbf79630482f5d43ff42d2f3fdce0c"} Dec 01 08:41:56 crc kubenswrapper[4813]: I1201 08:41:56.002570 4813 generic.go:334] "Generic (PLEG): container finished" podID="d2aac47a-b117-42c7-afa3-2c989d9536a5" containerID="1e3491836991afee1216fe2cfd188e02e51dc87d73f2657ecab8bbe29bb6d6ae" exitCode=0 Dec 01 08:41:56 crc kubenswrapper[4813]: I1201 08:41:56.002629 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-5ptsh" event={"ID":"d2aac47a-b117-42c7-afa3-2c989d9536a5","Type":"ContainerDied","Data":"1e3491836991afee1216fe2cfd188e02e51dc87d73f2657ecab8bbe29bb6d6ae"} Dec 01 08:41:56 crc kubenswrapper[4813]: I1201 08:41:56.009262 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 08:41:56 crc kubenswrapper[4813]: E1201 08:41:56.009644 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 08:41:56.50960719 +0000 UTC m=+101.132528776 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:41:56 crc kubenswrapper[4813]: I1201 08:41:56.009702 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mj64s\" (UID: \"a2b231fa-0ecc-44da-9259-a7dc7261127c\") " pod="openshift-image-registry/image-registry-697d97f7c8-mj64s" Dec 01 08:41:56 crc kubenswrapper[4813]: E1201 08:41:56.010180 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 08:41:56.510170517 +0000 UTC m=+101.133092103 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mj64s" (UID: "a2b231fa-0ecc-44da-9259-a7dc7261127c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:41:56 crc kubenswrapper[4813]: I1201 08:41:56.016862 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-s4vx5" event={"ID":"cf79ef17-ec75-4d0b-b665-ce24618a92df","Type":"ContainerStarted","Data":"8a1ff936fe936e75bd6af3806f4adcecd228f37c1382a29fffdf338f8622460b"} Dec 01 08:41:56 crc kubenswrapper[4813]: I1201 08:41:56.026441 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-ct7w8" event={"ID":"e17c11d6-eb41-4391-b513-a817eeffcdb4","Type":"ContainerStarted","Data":"7884792cc3b07ecb84a0069960782aefaa7b4569546e6f46c981d16c050b9c7b"} Dec 01 08:41:56 crc kubenswrapper[4813]: I1201 08:41:56.030852 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-9cxt7" event={"ID":"159a2fad-3042-43e6-8df3-16f1157db0b1","Type":"ContainerStarted","Data":"6398604c120564edceb7be97bc3dacb0d08205abb6d56a1693ee845bea4f6242"} Dec 01 08:41:56 crc kubenswrapper[4813]: I1201 08:41:56.042058 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lt4dx" event={"ID":"7f38918f-c551-4e3f-bb5e-d5dcac6ac826","Type":"ContainerStarted","Data":"fbe2c8949dc3228880746e2e5b7c2aa9a09f36a7fa24da151e83a66aec4e4310"} Dec 01 08:41:56 crc kubenswrapper[4813]: I1201 08:41:56.042100 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lt4dx" event={"ID":"7f38918f-c551-4e3f-bb5e-d5dcac6ac826","Type":"ContainerStarted","Data":"e71f66209edcf962c9a7df7b09c83b0d5f2e30cc4b5bf369941da9df4446b579"} Dec 01 08:41:56 crc kubenswrapper[4813]: I1201 08:41:56.048255 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-gpd6h" event={"ID":"4828f51e-d209-4d31-a608-7d0951d311f4","Type":"ContainerStarted","Data":"1b2cb11309ebd7c42494cb55230bdd2cdddb7e67f63ddfd4aa7839e8fba5b69c"} Dec 01 08:41:56 crc kubenswrapper[4813]: I1201 08:41:56.066451 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-n72r5" event={"ID":"b35b7145-810b-4945-9819-6d9bb650ec73","Type":"ContainerStarted","Data":"ef59fd738ba314d53a3728bc2b905a01614de2a0b9efd08e0438b8d1c9dd89e8"} Dec 01 08:41:56 crc kubenswrapper[4813]: I1201 08:41:56.066495 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-n72r5" event={"ID":"b35b7145-810b-4945-9819-6d9bb650ec73","Type":"ContainerStarted","Data":"9cc32a8fd67a5a74ef25d38056d469a8c4a8d8d6e603fea3731dc3be34635c56"} Dec 01 08:41:56 crc kubenswrapper[4813]: I1201 08:41:56.069262 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-vkbsg" event={"ID":"3c957bbc-fa1d-4188-9f9f-0b67be4c56c7","Type":"ContainerStarted","Data":"ca828ba09e5435928c66fd023df3a6a924f997dad5e6a9d85d24801fe8bc1e93"} Dec 01 08:41:56 crc kubenswrapper[4813]: I1201 08:41:56.069307 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-vkbsg" event={"ID":"3c957bbc-fa1d-4188-9f9f-0b67be4c56c7","Type":"ContainerStarted","Data":"e211699f23702d1a8144e129e3d2827263d317f25ed14ba9051acf24a30b91d2"} Dec 01 08:41:56 crc kubenswrapper[4813]: I1201 08:41:56.077716 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-wtwtl" event={"ID":"0e3d53a2-6105-4c2f-a74b-0b8dcd3b0706","Type":"ContainerStarted","Data":"52bc4541d50e9464c81bad6c566fede685b8ade11b2cd9829a0df05cd1f6bf08"} Dec 01 08:41:56 crc kubenswrapper[4813]: I1201 08:41:56.079191 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-wtwtl" event={"ID":"0e3d53a2-6105-4c2f-a74b-0b8dcd3b0706","Type":"ContainerStarted","Data":"bf6ae0008dbe1f3eac61299ba1b0bcb828b15b6f300facfb26e755552ae151ab"} Dec 01 08:41:56 crc kubenswrapper[4813]: I1201 08:41:56.079250 4813 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-l4zh5 container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.8:8443/healthz\": dial tcp 10.217.0.8:8443: connect: connection refused" start-of-body= Dec 01 08:41:56 crc kubenswrapper[4813]: I1201 08:41:56.079325 4813 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-l4zh5" podUID="705b2456-4ba7-4775-84ca-4dcea64b6755" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.8:8443/healthz\": dial tcp 10.217.0.8:8443: connect: connection refused" Dec 01 08:41:56 crc kubenswrapper[4813]: I1201 08:41:56.112465 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 08:41:56 crc kubenswrapper[4813]: E1201 08:41:56.114058 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 08:41:56.614033368 +0000 UTC m=+101.236954954 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:41:56 crc kubenswrapper[4813]: W1201 08:41:56.127229 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podaa7f6b23_f07d_44a9_8963_c2edbbb83918.slice/crio-8ac2af2664f23d243572bcf50398580d87292bc9b943dfa97ad8aa7dc91c90ab WatchSource:0}: Error finding container 8ac2af2664f23d243572bcf50398580d87292bc9b943dfa97ad8aa7dc91c90ab: Status 404 returned error can't find the container with id 8ac2af2664f23d243572bcf50398580d87292bc9b943dfa97ad8aa7dc91c90ab Dec 01 08:41:56 crc kubenswrapper[4813]: I1201 08:41:56.303091 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mj64s\" (UID: \"a2b231fa-0ecc-44da-9259-a7dc7261127c\") " pod="openshift-image-registry/image-registry-697d97f7c8-mj64s" Dec 01 08:41:56 crc kubenswrapper[4813]: E1201 08:41:56.303711 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 08:41:56.803694341 +0000 UTC m=+101.426615927 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mj64s" (UID: "a2b231fa-0ecc-44da-9259-a7dc7261127c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:41:56 crc kubenswrapper[4813]: I1201 08:41:56.407170 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 08:41:56 crc kubenswrapper[4813]: E1201 08:41:56.408731 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 08:41:56.908708487 +0000 UTC m=+101.531630073 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:41:56 crc kubenswrapper[4813]: I1201 08:41:56.509694 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mj64s\" (UID: \"a2b231fa-0ecc-44da-9259-a7dc7261127c\") " pod="openshift-image-registry/image-registry-697d97f7c8-mj64s" Dec 01 08:41:56 crc kubenswrapper[4813]: E1201 08:41:56.510149 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 08:41:57.010120472 +0000 UTC m=+101.633042058 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mj64s" (UID: "a2b231fa-0ecc-44da-9259-a7dc7261127c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:41:56 crc kubenswrapper[4813]: I1201 08:41:56.616526 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 08:41:56 crc kubenswrapper[4813]: E1201 08:41:56.616929 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 08:41:57.116912633 +0000 UTC m=+101.739834219 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:41:56 crc kubenswrapper[4813]: I1201 08:41:56.803929 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mj64s\" (UID: \"a2b231fa-0ecc-44da-9259-a7dc7261127c\") " pod="openshift-image-registry/image-registry-697d97f7c8-mj64s" Dec 01 08:41:56 crc kubenswrapper[4813]: E1201 08:41:56.804316 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 08:41:57.304304238 +0000 UTC m=+101.927225824 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mj64s" (UID: "a2b231fa-0ecc-44da-9259-a7dc7261127c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:41:56 crc kubenswrapper[4813]: I1201 08:41:56.905030 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 08:41:56 crc kubenswrapper[4813]: E1201 08:41:56.905527 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 08:41:57.405508183 +0000 UTC m=+102.028429769 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:41:56 crc kubenswrapper[4813]: I1201 08:41:56.905612 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mj64s\" (UID: \"a2b231fa-0ecc-44da-9259-a7dc7261127c\") " pod="openshift-image-registry/image-registry-697d97f7c8-mj64s" Dec 01 08:41:56 crc kubenswrapper[4813]: E1201 08:41:56.905974 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 08:41:57.405955764 +0000 UTC m=+102.028877350 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mj64s" (UID: "a2b231fa-0ecc-44da-9259-a7dc7261127c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:41:57 crc kubenswrapper[4813]: I1201 08:41:57.008896 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 08:41:57 crc kubenswrapper[4813]: E1201 08:41:57.009225 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 08:41:57.509207346 +0000 UTC m=+102.132128922 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:41:57 crc kubenswrapper[4813]: I1201 08:41:57.065074 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-7w2kz"] Dec 01 08:41:57 crc kubenswrapper[4813]: I1201 08:41:57.112629 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mj64s\" (UID: \"a2b231fa-0ecc-44da-9259-a7dc7261127c\") " pod="openshift-image-registry/image-registry-697d97f7c8-mj64s" Dec 01 08:41:57 crc kubenswrapper[4813]: E1201 08:41:57.113490 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 08:41:57.613478087 +0000 UTC m=+102.236399673 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mj64s" (UID: "a2b231fa-0ecc-44da-9259-a7dc7261127c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:41:57 crc kubenswrapper[4813]: I1201 08:41:57.113540 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-wqkxc" event={"ID":"156d5b06-7b82-4444-817f-c7bf13cb3c44","Type":"ContainerStarted","Data":"2909898ec839b92ab19478efa3dc0761e4b3cd49af98ed6cda14c903e962d418"} Dec 01 08:41:57 crc kubenswrapper[4813]: I1201 08:41:57.144449 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-5fgn9"] Dec 01 08:41:57 crc kubenswrapper[4813]: I1201 08:41:57.146278 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-ddnwl" event={"ID":"aa7f6b23-f07d-44a9-8963-c2edbbb83918","Type":"ContainerStarted","Data":"3fafdc3b058bd12a6da833132f099d4f811f61f282f7897cae48ae633f3ff400"} Dec 01 08:41:57 crc kubenswrapper[4813]: I1201 08:41:57.146331 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-ddnwl" event={"ID":"aa7f6b23-f07d-44a9-8963-c2edbbb83918","Type":"ContainerStarted","Data":"8ac2af2664f23d243572bcf50398580d87292bc9b943dfa97ad8aa7dc91c90ab"} Dec 01 08:41:57 crc kubenswrapper[4813]: I1201 08:41:57.155768 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-dvhpg" event={"ID":"df338eea-faa9-4dca-9c75-f12a2bbf87a4","Type":"ContainerStarted","Data":"c0e03440fff4a781a760b0a31044e1bb365aae17f7ed11f29864098b0274886c"} Dec 01 08:41:57 crc kubenswrapper[4813]: I1201 08:41:57.165683 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-sp4wh" event={"ID":"a905919c-6c7e-40be-8f43-b938a60f14f6","Type":"ContainerStarted","Data":"3afe8c054cf9eb618cd8139ad4a194860189f73e430f37d29508f739c866f452"} Dec 01 08:41:57 crc kubenswrapper[4813]: I1201 08:41:57.183232 4813 generic.go:334] "Generic (PLEG): container finished" podID="7f38918f-c551-4e3f-bb5e-d5dcac6ac826" containerID="fbe2c8949dc3228880746e2e5b7c2aa9a09f36a7fa24da151e83a66aec4e4310" exitCode=0 Dec 01 08:41:57 crc kubenswrapper[4813]: I1201 08:41:57.183361 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lt4dx" event={"ID":"7f38918f-c551-4e3f-bb5e-d5dcac6ac826","Type":"ContainerDied","Data":"fbe2c8949dc3228880746e2e5b7c2aa9a09f36a7fa24da151e83a66aec4e4310"} Dec 01 08:41:57 crc kubenswrapper[4813]: I1201 08:41:57.186400 4813 generic.go:334] "Generic (PLEG): container finished" podID="ab324c2b-a49d-4874-b9c0-a869b226ce98" containerID="89dccaf22e4249e3f80a392da053bae578eb943649fb2740e85dae3b9a25c8c5" exitCode=0 Dec 01 08:41:57 crc kubenswrapper[4813]: I1201 08:41:57.186469 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-hmh5v" event={"ID":"ab324c2b-a49d-4874-b9c0-a869b226ce98","Type":"ContainerDied","Data":"89dccaf22e4249e3f80a392da053bae578eb943649fb2740e85dae3b9a25c8c5"} Dec 01 08:41:57 crc kubenswrapper[4813]: I1201 08:41:57.215902 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-8274k" event={"ID":"57c3bc40-b325-4343-a01b-6705e9e8eb77","Type":"ContainerStarted","Data":"b6b5d773a9d75020216d88df031f85f23730d49d1cde29623bdec3b99ee892d6"} Dec 01 08:41:57 crc kubenswrapper[4813]: I1201 08:41:57.229923 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 08:41:57 crc kubenswrapper[4813]: E1201 08:41:57.230248 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 08:41:57.730225931 +0000 UTC m=+102.353147517 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:41:57 crc kubenswrapper[4813]: I1201 08:41:57.233982 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mj64s\" (UID: \"a2b231fa-0ecc-44da-9259-a7dc7261127c\") " pod="openshift-image-registry/image-registry-697d97f7c8-mj64s" Dec 01 08:41:57 crc kubenswrapper[4813]: E1201 08:41:57.235474 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 08:41:57.73545615 +0000 UTC m=+102.358377746 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mj64s" (UID: "a2b231fa-0ecc-44da-9259-a7dc7261127c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:41:57 crc kubenswrapper[4813]: I1201 08:41:57.342507 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 08:41:57 crc kubenswrapper[4813]: E1201 08:41:57.343039 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 08:41:57.843014927 +0000 UTC m=+102.465936513 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:41:57 crc kubenswrapper[4813]: I1201 08:41:57.343193 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mj64s\" (UID: \"a2b231fa-0ecc-44da-9259-a7dc7261127c\") " pod="openshift-image-registry/image-registry-697d97f7c8-mj64s" Dec 01 08:41:57 crc kubenswrapper[4813]: E1201 08:41:57.343524 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 08:41:57.843514661 +0000 UTC m=+102.466436247 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mj64s" (UID: "a2b231fa-0ecc-44da-9259-a7dc7261127c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:41:57 crc kubenswrapper[4813]: I1201 08:41:57.356919 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-s4vx5" event={"ID":"cf79ef17-ec75-4d0b-b665-ce24618a92df","Type":"ContainerStarted","Data":"ebf26999fffcacccc1e03e4d3d38b3168bf36075bab3e046e6748fb97df48792"} Dec 01 08:41:57 crc kubenswrapper[4813]: I1201 08:41:57.358066 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-ct7w8" Dec 01 08:41:57 crc kubenswrapper[4813]: I1201 08:41:57.388263 4813 patch_prober.go:28] interesting pod/downloads-7954f5f757-ct7w8 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" start-of-body= Dec 01 08:41:57 crc kubenswrapper[4813]: I1201 08:41:57.388311 4813 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-ct7w8" podUID="e17c11d6-eb41-4391-b513-a817eeffcdb4" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" Dec 01 08:41:57 crc kubenswrapper[4813]: I1201 08:41:57.406226 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-gpd6h" podStartSLOduration=77.406202023 podStartE2EDuration="1m17.406202023s" podCreationTimestamp="2025-12-01 08:40:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:41:57.405950681 +0000 UTC m=+102.028872267" watchObservedRunningTime="2025-12-01 08:41:57.406202023 +0000 UTC m=+102.029123609" Dec 01 08:41:57 crc kubenswrapper[4813]: I1201 08:41:57.446246 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 08:41:57 crc kubenswrapper[4813]: E1201 08:41:57.446328 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 08:41:57.946311871 +0000 UTC m=+102.569233447 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:41:57 crc kubenswrapper[4813]: I1201 08:41:57.446944 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mj64s\" (UID: \"a2b231fa-0ecc-44da-9259-a7dc7261127c\") " pod="openshift-image-registry/image-registry-697d97f7c8-mj64s" Dec 01 08:41:57 crc kubenswrapper[4813]: E1201 08:41:57.448530 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 08:41:57.948515496 +0000 UTC m=+102.571437082 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mj64s" (UID: "a2b231fa-0ecc-44da-9259-a7dc7261127c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:41:57 crc kubenswrapper[4813]: I1201 08:41:57.477767 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-wtwtl" podStartSLOduration=77.477749117 podStartE2EDuration="1m17.477749117s" podCreationTimestamp="2025-12-01 08:40:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:41:57.476383812 +0000 UTC m=+102.099305408" watchObservedRunningTime="2025-12-01 08:41:57.477749117 +0000 UTC m=+102.100670703" Dec 01 08:41:57 crc kubenswrapper[4813]: I1201 08:41:57.549152 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 08:41:57 crc kubenswrapper[4813]: E1201 08:41:57.549715 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 08:41:58.04969926 +0000 UTC m=+102.672620846 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:41:57 crc kubenswrapper[4813]: E1201 08:41:57.552215 4813 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podab324c2b_a49d_4874_b9c0_a869b226ce98.slice/crio-conmon-89dccaf22e4249e3f80a392da053bae578eb943649fb2740e85dae3b9a25c8c5.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podab324c2b_a49d_4874_b9c0_a869b226ce98.slice/crio-89dccaf22e4249e3f80a392da053bae578eb943649fb2740e85dae3b9a25c8c5.scope\": RecentStats: unable to find data in memory cache]" Dec 01 08:41:57 crc kubenswrapper[4813]: I1201 08:41:57.577539 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-9cxt7" podStartSLOduration=77.577516893 podStartE2EDuration="1m17.577516893s" podCreationTimestamp="2025-12-01 08:40:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:41:57.542238955 +0000 UTC m=+102.165160541" watchObservedRunningTime="2025-12-01 08:41:57.577516893 +0000 UTC m=+102.200438479" Dec 01 08:41:57 crc kubenswrapper[4813]: I1201 08:41:57.650271 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mj64s\" (UID: \"a2b231fa-0ecc-44da-9259-a7dc7261127c\") " pod="openshift-image-registry/image-registry-697d97f7c8-mj64s" Dec 01 08:41:57 crc kubenswrapper[4813]: E1201 08:41:57.650599 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 08:41:58.15058892 +0000 UTC m=+102.773510506 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mj64s" (UID: "a2b231fa-0ecc-44da-9259-a7dc7261127c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:41:57 crc kubenswrapper[4813]: I1201 08:41:57.679077 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-ddnwl" podStartSLOduration=5.679052034 podStartE2EDuration="5.679052034s" podCreationTimestamp="2025-12-01 08:41:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:41:57.630331296 +0000 UTC m=+102.253252882" watchObservedRunningTime="2025-12-01 08:41:57.679052034 +0000 UTC m=+102.301973610" Dec 01 08:41:57 crc kubenswrapper[4813]: I1201 08:41:57.681382 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-9ql8j" podStartSLOduration=77.681364394 podStartE2EDuration="1m17.681364394s" podCreationTimestamp="2025-12-01 08:40:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:41:57.679293955 +0000 UTC m=+102.302215541" watchObservedRunningTime="2025-12-01 08:41:57.681364394 +0000 UTC m=+102.304285980" Dec 01 08:41:57 crc kubenswrapper[4813]: I1201 08:41:57.713059 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-s4vx5" podStartSLOduration=77.713038931 podStartE2EDuration="1m17.713038931s" podCreationTimestamp="2025-12-01 08:40:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:41:57.709782726 +0000 UTC m=+102.332704312" watchObservedRunningTime="2025-12-01 08:41:57.713038931 +0000 UTC m=+102.335960517" Dec 01 08:41:57 crc kubenswrapper[4813]: I1201 08:41:57.750907 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 08:41:57 crc kubenswrapper[4813]: E1201 08:41:57.751366 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 08:41:58.251346093 +0000 UTC m=+102.874267679 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:41:57 crc kubenswrapper[4813]: I1201 08:41:57.759309 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-k9t5k" podStartSLOduration=77.759291811 podStartE2EDuration="1m17.759291811s" podCreationTimestamp="2025-12-01 08:40:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:41:57.757301977 +0000 UTC m=+102.380223553" watchObservedRunningTime="2025-12-01 08:41:57.759291811 +0000 UTC m=+102.382213397" Dec 01 08:41:57 crc kubenswrapper[4813]: I1201 08:41:57.809045 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-gnd5s" podStartSLOduration=77.809022907 podStartE2EDuration="1m17.809022907s" podCreationTimestamp="2025-12-01 08:40:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:41:57.78868571 +0000 UTC m=+102.411607296" watchObservedRunningTime="2025-12-01 08:41:57.809022907 +0000 UTC m=+102.431944493" Dec 01 08:41:57 crc kubenswrapper[4813]: I1201 08:41:57.810421 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-l4zh5" podStartSLOduration=77.810416513 podStartE2EDuration="1m17.810416513s" podCreationTimestamp="2025-12-01 08:40:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:41:57.809698309 +0000 UTC m=+102.432619895" watchObservedRunningTime="2025-12-01 08:41:57.810416513 +0000 UTC m=+102.433338099" Dec 01 08:41:57 crc kubenswrapper[4813]: I1201 08:41:57.826652 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-n72r5" podStartSLOduration=77.826633225 podStartE2EDuration="1m17.826633225s" podCreationTimestamp="2025-12-01 08:40:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:41:57.822405384 +0000 UTC m=+102.445326970" watchObservedRunningTime="2025-12-01 08:41:57.826633225 +0000 UTC m=+102.449554811" Dec 01 08:41:57 crc kubenswrapper[4813]: I1201 08:41:57.849672 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-vkbsg" podStartSLOduration=77.84964856 podStartE2EDuration="1m17.84964856s" podCreationTimestamp="2025-12-01 08:40:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:41:57.849393348 +0000 UTC m=+102.472314934" watchObservedRunningTime="2025-12-01 08:41:57.84964856 +0000 UTC m=+102.472570146" Dec 01 08:41:57 crc kubenswrapper[4813]: I1201 08:41:57.850091 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-dvhpg" podStartSLOduration=77.850084481 podStartE2EDuration="1m17.850084481s" podCreationTimestamp="2025-12-01 08:40:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:41:57.836317226 +0000 UTC m=+102.459238812" watchObservedRunningTime="2025-12-01 08:41:57.850084481 +0000 UTC m=+102.473006067" Dec 01 08:41:57 crc kubenswrapper[4813]: I1201 08:41:57.853145 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mj64s\" (UID: \"a2b231fa-0ecc-44da-9259-a7dc7261127c\") " pod="openshift-image-registry/image-registry-697d97f7c8-mj64s" Dec 01 08:41:57 crc kubenswrapper[4813]: E1201 08:41:57.853476 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 08:41:58.353457161 +0000 UTC m=+102.976378767 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mj64s" (UID: "a2b231fa-0ecc-44da-9259-a7dc7261127c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:41:57 crc kubenswrapper[4813]: I1201 08:41:57.891628 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-7q5lc" podStartSLOduration=77.891606776 podStartE2EDuration="1m17.891606776s" podCreationTimestamp="2025-12-01 08:40:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:41:57.889056375 +0000 UTC m=+102.511977961" watchObservedRunningTime="2025-12-01 08:41:57.891606776 +0000 UTC m=+102.514528362" Dec 01 08:41:57 crc kubenswrapper[4813]: I1201 08:41:57.933176 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-ct7w8" podStartSLOduration=77.933156023 podStartE2EDuration="1m17.933156023s" podCreationTimestamp="2025-12-01 08:40:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:41:57.928237739 +0000 UTC m=+102.551159325" watchObservedRunningTime="2025-12-01 08:41:57.933156023 +0000 UTC m=+102.556077609" Dec 01 08:41:57 crc kubenswrapper[4813]: I1201 08:41:57.956567 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 08:41:57 crc kubenswrapper[4813]: E1201 08:41:57.956682 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 08:41:58.456656411 +0000 UTC m=+103.079577997 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:41:57 crc kubenswrapper[4813]: I1201 08:41:57.956769 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mj64s\" (UID: \"a2b231fa-0ecc-44da-9259-a7dc7261127c\") " pod="openshift-image-registry/image-registry-697d97f7c8-mj64s" Dec 01 08:41:57 crc kubenswrapper[4813]: E1201 08:41:57.957234 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 08:41:58.457222488 +0000 UTC m=+103.080144074 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mj64s" (UID: "a2b231fa-0ecc-44da-9259-a7dc7261127c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:41:58 crc kubenswrapper[4813]: I1201 08:41:58.058526 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 08:41:58 crc kubenswrapper[4813]: E1201 08:41:58.059086 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 08:41:58.559064923 +0000 UTC m=+103.181986509 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:41:58 crc kubenswrapper[4813]: I1201 08:41:58.159946 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mj64s\" (UID: \"a2b231fa-0ecc-44da-9259-a7dc7261127c\") " pod="openshift-image-registry/image-registry-697d97f7c8-mj64s" Dec 01 08:41:58 crc kubenswrapper[4813]: E1201 08:41:58.160781 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 08:41:58.660764991 +0000 UTC m=+103.283686577 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mj64s" (UID: "a2b231fa-0ecc-44da-9259-a7dc7261127c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:41:58 crc kubenswrapper[4813]: I1201 08:41:58.263594 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 08:41:58 crc kubenswrapper[4813]: E1201 08:41:58.264091 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 08:41:58.764074896 +0000 UTC m=+103.386996482 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:41:58 crc kubenswrapper[4813]: I1201 08:41:58.275918 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-2mb7d"] Dec 01 08:41:58 crc kubenswrapper[4813]: I1201 08:41:58.351575 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-8274k" event={"ID":"57c3bc40-b325-4343-a01b-6705e9e8eb77","Type":"ContainerStarted","Data":"3c6251908e3ad1b073a5850f6a20633f31e2289582676365ae67d37785e0b57f"} Dec 01 08:41:58 crc kubenswrapper[4813]: I1201 08:41:58.359337 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-2mb7d" event={"ID":"0f0340e8-faa5-447b-a5cd-a878e982d98f","Type":"ContainerStarted","Data":"81ace5696b7b1e97479c500c0dae3e9c2f5ad6341c22e14161695ed61e7ab950"} Dec 01 08:41:58 crc kubenswrapper[4813]: I1201 08:41:58.363111 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-7w2kz" event={"ID":"8b7d482d-6638-41f3-93b3-0b6aebf7401b","Type":"ContainerStarted","Data":"597e51be34648ce41aea7160576b25313cc2c7ce1e063b9c7ce327bec5efa822"} Dec 01 08:41:58 crc kubenswrapper[4813]: I1201 08:41:58.363294 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-7w2kz" event={"ID":"8b7d482d-6638-41f3-93b3-0b6aebf7401b","Type":"ContainerStarted","Data":"a55f38af42ae6b985523cf5a61a4f68672eb96c9508eda51c71f30658e1b9c87"} Dec 01 08:41:58 crc kubenswrapper[4813]: I1201 08:41:58.364197 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-7w2kz" Dec 01 08:41:58 crc kubenswrapper[4813]: I1201 08:41:58.365361 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mj64s\" (UID: \"a2b231fa-0ecc-44da-9259-a7dc7261127c\") " pod="openshift-image-registry/image-registry-697d97f7c8-mj64s" Dec 01 08:41:58 crc kubenswrapper[4813]: E1201 08:41:58.366938 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 08:41:58.866923109 +0000 UTC m=+103.489844695 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mj64s" (UID: "a2b231fa-0ecc-44da-9259-a7dc7261127c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:41:58 crc kubenswrapper[4813]: I1201 08:41:58.367774 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-hmh5v" event={"ID":"ab324c2b-a49d-4874-b9c0-a869b226ce98","Type":"ContainerStarted","Data":"74fa44d12e9f30a6bce6025e8c9580a70f8ec7f5d9999b4b16ae313f11cbb3a1"} Dec 01 08:41:58 crc kubenswrapper[4813]: I1201 08:41:58.367959 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-hmh5v" Dec 01 08:41:58 crc kubenswrapper[4813]: I1201 08:41:58.376434 4813 patch_prober.go:28] interesting pod/console-operator-58897d9998-7w2kz container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.23:8443/readyz\": dial tcp 10.217.0.23:8443: connect: connection refused" start-of-body= Dec 01 08:41:58 crc kubenswrapper[4813]: I1201 08:41:58.376484 4813 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-7w2kz" podUID="8b7d482d-6638-41f3-93b3-0b6aebf7401b" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.23:8443/readyz\": dial tcp 10.217.0.23:8443: connect: connection refused" Dec 01 08:41:58 crc kubenswrapper[4813]: I1201 08:41:58.403937 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-5ptsh" event={"ID":"d2aac47a-b117-42c7-afa3-2c989d9536a5","Type":"ContainerStarted","Data":"6132879ee7144ebbe8832ce300f47946cc00dd95702fd0f91b5b1a8559130988"} Dec 01 08:41:58 crc kubenswrapper[4813]: I1201 08:41:58.404011 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-5ptsh" event={"ID":"d2aac47a-b117-42c7-afa3-2c989d9536a5","Type":"ContainerStarted","Data":"75c7c446ed5bd43cafe5120c814534f095e97866bc3b7e386491bd028c87f55f"} Dec 01 08:41:58 crc kubenswrapper[4813]: I1201 08:41:58.415654 4813 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-5fgn9 container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.16:6443/healthz\": dial tcp 10.217.0.16:6443: connect: connection refused" start-of-body= Dec 01 08:41:58 crc kubenswrapper[4813]: I1201 08:41:58.416008 4813 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-5fgn9" podUID="52120ed5-ff4f-4df5-8c60-3342df484be8" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.16:6443/healthz\": dial tcp 10.217.0.16:6443: connect: connection refused" Dec 01 08:41:58 crc kubenswrapper[4813]: I1201 08:41:58.420316 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-8274k" podStartSLOduration=78.420295848 podStartE2EDuration="1m18.420295848s" podCreationTimestamp="2025-12-01 08:40:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:41:58.376158568 +0000 UTC m=+102.999080154" watchObservedRunningTime="2025-12-01 08:41:58.420295848 +0000 UTC m=+103.043217444" Dec 01 08:41:58 crc kubenswrapper[4813]: I1201 08:41:58.422225 4813 patch_prober.go:28] interesting pod/downloads-7954f5f757-ct7w8 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" start-of-body= Dec 01 08:41:58 crc kubenswrapper[4813]: I1201 08:41:58.428590 4813 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-ct7w8" podUID="e17c11d6-eb41-4391-b513-a817eeffcdb4" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" Dec 01 08:41:58 crc kubenswrapper[4813]: I1201 08:41:58.433003 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-7w2kz" podStartSLOduration=78.432979562 podStartE2EDuration="1m18.432979562s" podCreationTimestamp="2025-12-01 08:40:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:41:58.419022878 +0000 UTC m=+103.041944464" watchObservedRunningTime="2025-12-01 08:41:58.432979562 +0000 UTC m=+103.055901148" Dec 01 08:41:58 crc kubenswrapper[4813]: I1201 08:41:58.435240 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-5fgn9" event={"ID":"52120ed5-ff4f-4df5-8c60-3342df484be8","Type":"ContainerStarted","Data":"fad7917420effa27fbb37fc9bc2b08f930b871c20e49fe40a368976ec808175f"} Dec 01 08:41:58 crc kubenswrapper[4813]: I1201 08:41:58.435293 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-5fgn9" Dec 01 08:41:58 crc kubenswrapper[4813]: I1201 08:41:58.435304 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-5fgn9" event={"ID":"52120ed5-ff4f-4df5-8c60-3342df484be8","Type":"ContainerStarted","Data":"8cf41b039137a394ada809bf9390a7178c63a69753f291865d30047a66675172"} Dec 01 08:41:58 crc kubenswrapper[4813]: I1201 08:41:58.435314 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-sp4wh" event={"ID":"a905919c-6c7e-40be-8f43-b938a60f14f6","Type":"ContainerStarted","Data":"6800da3736d0c22dc48364c19e0635b1f1f85bba3c0037425bd912045d8e97e2"} Dec 01 08:41:58 crc kubenswrapper[4813]: I1201 08:41:58.435324 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lt4dx" event={"ID":"7f38918f-c551-4e3f-bb5e-d5dcac6ac826","Type":"ContainerStarted","Data":"de054dbd0812bd268003a34b4e993c4e2972b15ccc16ba117e85b70e514c1f54"} Dec 01 08:41:58 crc kubenswrapper[4813]: I1201 08:41:58.435530 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-jb992"] Dec 01 08:41:58 crc kubenswrapper[4813]: I1201 08:41:58.466269 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 08:41:58 crc kubenswrapper[4813]: E1201 08:41:58.466563 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 08:41:58.966536938 +0000 UTC m=+103.589458524 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:41:58 crc kubenswrapper[4813]: I1201 08:41:58.467418 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mj64s\" (UID: \"a2b231fa-0ecc-44da-9259-a7dc7261127c\") " pod="openshift-image-registry/image-registry-697d97f7c8-mj64s" Dec 01 08:41:58 crc kubenswrapper[4813]: E1201 08:41:58.468477 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 08:41:58.96845998 +0000 UTC m=+103.591381566 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mj64s" (UID: "a2b231fa-0ecc-44da-9259-a7dc7261127c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:41:58 crc kubenswrapper[4813]: I1201 08:41:58.515032 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-hmh5v" podStartSLOduration=78.515010075 podStartE2EDuration="1m18.515010075s" podCreationTimestamp="2025-12-01 08:40:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:41:58.513582027 +0000 UTC m=+103.136503613" watchObservedRunningTime="2025-12-01 08:41:58.515010075 +0000 UTC m=+103.137931661" Dec 01 08:41:58 crc kubenswrapper[4813]: I1201 08:41:58.549943 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-m8jlq"] Dec 01 08:41:58 crc kubenswrapper[4813]: I1201 08:41:58.553163 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-sjjm7"] Dec 01 08:41:58 crc kubenswrapper[4813]: I1201 08:41:58.568704 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 08:41:58 crc kubenswrapper[4813]: E1201 08:41:58.569081 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 08:41:59.069066056 +0000 UTC m=+103.691987642 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:41:58 crc kubenswrapper[4813]: I1201 08:41:58.586209 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-5fgn9" podStartSLOduration=78.586191771 podStartE2EDuration="1m18.586191771s" podCreationTimestamp="2025-12-01 08:40:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:41:58.583538465 +0000 UTC m=+103.206460051" watchObservedRunningTime="2025-12-01 08:41:58.586191771 +0000 UTC m=+103.209113347" Dec 01 08:41:58 crc kubenswrapper[4813]: I1201 08:41:58.672829 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mj64s\" (UID: \"a2b231fa-0ecc-44da-9259-a7dc7261127c\") " pod="openshift-image-registry/image-registry-697d97f7c8-mj64s" Dec 01 08:41:58 crc kubenswrapper[4813]: E1201 08:41:58.673209 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 08:41:59.17319638 +0000 UTC m=+103.796117966 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mj64s" (UID: "a2b231fa-0ecc-44da-9259-a7dc7261127c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:41:58 crc kubenswrapper[4813]: I1201 08:41:58.688723 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-sp4wh" podStartSLOduration=78.688701308 podStartE2EDuration="1m18.688701308s" podCreationTimestamp="2025-12-01 08:40:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:41:58.632388839 +0000 UTC m=+103.255310435" watchObservedRunningTime="2025-12-01 08:41:58.688701308 +0000 UTC m=+103.311622884" Dec 01 08:41:58 crc kubenswrapper[4813]: I1201 08:41:58.690075 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-5ptsh" podStartSLOduration=78.690067443 podStartE2EDuration="1m18.690067443s" podCreationTimestamp="2025-12-01 08:40:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:41:58.688391473 +0000 UTC m=+103.311313059" watchObservedRunningTime="2025-12-01 08:41:58.690067443 +0000 UTC m=+103.312989029" Dec 01 08:41:58 crc kubenswrapper[4813]: I1201 08:41:58.691391 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-6sclt"] Dec 01 08:41:58 crc kubenswrapper[4813]: I1201 08:41:58.696596 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-49bkd"] Dec 01 08:41:58 crc kubenswrapper[4813]: I1201 08:41:58.707101 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-vcwrd"] Dec 01 08:41:58 crc kubenswrapper[4813]: I1201 08:41:58.724173 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-qsskc"] Dec 01 08:41:58 crc kubenswrapper[4813]: I1201 08:41:58.734522 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-wqkxc" podStartSLOduration=78.734501927 podStartE2EDuration="1m18.734501927s" podCreationTimestamp="2025-12-01 08:40:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:41:58.724025878 +0000 UTC m=+103.346947464" watchObservedRunningTime="2025-12-01 08:41:58.734501927 +0000 UTC m=+103.357423513" Dec 01 08:41:58 crc kubenswrapper[4813]: I1201 08:41:58.734804 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-hf8fp"] Dec 01 08:41:58 crc kubenswrapper[4813]: W1201 08:41:58.748280 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1e597d43_ddb6_4cdb_8ce7_94b7427fb2f0.slice/crio-93b3e60bee5186b0ab4991f22ce1c4f6e483709752793577966215db1e1f3504 WatchSource:0}: Error finding container 93b3e60bee5186b0ab4991f22ce1c4f6e483709752793577966215db1e1f3504: Status 404 returned error can't find the container with id 93b3e60bee5186b0ab4991f22ce1c4f6e483709752793577966215db1e1f3504 Dec 01 08:41:58 crc kubenswrapper[4813]: I1201 08:41:58.774008 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lt4dx" podStartSLOduration=78.773985765 podStartE2EDuration="1m18.773985765s" podCreationTimestamp="2025-12-01 08:40:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:41:58.768950965 +0000 UTC m=+103.391872541" watchObservedRunningTime="2025-12-01 08:41:58.773985765 +0000 UTC m=+103.396907351" Dec 01 08:41:58 crc kubenswrapper[4813]: I1201 08:41:58.778038 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-f8kzc"] Dec 01 08:41:58 crc kubenswrapper[4813]: I1201 08:41:58.778921 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 08:41:58 crc kubenswrapper[4813]: E1201 08:41:58.779422 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 08:41:59.279401633 +0000 UTC m=+103.902323219 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:41:58 crc kubenswrapper[4813]: W1201 08:41:58.795310 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod525b7be7_bc47_4ff1_ada6_0521aa1c8620.slice/crio-e7406db1e0f016238e0ae7d786269739ab199da0767200fb45f61599fc4fbfe6 WatchSource:0}: Error finding container e7406db1e0f016238e0ae7d786269739ab199da0767200fb45f61599fc4fbfe6: Status 404 returned error can't find the container with id e7406db1e0f016238e0ae7d786269739ab199da0767200fb45f61599fc4fbfe6 Dec 01 08:41:58 crc kubenswrapper[4813]: I1201 08:41:58.815784 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-l5xp9"] Dec 01 08:41:58 crc kubenswrapper[4813]: I1201 08:41:58.817552 4813 patch_prober.go:28] interesting pod/router-default-5444994796-sp4wh container/router namespace/openshift-ingress: Startup probe status=failure output="Get \"http://localhost:1936/healthz/ready\": dial tcp [::1]:1936: connect: connection refused" start-of-body= Dec 01 08:41:58 crc kubenswrapper[4813]: I1201 08:41:58.817590 4813 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-sp4wh" podUID="a905919c-6c7e-40be-8f43-b938a60f14f6" containerName="router" probeResult="failure" output="Get \"http://localhost:1936/healthz/ready\": dial tcp [::1]:1936: connect: connection refused" Dec 01 08:41:58 crc kubenswrapper[4813]: I1201 08:41:58.828575 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-sp4wh" Dec 01 08:41:58 crc kubenswrapper[4813]: I1201 08:41:58.828616 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-n4dp7"] Dec 01 08:41:58 crc kubenswrapper[4813]: I1201 08:41:58.843796 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29409630-45pn5"] Dec 01 08:41:58 crc kubenswrapper[4813]: I1201 08:41:58.854253 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-dh427"] Dec 01 08:41:58 crc kubenswrapper[4813]: I1201 08:41:58.861526 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-dwkm7"] Dec 01 08:41:58 crc kubenswrapper[4813]: I1201 08:41:58.868683 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-j7v65"] Dec 01 08:41:58 crc kubenswrapper[4813]: I1201 08:41:58.876946 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-fxgbc"] Dec 01 08:41:58 crc kubenswrapper[4813]: I1201 08:41:58.879597 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-lfm7c"] Dec 01 08:41:58 crc kubenswrapper[4813]: I1201 08:41:58.880317 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mj64s\" (UID: \"a2b231fa-0ecc-44da-9259-a7dc7261127c\") " pod="openshift-image-registry/image-registry-697d97f7c8-mj64s" Dec 01 08:41:58 crc kubenswrapper[4813]: E1201 08:41:58.880673 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 08:41:59.380658279 +0000 UTC m=+104.003579865 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mj64s" (UID: "a2b231fa-0ecc-44da-9259-a7dc7261127c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:41:58 crc kubenswrapper[4813]: I1201 08:41:58.916920 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-5z87f"] Dec 01 08:41:58 crc kubenswrapper[4813]: W1201 08:41:58.933327 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5aa447ba_69b3_4291_9db5_25f4aeee2370.slice/crio-2a7f424fe3549a63e5f67f2b2b5e6f2a009e63a197fefedc2d4517df620f3074 WatchSource:0}: Error finding container 2a7f424fe3549a63e5f67f2b2b5e6f2a009e63a197fefedc2d4517df620f3074: Status 404 returned error can't find the container with id 2a7f424fe3549a63e5f67f2b2b5e6f2a009e63a197fefedc2d4517df620f3074 Dec 01 08:41:58 crc kubenswrapper[4813]: I1201 08:41:58.982557 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 08:41:58 crc kubenswrapper[4813]: E1201 08:41:58.982993 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 08:41:59.482935645 +0000 UTC m=+104.105857231 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:41:58 crc kubenswrapper[4813]: I1201 08:41:58.983685 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mj64s\" (UID: \"a2b231fa-0ecc-44da-9259-a7dc7261127c\") " pod="openshift-image-registry/image-registry-697d97f7c8-mj64s" Dec 01 08:41:58 crc kubenswrapper[4813]: E1201 08:41:58.984235 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 08:41:59.484226746 +0000 UTC m=+104.107148332 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mj64s" (UID: "a2b231fa-0ecc-44da-9259-a7dc7261127c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:41:59 crc kubenswrapper[4813]: I1201 08:41:59.096544 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 08:41:59 crc kubenswrapper[4813]: E1201 08:41:59.097096 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 08:41:59.597079595 +0000 UTC m=+104.220001181 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:41:59 crc kubenswrapper[4813]: I1201 08:41:59.097767 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-5ptsh" Dec 01 08:41:59 crc kubenswrapper[4813]: I1201 08:41:59.098184 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-5ptsh" Dec 01 08:41:59 crc kubenswrapper[4813]: I1201 08:41:59.165784 4813 patch_prober.go:28] interesting pod/apiserver-76f77b778f-5ptsh container/openshift-apiserver namespace/openshift-apiserver: Startup probe status=failure output="Get \"https://10.217.0.11:8443/livez\": dial tcp 10.217.0.11:8443: connect: connection refused" start-of-body= Dec 01 08:41:59 crc kubenswrapper[4813]: I1201 08:41:59.165877 4813 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-apiserver/apiserver-76f77b778f-5ptsh" podUID="d2aac47a-b117-42c7-afa3-2c989d9536a5" containerName="openshift-apiserver" probeResult="failure" output="Get \"https://10.217.0.11:8443/livez\": dial tcp 10.217.0.11:8443: connect: connection refused" Dec 01 08:41:59 crc kubenswrapper[4813]: I1201 08:41:59.198764 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mj64s\" (UID: \"a2b231fa-0ecc-44da-9259-a7dc7261127c\") " pod="openshift-image-registry/image-registry-697d97f7c8-mj64s" Dec 01 08:41:59 crc kubenswrapper[4813]: E1201 08:41:59.200379 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 08:41:59.700362089 +0000 UTC m=+104.323283675 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mj64s" (UID: "a2b231fa-0ecc-44da-9259-a7dc7261127c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:41:59 crc kubenswrapper[4813]: I1201 08:41:59.302011 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 08:41:59 crc kubenswrapper[4813]: E1201 08:41:59.302397 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 08:41:59.802376912 +0000 UTC m=+104.425298498 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:41:59 crc kubenswrapper[4813]: I1201 08:41:59.370072 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lt4dx" Dec 01 08:41:59 crc kubenswrapper[4813]: I1201 08:41:59.370424 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lt4dx" Dec 01 08:41:59 crc kubenswrapper[4813]: I1201 08:41:59.398463 4813 patch_prober.go:28] interesting pod/apiserver-7bbb656c7d-lt4dx container/oauth-apiserver namespace/openshift-oauth-apiserver: Startup probe status=failure output="Get \"https://10.217.0.9:8443/livez\": dial tcp 10.217.0.9:8443: connect: connection refused" start-of-body= Dec 01 08:41:59 crc kubenswrapper[4813]: I1201 08:41:59.398527 4813 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lt4dx" podUID="7f38918f-c551-4e3f-bb5e-d5dcac6ac826" containerName="oauth-apiserver" probeResult="failure" output="Get \"https://10.217.0.9:8443/livez\": dial tcp 10.217.0.9:8443: connect: connection refused" Dec 01 08:41:59 crc kubenswrapper[4813]: I1201 08:41:59.403343 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mj64s\" (UID: \"a2b231fa-0ecc-44da-9259-a7dc7261127c\") " pod="openshift-image-registry/image-registry-697d97f7c8-mj64s" Dec 01 08:41:59 crc kubenswrapper[4813]: E1201 08:41:59.403762 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 08:41:59.903745215 +0000 UTC m=+104.526666801 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mj64s" (UID: "a2b231fa-0ecc-44da-9259-a7dc7261127c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:41:59 crc kubenswrapper[4813]: I1201 08:41:59.467860 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-5z87f" event={"ID":"a7588f88-9a71-4e13-9fc3-a34cbfa6a090","Type":"ContainerStarted","Data":"41dd835dbbb56a2836ccaee717f805eadfdab83056970e0df05bda4058058ef7"} Dec 01 08:41:59 crc kubenswrapper[4813]: I1201 08:41:59.501903 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-n4dp7" event={"ID":"5aa447ba-69b3-4291-9db5-25f4aeee2370","Type":"ContainerStarted","Data":"2a7f424fe3549a63e5f67f2b2b5e6f2a009e63a197fefedc2d4517df620f3074"} Dec 01 08:41:59 crc kubenswrapper[4813]: I1201 08:41:59.504136 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 08:41:59 crc kubenswrapper[4813]: E1201 08:41:59.504752 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 08:42:00.004706628 +0000 UTC m=+104.627628244 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:41:59 crc kubenswrapper[4813]: I1201 08:41:59.532680 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-lfm7c" event={"ID":"469840fe-2ac3-4431-9b2b-3b41451e6cbc","Type":"ContainerStarted","Data":"57ca8bd0df3836842c4d35c59d860dc97246bfe169a0ca3fdbcc38e7c564dfe9"} Dec 01 08:41:59 crc kubenswrapper[4813]: I1201 08:41:59.554142 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-6sclt" event={"ID":"525b7be7-bc47-4ff1-ada6-0521aa1c8620","Type":"ContainerStarted","Data":"e7406db1e0f016238e0ae7d786269739ab199da0767200fb45f61599fc4fbfe6"} Dec 01 08:41:59 crc kubenswrapper[4813]: I1201 08:41:59.556153 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-dh427" event={"ID":"57a1918e-f28a-4890-b88c-bc80df5cf704","Type":"ContainerStarted","Data":"07f0e6adf7d98728f95215b0e7278b2a758c3874363d83fa1eddf907aed08224"} Dec 01 08:41:59 crc kubenswrapper[4813]: I1201 08:41:59.558416 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-vcwrd" event={"ID":"1e597d43-ddb6-4cdb-8ce7-94b7427fb2f0","Type":"ContainerStarted","Data":"93b3e60bee5186b0ab4991f22ce1c4f6e483709752793577966215db1e1f3504"} Dec 01 08:41:59 crc kubenswrapper[4813]: I1201 08:41:59.565647 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-f8kzc" event={"ID":"2e3f10d2-ac94-459c-9898-542e9f710747","Type":"ContainerStarted","Data":"9dee7bd6c68b831cef0b97c8eb984c53c1f6631906fcb4dfa58addf30a812034"} Dec 01 08:41:59 crc kubenswrapper[4813]: I1201 08:41:59.568164 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-m8jlq" event={"ID":"363f808f-a265-464a-a3a7-4128e3734b14","Type":"ContainerStarted","Data":"19aff26e84be8f30d1250979ddaeb5d49a3c9d9602f5edeb8e997956522e5cf2"} Dec 01 08:41:59 crc kubenswrapper[4813]: I1201 08:41:59.569609 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29409630-45pn5" event={"ID":"95e0dc94-8aec-42d0-8b19-7cdf12fa7dee","Type":"ContainerStarted","Data":"97c05336981e5cd1c1cddcc9c2fcc2d3a3c658e8a05351e240d02255ceaa7081"} Dec 01 08:41:59 crc kubenswrapper[4813]: I1201 08:41:59.590553 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-jb992" event={"ID":"7d654f9a-b190-47bc-89ce-0caae2ec1513","Type":"ContainerStarted","Data":"6fe09de3f009c52c6bec1ba4ac767a29b142cbf85c78a28b72aed576e3d553d6"} Dec 01 08:41:59 crc kubenswrapper[4813]: I1201 08:41:59.590604 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-jb992" event={"ID":"7d654f9a-b190-47bc-89ce-0caae2ec1513","Type":"ContainerStarted","Data":"fbf3705e1d2f9560c84e985c9ebb31aa32faf34075b46fc33908148c396eeef6"} Dec 01 08:41:59 crc kubenswrapper[4813]: I1201 08:41:59.591677 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-dwkm7" event={"ID":"f3561d3e-0e57-4243-83df-9a475da69fb6","Type":"ContainerStarted","Data":"29db19f382a961d0faabc90338091f8b2266e5f44ff3b4150baaa1a5fe028078"} Dec 01 08:41:59 crc kubenswrapper[4813]: I1201 08:41:59.592502 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-l5xp9" event={"ID":"7caa1280-3fcc-47b0-ab38-8aa5479fcfce","Type":"ContainerStarted","Data":"b5ea8d45ed51f7480efb521de8ad79d988630669fdf7079c93c527583758e173"} Dec 01 08:41:59 crc kubenswrapper[4813]: I1201 08:41:59.679584 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mj64s\" (UID: \"a2b231fa-0ecc-44da-9259-a7dc7261127c\") " pod="openshift-image-registry/image-registry-697d97f7c8-mj64s" Dec 01 08:41:59 crc kubenswrapper[4813]: E1201 08:41:59.679885 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 08:42:00.179871541 +0000 UTC m=+104.802793127 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mj64s" (UID: "a2b231fa-0ecc-44da-9259-a7dc7261127c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:41:59 crc kubenswrapper[4813]: I1201 08:41:59.687170 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-49bkd" event={"ID":"adf5f998-8117-4362-868e-b7cdbbf91e0e","Type":"ContainerStarted","Data":"2a622f1764d5533c4e8d9e2161b0c0a0c97363a2418f610cffe384d28f979231"} Dec 01 08:41:59 crc kubenswrapper[4813]: I1201 08:41:59.702530 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-2mb7d" event={"ID":"0f0340e8-faa5-447b-a5cd-a878e982d98f","Type":"ContainerStarted","Data":"a886c4c7bb5d2c717fea11a48e826ddbcdd7e2724cea3114371eee53b1ed91d1"} Dec 01 08:41:59 crc kubenswrapper[4813]: I1201 08:41:59.704162 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-2mb7d" Dec 01 08:41:59 crc kubenswrapper[4813]: I1201 08:41:59.712780 4813 patch_prober.go:28] interesting pod/route-controller-manager-6576b87f9c-2mb7d container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.19:8443/healthz\": dial tcp 10.217.0.19:8443: connect: connection refused" start-of-body= Dec 01 08:41:59 crc kubenswrapper[4813]: I1201 08:41:59.712847 4813 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-2mb7d" podUID="0f0340e8-faa5-447b-a5cd-a878e982d98f" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.19:8443/healthz\": dial tcp 10.217.0.19:8443: connect: connection refused" Dec 01 08:41:59 crc kubenswrapper[4813]: I1201 08:41:59.724298 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-j7v65" event={"ID":"61054467-59d0-40a9-a60c-09c6c6106acd","Type":"ContainerStarted","Data":"b913eecd10d42329c282b3ec044d50d40511501f8934a73060eda0e0645e3319"} Dec 01 08:41:59 crc kubenswrapper[4813]: I1201 08:41:59.742375 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-sjjm7" event={"ID":"dadb7bca-672d-434b-897d-7eb83059f189","Type":"ContainerStarted","Data":"da2e2f521a38ba87470c2588bcb0e94bcf4970c6ac7e30bb7e671485d7903a27"} Dec 01 08:41:59 crc kubenswrapper[4813]: I1201 08:41:59.742424 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-sjjm7" event={"ID":"dadb7bca-672d-434b-897d-7eb83059f189","Type":"ContainerStarted","Data":"05043f0f2e6b5fd9e1c11416f10c62cb87a15ad5a131982fa43e25fba4fbd5f3"} Dec 01 08:41:59 crc kubenswrapper[4813]: I1201 08:41:59.767619 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-hf8fp" event={"ID":"489e9f98-0728-4a72-80d4-3d88a77823f5","Type":"ContainerStarted","Data":"578e37160648db54f251afa89ef15964abb8d777d214d004efc5754168ed17a5"} Dec 01 08:41:59 crc kubenswrapper[4813]: I1201 08:41:59.767683 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-hf8fp" event={"ID":"489e9f98-0728-4a72-80d4-3d88a77823f5","Type":"ContainerStarted","Data":"36799f2d2bf9aa30301341ee2ac5a50b53d2b90c331b96eba4a86c0b7593659b"} Dec 01 08:41:59 crc kubenswrapper[4813]: I1201 08:41:59.768144 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-2mb7d" podStartSLOduration=79.76812198 podStartE2EDuration="1m19.76812198s" podCreationTimestamp="2025-12-01 08:40:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:41:59.752167471 +0000 UTC m=+104.375089057" watchObservedRunningTime="2025-12-01 08:41:59.76812198 +0000 UTC m=+104.391043556" Dec 01 08:41:59 crc kubenswrapper[4813]: I1201 08:41:59.768888 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-hf8fp" Dec 01 08:41:59 crc kubenswrapper[4813]: I1201 08:41:59.778662 4813 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-hf8fp container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.29:8443/healthz\": dial tcp 10.217.0.29:8443: connect: connection refused" start-of-body= Dec 01 08:41:59 crc kubenswrapper[4813]: I1201 08:41:59.778729 4813 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-hf8fp" podUID="489e9f98-0728-4a72-80d4-3d88a77823f5" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.29:8443/healthz\": dial tcp 10.217.0.29:8443: connect: connection refused" Dec 01 08:41:59 crc kubenswrapper[4813]: I1201 08:41:59.782427 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 08:41:59 crc kubenswrapper[4813]: E1201 08:41:59.783870 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 08:42:00.283850158 +0000 UTC m=+104.906771744 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:41:59 crc kubenswrapper[4813]: I1201 08:41:59.811261 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-qsskc" event={"ID":"c342090f-f331-484d-b287-929d8303f5d7","Type":"ContainerStarted","Data":"9bf28095a6c6ccd8c16c480dd35263958c6ff846cbc75ff9300b59da85db1125"} Dec 01 08:41:59 crc kubenswrapper[4813]: I1201 08:41:59.820635 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-hf8fp" podStartSLOduration=79.820614837 podStartE2EDuration="1m19.820614837s" podCreationTimestamp="2025-12-01 08:40:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:41:59.812276331 +0000 UTC m=+104.435197937" watchObservedRunningTime="2025-12-01 08:41:59.820614837 +0000 UTC m=+104.443536423" Dec 01 08:41:59 crc kubenswrapper[4813]: I1201 08:41:59.823706 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-fxgbc" event={"ID":"2fd313fd-c0b1-4aae-8b3b-2d078a7fc7d0","Type":"ContainerStarted","Data":"45429916d0b4eb4b91d49eb9e3c24839efc208b79cbe4bc9eb363e2200817cfd"} Dec 01 08:41:59 crc kubenswrapper[4813]: I1201 08:41:59.824129 4813 patch_prober.go:28] interesting pod/console-operator-58897d9998-7w2kz container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.23:8443/readyz\": dial tcp 10.217.0.23:8443: connect: connection refused" start-of-body= Dec 01 08:41:59 crc kubenswrapper[4813]: I1201 08:41:59.824198 4813 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-7w2kz" podUID="8b7d482d-6638-41f3-93b3-0b6aebf7401b" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.23:8443/readyz\": dial tcp 10.217.0.23:8443: connect: connection refused" Dec 01 08:41:59 crc kubenswrapper[4813]: I1201 08:41:59.830154 4813 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-5fgn9 container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.16:6443/healthz\": dial tcp 10.217.0.16:6443: connect: connection refused" start-of-body= Dec 01 08:41:59 crc kubenswrapper[4813]: I1201 08:41:59.830238 4813 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-5fgn9" podUID="52120ed5-ff4f-4df5-8c60-3342df484be8" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.16:6443/healthz\": dial tcp 10.217.0.16:6443: connect: connection refused" Dec 01 08:41:59 crc kubenswrapper[4813]: I1201 08:41:59.835178 4813 patch_prober.go:28] interesting pod/router-default-5444994796-sp4wh container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 01 08:41:59 crc kubenswrapper[4813]: [-]has-synced failed: reason withheld Dec 01 08:41:59 crc kubenswrapper[4813]: [+]process-running ok Dec 01 08:41:59 crc kubenswrapper[4813]: healthz check failed Dec 01 08:41:59 crc kubenswrapper[4813]: I1201 08:41:59.835226 4813 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-sp4wh" podUID="a905919c-6c7e-40be-8f43-b938a60f14f6" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 01 08:41:59 crc kubenswrapper[4813]: I1201 08:41:59.886664 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mj64s\" (UID: \"a2b231fa-0ecc-44da-9259-a7dc7261127c\") " pod="openshift-image-registry/image-registry-697d97f7c8-mj64s" Dec 01 08:41:59 crc kubenswrapper[4813]: I1201 08:41:59.886766 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0789e3f7-5b08-4c04-8596-9ba29f50cdd0-metrics-certs\") pod \"network-metrics-daemon-96mbr\" (UID: \"0789e3f7-5b08-4c04-8596-9ba29f50cdd0\") " pod="openshift-multus/network-metrics-daemon-96mbr" Dec 01 08:41:59 crc kubenswrapper[4813]: E1201 08:41:59.889774 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 08:42:00.389754497 +0000 UTC m=+105.012676183 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mj64s" (UID: "a2b231fa-0ecc-44da-9259-a7dc7261127c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:41:59 crc kubenswrapper[4813]: I1201 08:41:59.950340 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0789e3f7-5b08-4c04-8596-9ba29f50cdd0-metrics-certs\") pod \"network-metrics-daemon-96mbr\" (UID: \"0789e3f7-5b08-4c04-8596-9ba29f50cdd0\") " pod="openshift-multus/network-metrics-daemon-96mbr" Dec 01 08:41:59 crc kubenswrapper[4813]: I1201 08:41:59.993935 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 08:41:59 crc kubenswrapper[4813]: E1201 08:41:59.994662 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 08:42:00.494643437 +0000 UTC m=+105.117565023 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:41:59 crc kubenswrapper[4813]: I1201 08:41:59.994769 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-96mbr" Dec 01 08:42:00 crc kubenswrapper[4813]: I1201 08:42:00.097859 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mj64s\" (UID: \"a2b231fa-0ecc-44da-9259-a7dc7261127c\") " pod="openshift-image-registry/image-registry-697d97f7c8-mj64s" Dec 01 08:42:00 crc kubenswrapper[4813]: E1201 08:42:00.098671 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 08:42:00.598653595 +0000 UTC m=+105.221575181 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mj64s" (UID: "a2b231fa-0ecc-44da-9259-a7dc7261127c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:42:00 crc kubenswrapper[4813]: I1201 08:42:00.208623 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 08:42:00 crc kubenswrapper[4813]: E1201 08:42:00.208859 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 08:42:00.708844537 +0000 UTC m=+105.331766123 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:42:00 crc kubenswrapper[4813]: I1201 08:42:00.314053 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mj64s\" (UID: \"a2b231fa-0ecc-44da-9259-a7dc7261127c\") " pod="openshift-image-registry/image-registry-697d97f7c8-mj64s" Dec 01 08:42:00 crc kubenswrapper[4813]: E1201 08:42:00.314548 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 08:42:00.814530485 +0000 UTC m=+105.437452071 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mj64s" (UID: "a2b231fa-0ecc-44da-9259-a7dc7261127c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:42:00 crc kubenswrapper[4813]: I1201 08:42:00.416178 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 08:42:00 crc kubenswrapper[4813]: E1201 08:42:00.416530 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 08:42:00.916500207 +0000 UTC m=+105.539421803 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:42:00 crc kubenswrapper[4813]: I1201 08:42:00.417624 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mj64s\" (UID: \"a2b231fa-0ecc-44da-9259-a7dc7261127c\") " pod="openshift-image-registry/image-registry-697d97f7c8-mj64s" Dec 01 08:42:00 crc kubenswrapper[4813]: E1201 08:42:00.419893 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 08:42:00.919873257 +0000 UTC m=+105.542794843 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mj64s" (UID: "a2b231fa-0ecc-44da-9259-a7dc7261127c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:42:00 crc kubenswrapper[4813]: I1201 08:42:00.527703 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 08:42:00 crc kubenswrapper[4813]: E1201 08:42:00.528216 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 08:42:01.028196641 +0000 UTC m=+105.651118227 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:42:00 crc kubenswrapper[4813]: I1201 08:42:00.629415 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mj64s\" (UID: \"a2b231fa-0ecc-44da-9259-a7dc7261127c\") " pod="openshift-image-registry/image-registry-697d97f7c8-mj64s" Dec 01 08:42:00 crc kubenswrapper[4813]: E1201 08:42:00.630769 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 08:42:01.13075672 +0000 UTC m=+105.753678296 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mj64s" (UID: "a2b231fa-0ecc-44da-9259-a7dc7261127c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:42:00 crc kubenswrapper[4813]: I1201 08:42:00.731383 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 08:42:00 crc kubenswrapper[4813]: E1201 08:42:00.732057 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 08:42:01.232038208 +0000 UTC m=+105.854959794 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:42:00 crc kubenswrapper[4813]: I1201 08:42:00.818051 4813 patch_prober.go:28] interesting pod/router-default-5444994796-sp4wh container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 01 08:42:00 crc kubenswrapper[4813]: [-]has-synced failed: reason withheld Dec 01 08:42:00 crc kubenswrapper[4813]: [+]process-running ok Dec 01 08:42:00 crc kubenswrapper[4813]: healthz check failed Dec 01 08:42:00 crc kubenswrapper[4813]: I1201 08:42:00.818108 4813 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-sp4wh" podUID="a905919c-6c7e-40be-8f43-b938a60f14f6" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 01 08:42:00 crc kubenswrapper[4813]: I1201 08:42:00.833470 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mj64s\" (UID: \"a2b231fa-0ecc-44da-9259-a7dc7261127c\") " pod="openshift-image-registry/image-registry-697d97f7c8-mj64s" Dec 01 08:42:00 crc kubenswrapper[4813]: E1201 08:42:00.833925 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 08:42:01.333913245 +0000 UTC m=+105.956834821 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mj64s" (UID: "a2b231fa-0ecc-44da-9259-a7dc7261127c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:42:00 crc kubenswrapper[4813]: I1201 08:42:00.934352 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 08:42:00 crc kubenswrapper[4813]: E1201 08:42:00.934710 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 08:42:01.43469276 +0000 UTC m=+106.057614346 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:42:00 crc kubenswrapper[4813]: I1201 08:42:00.955827 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-fxgbc" event={"ID":"2fd313fd-c0b1-4aae-8b3b-2d078a7fc7d0","Type":"ContainerStarted","Data":"b71f3d2028f6b4644af530e799ca4d2a19d5592d24ff991e91f7798069f6b757"} Dec 01 08:42:01 crc kubenswrapper[4813]: I1201 08:42:01.005831 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-dwkm7" event={"ID":"f3561d3e-0e57-4243-83df-9a475da69fb6","Type":"ContainerStarted","Data":"9ce74a90bf05909764f299262ac3c11d5c64ee00e6d69f527c70ebd18c1f9b76"} Dec 01 08:42:01 crc kubenswrapper[4813]: I1201 08:42:01.031276 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-dh427" event={"ID":"57a1918e-f28a-4890-b88c-bc80df5cf704","Type":"ContainerStarted","Data":"cb6b52576286a3089e8c729d4da49c39104373964f5c4a2eeb001d937af09e9d"} Dec 01 08:42:01 crc kubenswrapper[4813]: I1201 08:42:01.038637 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-fxgbc" podStartSLOduration=81.038620914 podStartE2EDuration="1m21.038620914s" podCreationTimestamp="2025-12-01 08:40:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:42:01.024021919 +0000 UTC m=+105.646943505" watchObservedRunningTime="2025-12-01 08:42:01.038620914 +0000 UTC m=+105.661542500" Dec 01 08:42:01 crc kubenswrapper[4813]: I1201 08:42:01.042986 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mj64s\" (UID: \"a2b231fa-0ecc-44da-9259-a7dc7261127c\") " pod="openshift-image-registry/image-registry-697d97f7c8-mj64s" Dec 01 08:42:01 crc kubenswrapper[4813]: E1201 08:42:01.044104 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 08:42:01.544093294 +0000 UTC m=+106.167014880 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mj64s" (UID: "a2b231fa-0ecc-44da-9259-a7dc7261127c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:42:01 crc kubenswrapper[4813]: I1201 08:42:01.102579 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-dh427" podStartSLOduration=81.102560436 podStartE2EDuration="1m21.102560436s" podCreationTimestamp="2025-12-01 08:40:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:42:01.081788328 +0000 UTC m=+105.704709914" watchObservedRunningTime="2025-12-01 08:42:01.102560436 +0000 UTC m=+105.725482022" Dec 01 08:42:01 crc kubenswrapper[4813]: I1201 08:42:01.105324 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-m8jlq" event={"ID":"363f808f-a265-464a-a3a7-4128e3734b14","Type":"ContainerStarted","Data":"e709d40e9e318e883e8c7584f6ce71a776bc2732df3b69e47561d04f8fd200f5"} Dec 01 08:42:01 crc kubenswrapper[4813]: I1201 08:42:01.132658 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-n4dp7" event={"ID":"5aa447ba-69b3-4291-9db5-25f4aeee2370","Type":"ContainerStarted","Data":"8a6476cbd3567b63340e5d358d640a2fc724fc8fccfdf27821d34ef983addf69"} Dec 01 08:42:01 crc kubenswrapper[4813]: I1201 08:42:01.133568 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-n4dp7" Dec 01 08:42:01 crc kubenswrapper[4813]: I1201 08:42:01.144788 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 08:42:01 crc kubenswrapper[4813]: E1201 08:42:01.146002 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 08:42:01.645980432 +0000 UTC m=+106.268902018 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:42:01 crc kubenswrapper[4813]: I1201 08:42:01.152116 4813 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-n4dp7 container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.41:8443/healthz\": dial tcp 10.217.0.41:8443: connect: connection refused" start-of-body= Dec 01 08:42:01 crc kubenswrapper[4813]: I1201 08:42:01.152168 4813 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-n4dp7" podUID="5aa447ba-69b3-4291-9db5-25f4aeee2370" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.41:8443/healthz\": dial tcp 10.217.0.41:8443: connect: connection refused" Dec 01 08:42:01 crc kubenswrapper[4813]: I1201 08:42:01.247267 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mj64s\" (UID: \"a2b231fa-0ecc-44da-9259-a7dc7261127c\") " pod="openshift-image-registry/image-registry-697d97f7c8-mj64s" Dec 01 08:42:01 crc kubenswrapper[4813]: E1201 08:42:01.248642 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 08:42:01.748629765 +0000 UTC m=+106.371551351 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mj64s" (UID: "a2b231fa-0ecc-44da-9259-a7dc7261127c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:42:01 crc kubenswrapper[4813]: I1201 08:42:01.248777 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-qsskc" event={"ID":"c342090f-f331-484d-b287-929d8303f5d7","Type":"ContainerStarted","Data":"675727fde89d8e51f7ed3ff2b83d39e2917fa51a8b9eb117e74b96316e396b4f"} Dec 01 08:42:01 crc kubenswrapper[4813]: I1201 08:42:01.249223 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-qsskc" Dec 01 08:42:01 crc kubenswrapper[4813]: I1201 08:42:01.291709 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-lfm7c" event={"ID":"469840fe-2ac3-4431-9b2b-3b41451e6cbc","Type":"ContainerStarted","Data":"9e27e759cf47b84e0e5ff7525432ba63c91362bc3d97d4dd14942ecb4fcd6b37"} Dec 01 08:42:01 crc kubenswrapper[4813]: I1201 08:42:01.329007 4813 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-qsskc container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.20:8080/healthz\": dial tcp 10.217.0.20:8080: connect: connection refused" start-of-body= Dec 01 08:42:01 crc kubenswrapper[4813]: I1201 08:42:01.329054 4813 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-qsskc" podUID="c342090f-f331-484d-b287-929d8303f5d7" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.20:8080/healthz\": dial tcp 10.217.0.20:8080: connect: connection refused" Dec 01 08:42:01 crc kubenswrapper[4813]: I1201 08:42:01.347932 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-sjjm7" podStartSLOduration=81.347915699 podStartE2EDuration="1m21.347915699s" podCreationTimestamp="2025-12-01 08:40:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:42:01.347027667 +0000 UTC m=+105.969949253" watchObservedRunningTime="2025-12-01 08:42:01.347915699 +0000 UTC m=+105.970837275" Dec 01 08:42:01 crc kubenswrapper[4813]: I1201 08:42:01.350176 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 08:42:01 crc kubenswrapper[4813]: E1201 08:42:01.350757 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 08:42:01.850735323 +0000 UTC m=+106.473656909 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:42:01 crc kubenswrapper[4813]: I1201 08:42:01.355227 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-vcwrd" event={"ID":"1e597d43-ddb6-4cdb-8ce7-94b7427fb2f0","Type":"ContainerStarted","Data":"65bfee29111b38bfe44439726c7b1ad736c4dca0bb115e9b66c76d79392ca33d"} Dec 01 08:42:01 crc kubenswrapper[4813]: I1201 08:42:01.375924 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-qsskc" podStartSLOduration=81.375908111 podStartE2EDuration="1m21.375908111s" podCreationTimestamp="2025-12-01 08:40:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:42:01.374275793 +0000 UTC m=+105.997197379" watchObservedRunningTime="2025-12-01 08:42:01.375908111 +0000 UTC m=+105.998829697" Dec 01 08:42:01 crc kubenswrapper[4813]: I1201 08:42:01.379490 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-f8kzc" Dec 01 08:42:01 crc kubenswrapper[4813]: I1201 08:42:01.383632 4813 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-f8kzc container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.25:5443/healthz\": dial tcp 10.217.0.25:5443: connect: connection refused" start-of-body= Dec 01 08:42:01 crc kubenswrapper[4813]: I1201 08:42:01.383661 4813 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-f8kzc" podUID="2e3f10d2-ac94-459c-9898-542e9f710747" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.25:5443/healthz\": dial tcp 10.217.0.25:5443: connect: connection refused" Dec 01 08:42:01 crc kubenswrapper[4813]: I1201 08:42:01.409617 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-n4dp7" podStartSLOduration=81.409601673 podStartE2EDuration="1m21.409601673s" podCreationTimestamp="2025-12-01 08:40:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:42:01.405358832 +0000 UTC m=+106.028280418" watchObservedRunningTime="2025-12-01 08:42:01.409601673 +0000 UTC m=+106.032523259" Dec 01 08:42:01 crc kubenswrapper[4813]: I1201 08:42:01.411752 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-96mbr"] Dec 01 08:42:01 crc kubenswrapper[4813]: I1201 08:42:01.421525 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-6sclt" event={"ID":"525b7be7-bc47-4ff1-ada6-0521aa1c8620","Type":"ContainerStarted","Data":"583ae997533a73a25b0e1982bd950558bfd0c0195a295d8b4f663a4d6b1f883b"} Dec 01 08:42:01 crc kubenswrapper[4813]: I1201 08:42:01.456461 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mj64s\" (UID: \"a2b231fa-0ecc-44da-9259-a7dc7261127c\") " pod="openshift-image-registry/image-registry-697d97f7c8-mj64s" Dec 01 08:42:01 crc kubenswrapper[4813]: E1201 08:42:01.458874 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 08:42:01.958861587 +0000 UTC m=+106.581783173 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mj64s" (UID: "a2b231fa-0ecc-44da-9259-a7dc7261127c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:42:01 crc kubenswrapper[4813]: I1201 08:42:01.524803 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29409630-45pn5" event={"ID":"95e0dc94-8aec-42d0-8b19-7cdf12fa7dee","Type":"ContainerStarted","Data":"f6d3eb682a8ac8a11cfdaff1cc058f21bb6b5cfef699c8f6fdf59a4c88c3f826"} Dec 01 08:42:01 crc kubenswrapper[4813]: I1201 08:42:01.527439 4813 patch_prober.go:28] interesting pod/console-operator-58897d9998-7w2kz container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.23:8443/readyz\": dial tcp 10.217.0.23:8443: connect: connection refused" start-of-body= Dec 01 08:42:01 crc kubenswrapper[4813]: I1201 08:42:01.527483 4813 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-7w2kz" podUID="8b7d482d-6638-41f3-93b3-0b6aebf7401b" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.23:8443/readyz\": dial tcp 10.217.0.23:8443: connect: connection refused" Dec 01 08:42:01 crc kubenswrapper[4813]: I1201 08:42:01.527547 4813 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-hf8fp container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.29:8443/healthz\": dial tcp 10.217.0.29:8443: connect: connection refused" start-of-body= Dec 01 08:42:01 crc kubenswrapper[4813]: I1201 08:42:01.527562 4813 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-hf8fp" podUID="489e9f98-0728-4a72-80d4-3d88a77823f5" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.29:8443/healthz\": dial tcp 10.217.0.29:8443: connect: connection refused" Dec 01 08:42:01 crc kubenswrapper[4813]: I1201 08:42:01.557898 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 08:42:01 crc kubenswrapper[4813]: E1201 08:42:01.559355 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 08:42:02.059316556 +0000 UTC m=+106.682238142 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:42:01 crc kubenswrapper[4813]: I1201 08:42:01.563024 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-lfm7c" podStartSLOduration=81.563002131 podStartE2EDuration="1m21.563002131s" podCreationTimestamp="2025-12-01 08:40:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:42:01.446156892 +0000 UTC m=+106.069078478" watchObservedRunningTime="2025-12-01 08:42:01.563002131 +0000 UTC m=+106.185923737" Dec 01 08:42:01 crc kubenswrapper[4813]: I1201 08:42:01.563440 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-f8kzc" podStartSLOduration=81.563430612 podStartE2EDuration="1m21.563430612s" podCreationTimestamp="2025-12-01 08:40:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:42:01.548347324 +0000 UTC m=+106.171268910" watchObservedRunningTime="2025-12-01 08:42:01.563430612 +0000 UTC m=+106.186352198" Dec 01 08:42:01 crc kubenswrapper[4813]: I1201 08:42:01.588737 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-vcwrd" podStartSLOduration=81.588714295 podStartE2EDuration="1m21.588714295s" podCreationTimestamp="2025-12-01 08:40:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:42:01.5884159 +0000 UTC m=+106.211337476" watchObservedRunningTime="2025-12-01 08:42:01.588714295 +0000 UTC m=+106.211635881" Dec 01 08:42:01 crc kubenswrapper[4813]: I1201 08:42:01.768253 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mj64s\" (UID: \"a2b231fa-0ecc-44da-9259-a7dc7261127c\") " pod="openshift-image-registry/image-registry-697d97f7c8-mj64s" Dec 01 08:42:01 crc kubenswrapper[4813]: E1201 08:42:01.789079 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 08:42:02.289034975 +0000 UTC m=+106.911956561 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mj64s" (UID: "a2b231fa-0ecc-44da-9259-a7dc7261127c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:42:01 crc kubenswrapper[4813]: I1201 08:42:01.796154 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-6sclt" podStartSLOduration=81.796119832 podStartE2EDuration="1m21.796119832s" podCreationTimestamp="2025-12-01 08:40:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:42:01.636284928 +0000 UTC m=+106.259206514" watchObservedRunningTime="2025-12-01 08:42:01.796119832 +0000 UTC m=+106.419041418" Dec 01 08:42:01 crc kubenswrapper[4813]: I1201 08:42:01.822104 4813 patch_prober.go:28] interesting pod/router-default-5444994796-sp4wh container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 01 08:42:01 crc kubenswrapper[4813]: [-]has-synced failed: reason withheld Dec 01 08:42:01 crc kubenswrapper[4813]: [+]process-running ok Dec 01 08:42:01 crc kubenswrapper[4813]: healthz check failed Dec 01 08:42:01 crc kubenswrapper[4813]: I1201 08:42:01.822197 4813 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-sp4wh" podUID="a905919c-6c7e-40be-8f43-b938a60f14f6" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 01 08:42:01 crc kubenswrapper[4813]: I1201 08:42:01.833956 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-5z87f" podStartSLOduration=81.833935671 podStartE2EDuration="1m21.833935671s" podCreationTimestamp="2025-12-01 08:40:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:42:01.796383204 +0000 UTC m=+106.419304790" watchObservedRunningTime="2025-12-01 08:42:01.833935671 +0000 UTC m=+106.456857247" Dec 01 08:42:01 crc kubenswrapper[4813]: I1201 08:42:01.850109 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-49bkd" podStartSLOduration=9.85008906 podStartE2EDuration="9.85008906s" podCreationTimestamp="2025-12-01 08:41:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:42:01.847244344 +0000 UTC m=+106.470165930" watchObservedRunningTime="2025-12-01 08:42:01.85008906 +0000 UTC m=+106.473010646" Dec 01 08:42:01 crc kubenswrapper[4813]: I1201 08:42:01.878987 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 08:42:01 crc kubenswrapper[4813]: E1201 08:42:01.879711 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 08:42:02.379693088 +0000 UTC m=+107.002614674 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:42:02 crc kubenswrapper[4813]: I1201 08:42:01.980728 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mj64s\" (UID: \"a2b231fa-0ecc-44da-9259-a7dc7261127c\") " pod="openshift-image-registry/image-registry-697d97f7c8-mj64s" Dec 01 08:42:02 crc kubenswrapper[4813]: E1201 08:42:01.981080 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 08:42:02.481067731 +0000 UTC m=+107.103989317 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mj64s" (UID: "a2b231fa-0ecc-44da-9259-a7dc7261127c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:42:02 crc kubenswrapper[4813]: I1201 08:42:02.406745 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 08:42:02 crc kubenswrapper[4813]: E1201 08:42:02.407624 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 08:42:02.90755084 +0000 UTC m=+107.530472426 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:42:02 crc kubenswrapper[4813]: I1201 08:42:02.508340 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mj64s\" (UID: \"a2b231fa-0ecc-44da-9259-a7dc7261127c\") " pod="openshift-image-registry/image-registry-697d97f7c8-mj64s" Dec 01 08:42:02 crc kubenswrapper[4813]: E1201 08:42:02.510173 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 08:42:03.010158461 +0000 UTC m=+107.633080047 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mj64s" (UID: "a2b231fa-0ecc-44da-9259-a7dc7261127c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:42:02 crc kubenswrapper[4813]: I1201 08:42:02.530245 4813 patch_prober.go:28] interesting pod/route-controller-manager-6576b87f9c-2mb7d container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.19:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 01 08:42:02 crc kubenswrapper[4813]: I1201 08:42:02.530364 4813 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-2mb7d" podUID="0f0340e8-faa5-447b-a5cd-a878e982d98f" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.19:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 01 08:42:02 crc kubenswrapper[4813]: I1201 08:42:02.565531 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-f8kzc" event={"ID":"2e3f10d2-ac94-459c-9898-542e9f710747","Type":"ContainerStarted","Data":"6a31834f93702e5ece3b6f5c4dc1b94d7183a81ac3cdb4583b69307bdc881730"} Dec 01 08:42:02 crc kubenswrapper[4813]: I1201 08:42:02.566802 4813 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-f8kzc container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.25:5443/healthz\": dial tcp 10.217.0.25:5443: connect: connection refused" start-of-body= Dec 01 08:42:02 crc kubenswrapper[4813]: I1201 08:42:02.566844 4813 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-f8kzc" podUID="2e3f10d2-ac94-459c-9898-542e9f710747" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.25:5443/healthz\": dial tcp 10.217.0.25:5443: connect: connection refused" Dec 01 08:42:02 crc kubenswrapper[4813]: E1201 08:42:02.611787 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 08:42:03.111751865 +0000 UTC m=+107.734673451 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:42:02 crc kubenswrapper[4813]: I1201 08:42:02.611536 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 08:42:02 crc kubenswrapper[4813]: I1201 08:42:02.669530 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mj64s\" (UID: \"a2b231fa-0ecc-44da-9259-a7dc7261127c\") " pod="openshift-image-registry/image-registry-697d97f7c8-mj64s" Dec 01 08:42:02 crc kubenswrapper[4813]: E1201 08:42:02.670268 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 08:42:03.170258258 +0000 UTC m=+107.793179834 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mj64s" (UID: "a2b231fa-0ecc-44da-9259-a7dc7261127c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:42:02 crc kubenswrapper[4813]: I1201 08:42:02.677407 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-5z87f" event={"ID":"a7588f88-9a71-4e13-9fc3-a34cbfa6a090","Type":"ContainerStarted","Data":"e983cc1cdbcdfdb799a69bbd6ab79c064ad26831e8c80d402b07535c82b9cbb5"} Dec 01 08:42:02 crc kubenswrapper[4813]: I1201 08:42:02.700222 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-j7v65" event={"ID":"61054467-59d0-40a9-a60c-09c6c6106acd","Type":"ContainerStarted","Data":"b443b8e5752c9be49f32c51e696d52f652920ff75669843029d0dbb4d34d7869"} Dec 01 08:42:02 crc kubenswrapper[4813]: I1201 08:42:02.700297 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-j7v65" event={"ID":"61054467-59d0-40a9-a60c-09c6c6106acd","Type":"ContainerStarted","Data":"1a5f909f26b451be3c925ee37c545b0881a6b8759258c7b0cbbffa63cacbac74"} Dec 01 08:42:02 crc kubenswrapper[4813]: I1201 08:42:02.701464 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-j7v65" Dec 01 08:42:02 crc kubenswrapper[4813]: I1201 08:42:02.734316 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29409630-45pn5" podStartSLOduration=82.734267433 podStartE2EDuration="1m22.734267433s" podCreationTimestamp="2025-12-01 08:40:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:42:01.87385578 +0000 UTC m=+106.496777366" watchObservedRunningTime="2025-12-01 08:42:02.734267433 +0000 UTC m=+107.357189019" Dec 01 08:42:02 crc kubenswrapper[4813]: I1201 08:42:02.745344 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-jb992" event={"ID":"7d654f9a-b190-47bc-89ce-0caae2ec1513","Type":"ContainerStarted","Data":"bc0fe65995f52e3665f6aa614c70086375218cbce579017520bc6b03abf628ce"} Dec 01 08:42:02 crc kubenswrapper[4813]: I1201 08:42:02.770375 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 08:42:02 crc kubenswrapper[4813]: E1201 08:42:02.770475 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 08:42:03.270454525 +0000 UTC m=+107.893376111 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:42:02 crc kubenswrapper[4813]: I1201 08:42:02.770860 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mj64s\" (UID: \"a2b231fa-0ecc-44da-9259-a7dc7261127c\") " pod="openshift-image-registry/image-registry-697d97f7c8-mj64s" Dec 01 08:42:02 crc kubenswrapper[4813]: E1201 08:42:02.772724 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 08:42:03.272710012 +0000 UTC m=+107.895631608 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mj64s" (UID: "a2b231fa-0ecc-44da-9259-a7dc7261127c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:42:02 crc kubenswrapper[4813]: I1201 08:42:02.779519 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-96mbr" event={"ID":"0789e3f7-5b08-4c04-8596-9ba29f50cdd0","Type":"ContainerStarted","Data":"4bcdc7b7541994b1a825d7f555e8cf2ef2b16879a388f304029b7ea94b734d4a"} Dec 01 08:42:02 crc kubenswrapper[4813]: I1201 08:42:02.782000 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-jb992" podStartSLOduration=82.781955132 podStartE2EDuration="1m22.781955132s" podCreationTimestamp="2025-12-01 08:40:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:42:02.774048036 +0000 UTC m=+107.396969622" watchObservedRunningTime="2025-12-01 08:42:02.781955132 +0000 UTC m=+107.404876728" Dec 01 08:42:02 crc kubenswrapper[4813]: I1201 08:42:02.785519 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-j7v65" podStartSLOduration=10.78549771 podStartE2EDuration="10.78549771s" podCreationTimestamp="2025-12-01 08:41:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:42:02.737592161 +0000 UTC m=+107.360513747" watchObservedRunningTime="2025-12-01 08:42:02.78549771 +0000 UTC m=+107.408419296" Dec 01 08:42:02 crc kubenswrapper[4813]: I1201 08:42:02.809172 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-l5xp9" event={"ID":"7caa1280-3fcc-47b0-ab38-8aa5479fcfce","Type":"ContainerStarted","Data":"a7e09491968565bc67b09ddccd685bc88f82d58d8e053e8e9218c95d8d3d775f"} Dec 01 08:42:02 crc kubenswrapper[4813]: I1201 08:42:02.809250 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-l5xp9" event={"ID":"7caa1280-3fcc-47b0-ab38-8aa5479fcfce","Type":"ContainerStarted","Data":"1e171602156d27376f71b6d2c1274f34fb7991c79c31bf7c71c25cdfe352f776"} Dec 01 08:42:02 crc kubenswrapper[4813]: I1201 08:42:02.812023 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-l5xp9" Dec 01 08:42:02 crc kubenswrapper[4813]: I1201 08:42:02.823232 4813 patch_prober.go:28] interesting pod/router-default-5444994796-sp4wh container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 01 08:42:02 crc kubenswrapper[4813]: [-]has-synced failed: reason withheld Dec 01 08:42:02 crc kubenswrapper[4813]: [+]process-running ok Dec 01 08:42:02 crc kubenswrapper[4813]: healthz check failed Dec 01 08:42:02 crc kubenswrapper[4813]: I1201 08:42:02.823292 4813 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-sp4wh" podUID="a905919c-6c7e-40be-8f43-b938a60f14f6" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 01 08:42:02 crc kubenswrapper[4813]: I1201 08:42:02.851858 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-vcwrd" event={"ID":"1e597d43-ddb6-4cdb-8ce7-94b7427fb2f0","Type":"ContainerStarted","Data":"ecc48377b0cee1eee58e6590d833f6f984edfb723bf398ac93cd3aa0c861aa35"} Dec 01 08:42:03 crc kubenswrapper[4813]: I1201 08:42:02.874565 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 08:42:03 crc kubenswrapper[4813]: E1201 08:42:02.875086 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 08:42:03.375068112 +0000 UTC m=+107.997989698 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:42:03 crc kubenswrapper[4813]: I1201 08:42:02.978993 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mj64s\" (UID: \"a2b231fa-0ecc-44da-9259-a7dc7261127c\") " pod="openshift-image-registry/image-registry-697d97f7c8-mj64s" Dec 01 08:42:03 crc kubenswrapper[4813]: E1201 08:42:02.980424 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 08:42:03.480411103 +0000 UTC m=+108.103332689 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mj64s" (UID: "a2b231fa-0ecc-44da-9259-a7dc7261127c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:42:03 crc kubenswrapper[4813]: I1201 08:42:03.082458 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 08:42:03 crc kubenswrapper[4813]: E1201 08:42:03.082979 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 08:42:03.582945747 +0000 UTC m=+108.205867333 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:42:03 crc kubenswrapper[4813]: I1201 08:42:03.108181 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-l5xp9" podStartSLOduration=83.108155791 podStartE2EDuration="1m23.108155791s" podCreationTimestamp="2025-12-01 08:40:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:42:03.107208531 +0000 UTC m=+107.730130117" watchObservedRunningTime="2025-12-01 08:42:03.108155791 +0000 UTC m=+107.731077377" Dec 01 08:42:03 crc kubenswrapper[4813]: I1201 08:42:03.116216 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-49bkd" event={"ID":"adf5f998-8117-4362-868e-b7cdbbf91e0e","Type":"ContainerStarted","Data":"e6d8eda8d9b185ce4018c0ca4e0d1ffd4f1568f92d1a57bd38bcd0f368af02d8"} Dec 01 08:42:03 crc kubenswrapper[4813]: I1201 08:42:03.155168 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-sjjm7" event={"ID":"dadb7bca-672d-434b-897d-7eb83059f189","Type":"ContainerStarted","Data":"f087a90937aca264abe546a8ed9b499f726db427da08745056b8cf932842562f"} Dec 01 08:42:03 crc kubenswrapper[4813]: I1201 08:42:03.167260 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-dwkm7" event={"ID":"f3561d3e-0e57-4243-83df-9a475da69fb6","Type":"ContainerStarted","Data":"562965efef81d9c0951d55f3dfcb9e1de49ecf86800b495801088dd49197cc1e"} Dec 01 08:42:03 crc kubenswrapper[4813]: I1201 08:42:03.167809 4813 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-qsskc container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.20:8080/healthz\": dial tcp 10.217.0.20:8080: connect: connection refused" start-of-body= Dec 01 08:42:03 crc kubenswrapper[4813]: I1201 08:42:03.167842 4813 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-qsskc" podUID="c342090f-f331-484d-b287-929d8303f5d7" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.20:8080/healthz\": dial tcp 10.217.0.20:8080: connect: connection refused" Dec 01 08:42:03 crc kubenswrapper[4813]: I1201 08:42:03.168449 4813 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-n4dp7 container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.41:8443/healthz\": dial tcp 10.217.0.41:8443: connect: connection refused" start-of-body= Dec 01 08:42:03 crc kubenswrapper[4813]: I1201 08:42:03.168493 4813 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-n4dp7" podUID="5aa447ba-69b3-4291-9db5-25f4aeee2370" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.41:8443/healthz\": dial tcp 10.217.0.41:8443: connect: connection refused" Dec 01 08:42:03 crc kubenswrapper[4813]: I1201 08:42:03.168588 4813 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-hf8fp container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.29:8443/healthz\": dial tcp 10.217.0.29:8443: connect: connection refused" start-of-body= Dec 01 08:42:03 crc kubenswrapper[4813]: I1201 08:42:03.168609 4813 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-hf8fp" podUID="489e9f98-0728-4a72-80d4-3d88a77823f5" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.29:8443/healthz\": dial tcp 10.217.0.29:8443: connect: connection refused" Dec 01 08:42:03 crc kubenswrapper[4813]: I1201 08:42:03.512718 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mj64s\" (UID: \"a2b231fa-0ecc-44da-9259-a7dc7261127c\") " pod="openshift-image-registry/image-registry-697d97f7c8-mj64s" Dec 01 08:42:03 crc kubenswrapper[4813]: E1201 08:42:03.514362 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 08:42:04.014347049 +0000 UTC m=+108.637268635 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mj64s" (UID: "a2b231fa-0ecc-44da-9259-a7dc7261127c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:42:03 crc kubenswrapper[4813]: I1201 08:42:03.767252 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 08:42:03 crc kubenswrapper[4813]: E1201 08:42:03.769315 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 08:42:04.269290518 +0000 UTC m=+108.892212104 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:42:03 crc kubenswrapper[4813]: I1201 08:42:03.783406 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-9ql8j" Dec 01 08:42:03 crc kubenswrapper[4813]: I1201 08:42:03.783476 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-9ql8j" Dec 01 08:42:03 crc kubenswrapper[4813]: I1201 08:42:03.912431 4813 patch_prober.go:28] interesting pod/console-f9d7485db-9ql8j container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.10:8443/health\": dial tcp 10.217.0.10:8443: connect: connection refused" start-of-body= Dec 01 08:42:03 crc kubenswrapper[4813]: I1201 08:42:03.912501 4813 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-9ql8j" podUID="65bbc5b9-f6f0-40ad-9e47-4122bb5cabab" containerName="console" probeResult="failure" output="Get \"https://10.217.0.10:8443/health\": dial tcp 10.217.0.10:8443: connect: connection refused" Dec 01 08:42:03 crc kubenswrapper[4813]: I1201 08:42:03.913281 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mj64s\" (UID: \"a2b231fa-0ecc-44da-9259-a7dc7261127c\") " pod="openshift-image-registry/image-registry-697d97f7c8-mj64s" Dec 01 08:42:03 crc kubenswrapper[4813]: E1201 08:42:03.914674 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 08:42:04.414657337 +0000 UTC m=+109.037578923 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mj64s" (UID: "a2b231fa-0ecc-44da-9259-a7dc7261127c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:42:03 crc kubenswrapper[4813]: I1201 08:42:03.922400 4813 patch_prober.go:28] interesting pod/router-default-5444994796-sp4wh container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 01 08:42:03 crc kubenswrapper[4813]: [-]has-synced failed: reason withheld Dec 01 08:42:03 crc kubenswrapper[4813]: [+]process-running ok Dec 01 08:42:03 crc kubenswrapper[4813]: healthz check failed Dec 01 08:42:03 crc kubenswrapper[4813]: I1201 08:42:03.922463 4813 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-sp4wh" podUID="a905919c-6c7e-40be-8f43-b938a60f14f6" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 01 08:42:03 crc kubenswrapper[4813]: I1201 08:42:03.938256 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-dwkm7" podStartSLOduration=83.93823951 podStartE2EDuration="1m23.93823951s" podCreationTimestamp="2025-12-01 08:40:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:42:03.527730999 +0000 UTC m=+108.150652585" watchObservedRunningTime="2025-12-01 08:42:03.93823951 +0000 UTC m=+108.561161096" Dec 01 08:42:03 crc kubenswrapper[4813]: I1201 08:42:03.938523 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Dec 01 08:42:03 crc kubenswrapper[4813]: I1201 08:42:03.939392 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 01 08:42:03 crc kubenswrapper[4813]: I1201 08:42:03.940610 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-2mb7d" Dec 01 08:42:03 crc kubenswrapper[4813]: I1201 08:42:03.941696 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Dec 01 08:42:03 crc kubenswrapper[4813]: I1201 08:42:03.942031 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Dec 01 08:42:03 crc kubenswrapper[4813]: I1201 08:42:03.949881 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-hmh5v" Dec 01 08:42:03 crc kubenswrapper[4813]: I1201 08:42:03.960694 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Dec 01 08:42:04 crc kubenswrapper[4813]: I1201 08:42:04.016046 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 08:42:04 crc kubenswrapper[4813]: E1201 08:42:04.016770 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 08:42:04.516750658 +0000 UTC m=+109.139672244 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:42:04 crc kubenswrapper[4813]: I1201 08:42:04.025011 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-l4zh5" Dec 01 08:42:04 crc kubenswrapper[4813]: I1201 08:42:04.124023 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mj64s\" (UID: \"a2b231fa-0ecc-44da-9259-a7dc7261127c\") " pod="openshift-image-registry/image-registry-697d97f7c8-mj64s" Dec 01 08:42:04 crc kubenswrapper[4813]: I1201 08:42:04.124197 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/62fe4f70-1303-4787-ad08-24efee57578a-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"62fe4f70-1303-4787-ad08-24efee57578a\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 01 08:42:04 crc kubenswrapper[4813]: I1201 08:42:04.124315 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/62fe4f70-1303-4787-ad08-24efee57578a-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"62fe4f70-1303-4787-ad08-24efee57578a\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 01 08:42:04 crc kubenswrapper[4813]: E1201 08:42:04.125409 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 08:42:04.62538945 +0000 UTC m=+109.248311036 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mj64s" (UID: "a2b231fa-0ecc-44da-9259-a7dc7261127c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:42:04 crc kubenswrapper[4813]: I1201 08:42:04.135529 4813 patch_prober.go:28] interesting pod/downloads-7954f5f757-ct7w8 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" start-of-body= Dec 01 08:42:04 crc kubenswrapper[4813]: I1201 08:42:04.135580 4813 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-ct7w8" podUID="e17c11d6-eb41-4391-b513-a817eeffcdb4" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" Dec 01 08:42:04 crc kubenswrapper[4813]: I1201 08:42:04.135596 4813 patch_prober.go:28] interesting pod/downloads-7954f5f757-ct7w8 container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" start-of-body= Dec 01 08:42:04 crc kubenswrapper[4813]: I1201 08:42:04.135655 4813 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-ct7w8" podUID="e17c11d6-eb41-4391-b513-a817eeffcdb4" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" Dec 01 08:42:04 crc kubenswrapper[4813]: I1201 08:42:04.185669 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-96mbr" event={"ID":"0789e3f7-5b08-4c04-8596-9ba29f50cdd0","Type":"ContainerStarted","Data":"c60078c9bd8097a23c0b9891a5b0dea935aa7ac17b5fa06798318ec8bdeac600"} Dec 01 08:42:04 crc kubenswrapper[4813]: I1201 08:42:04.185707 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-96mbr" event={"ID":"0789e3f7-5b08-4c04-8596-9ba29f50cdd0","Type":"ContainerStarted","Data":"0e60cd03633e762e5f0fbe61bb1c22317e81abf34ef811fd7397828643315ee1"} Dec 01 08:42:04 crc kubenswrapper[4813]: I1201 08:42:04.186220 4813 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-f8kzc container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.25:5443/healthz\": dial tcp 10.217.0.25:5443: connect: connection refused" start-of-body= Dec 01 08:42:04 crc kubenswrapper[4813]: I1201 08:42:04.186254 4813 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-f8kzc" podUID="2e3f10d2-ac94-459c-9898-542e9f710747" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.25:5443/healthz\": dial tcp 10.217.0.25:5443: connect: connection refused" Dec 01 08:42:04 crc kubenswrapper[4813]: I1201 08:42:04.219413 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-n4dp7" Dec 01 08:42:04 crc kubenswrapper[4813]: I1201 08:42:04.226497 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 08:42:04 crc kubenswrapper[4813]: E1201 08:42:04.226590 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 08:42:04.726571473 +0000 UTC m=+109.349493059 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:42:04 crc kubenswrapper[4813]: I1201 08:42:04.226944 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mj64s\" (UID: \"a2b231fa-0ecc-44da-9259-a7dc7261127c\") " pod="openshift-image-registry/image-registry-697d97f7c8-mj64s" Dec 01 08:42:04 crc kubenswrapper[4813]: I1201 08:42:04.227067 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/62fe4f70-1303-4787-ad08-24efee57578a-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"62fe4f70-1303-4787-ad08-24efee57578a\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 01 08:42:04 crc kubenswrapper[4813]: I1201 08:42:04.227208 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/62fe4f70-1303-4787-ad08-24efee57578a-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"62fe4f70-1303-4787-ad08-24efee57578a\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 01 08:42:04 crc kubenswrapper[4813]: E1201 08:42:04.228366 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 08:42:04.728356228 +0000 UTC m=+109.351277804 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mj64s" (UID: "a2b231fa-0ecc-44da-9259-a7dc7261127c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:42:04 crc kubenswrapper[4813]: I1201 08:42:04.229371 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/62fe4f70-1303-4787-ad08-24efee57578a-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"62fe4f70-1303-4787-ad08-24efee57578a\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 01 08:42:04 crc kubenswrapper[4813]: I1201 08:42:04.243396 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-96mbr" podStartSLOduration=84.243374289 podStartE2EDuration="1m24.243374289s" podCreationTimestamp="2025-12-01 08:40:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:42:04.241404418 +0000 UTC m=+108.864326004" watchObservedRunningTime="2025-12-01 08:42:04.243374289 +0000 UTC m=+108.866295875" Dec 01 08:42:04 crc kubenswrapper[4813]: I1201 08:42:04.297265 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/62fe4f70-1303-4787-ad08-24efee57578a-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"62fe4f70-1303-4787-ad08-24efee57578a\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 01 08:42:04 crc kubenswrapper[4813]: I1201 08:42:04.332122 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 08:42:04 crc kubenswrapper[4813]: E1201 08:42:04.332684 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 08:42:04.832668437 +0000 UTC m=+109.455590023 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:42:04 crc kubenswrapper[4813]: I1201 08:42:04.443694 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mj64s\" (UID: \"a2b231fa-0ecc-44da-9259-a7dc7261127c\") " pod="openshift-image-registry/image-registry-697d97f7c8-mj64s" Dec 01 08:42:04 crc kubenswrapper[4813]: E1201 08:42:04.444337 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 08:42:04.944324482 +0000 UTC m=+109.567246068 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mj64s" (UID: "a2b231fa-0ecc-44da-9259-a7dc7261127c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:42:04 crc kubenswrapper[4813]: I1201 08:42:04.466313 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lt4dx" Dec 01 08:42:04 crc kubenswrapper[4813]: I1201 08:42:04.474641 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lt4dx" Dec 01 08:42:04 crc kubenswrapper[4813]: I1201 08:42:04.546934 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 08:42:04 crc kubenswrapper[4813]: E1201 08:42:04.547447 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 08:42:05.047430854 +0000 UTC m=+109.670352440 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:42:04 crc kubenswrapper[4813]: I1201 08:42:04.575187 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 01 08:42:04 crc kubenswrapper[4813]: I1201 08:42:04.648789 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mj64s\" (UID: \"a2b231fa-0ecc-44da-9259-a7dc7261127c\") " pod="openshift-image-registry/image-registry-697d97f7c8-mj64s" Dec 01 08:42:04 crc kubenswrapper[4813]: E1201 08:42:04.650305 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 08:42:05.150289149 +0000 UTC m=+109.773210735 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mj64s" (UID: "a2b231fa-0ecc-44da-9259-a7dc7261127c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:42:04 crc kubenswrapper[4813]: I1201 08:42:04.784696 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 08:42:04 crc kubenswrapper[4813]: E1201 08:42:04.785089 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 08:42:05.285063383 +0000 UTC m=+109.907984969 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:42:04 crc kubenswrapper[4813]: I1201 08:42:04.835841 4813 patch_prober.go:28] interesting pod/router-default-5444994796-sp4wh container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 01 08:42:04 crc kubenswrapper[4813]: [-]has-synced failed: reason withheld Dec 01 08:42:04 crc kubenswrapper[4813]: [+]process-running ok Dec 01 08:42:04 crc kubenswrapper[4813]: healthz check failed Dec 01 08:42:04 crc kubenswrapper[4813]: I1201 08:42:04.835907 4813 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-sp4wh" podUID="a905919c-6c7e-40be-8f43-b938a60f14f6" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 01 08:42:04 crc kubenswrapper[4813]: I1201 08:42:04.886615 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mj64s\" (UID: \"a2b231fa-0ecc-44da-9259-a7dc7261127c\") " pod="openshift-image-registry/image-registry-697d97f7c8-mj64s" Dec 01 08:42:04 crc kubenswrapper[4813]: E1201 08:42:04.887032 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 08:42:05.38701449 +0000 UTC m=+110.009936076 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mj64s" (UID: "a2b231fa-0ecc-44da-9259-a7dc7261127c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:42:04 crc kubenswrapper[4813]: I1201 08:42:04.987709 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 08:42:04 crc kubenswrapper[4813]: E1201 08:42:04.987873 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 08:42:05.487842062 +0000 UTC m=+110.110763648 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:42:04 crc kubenswrapper[4813]: I1201 08:42:04.988098 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mj64s\" (UID: \"a2b231fa-0ecc-44da-9259-a7dc7261127c\") " pod="openshift-image-registry/image-registry-697d97f7c8-mj64s" Dec 01 08:42:04 crc kubenswrapper[4813]: E1201 08:42:04.988485 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 08:42:05.488467381 +0000 UTC m=+110.111388957 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mj64s" (UID: "a2b231fa-0ecc-44da-9259-a7dc7261127c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:42:05 crc kubenswrapper[4813]: I1201 08:42:05.106581 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 08:42:05 crc kubenswrapper[4813]: E1201 08:42:05.106932 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 08:42:05.606906564 +0000 UTC m=+110.229828150 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:42:05 crc kubenswrapper[4813]: I1201 08:42:05.207684 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mj64s\" (UID: \"a2b231fa-0ecc-44da-9259-a7dc7261127c\") " pod="openshift-image-registry/image-registry-697d97f7c8-mj64s" Dec 01 08:42:05 crc kubenswrapper[4813]: E1201 08:42:05.208139 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 08:42:05.708120318 +0000 UTC m=+110.331041904 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mj64s" (UID: "a2b231fa-0ecc-44da-9259-a7dc7261127c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:42:05 crc kubenswrapper[4813]: I1201 08:42:05.297143 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-m8jlq" event={"ID":"363f808f-a265-464a-a3a7-4128e3734b14","Type":"ContainerStarted","Data":"f27c8392365ef5482ac3431240b0e6fe59d60b52d354e58bdd615cc34ae4c010"} Dec 01 08:42:05 crc kubenswrapper[4813]: I1201 08:42:05.297230 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-m8jlq" event={"ID":"363f808f-a265-464a-a3a7-4128e3734b14","Type":"ContainerStarted","Data":"f2d91925add028c1cb28d3aab0f5335004e3a9ac4c0a6780630146cabbce97e8"} Dec 01 08:42:05 crc kubenswrapper[4813]: I1201 08:42:05.309534 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 08:42:05 crc kubenswrapper[4813]: E1201 08:42:05.310142 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 08:42:05.810124977 +0000 UTC m=+110.433046563 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:42:05 crc kubenswrapper[4813]: I1201 08:42:05.411912 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mj64s\" (UID: \"a2b231fa-0ecc-44da-9259-a7dc7261127c\") " pod="openshift-image-registry/image-registry-697d97f7c8-mj64s" Dec 01 08:42:05 crc kubenswrapper[4813]: E1201 08:42:05.414120 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 08:42:05.914106816 +0000 UTC m=+110.537028402 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mj64s" (UID: "a2b231fa-0ecc-44da-9259-a7dc7261127c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:42:05 crc kubenswrapper[4813]: I1201 08:42:05.471563 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-7w2kz" Dec 01 08:42:05 crc kubenswrapper[4813]: I1201 08:42:05.471743 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-5fgn9" Dec 01 08:42:05 crc kubenswrapper[4813]: I1201 08:42:05.520558 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 08:42:05 crc kubenswrapper[4813]: E1201 08:42:05.521112 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 08:42:06.021094197 +0000 UTC m=+110.644015783 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:42:05 crc kubenswrapper[4813]: I1201 08:42:05.594374 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-hf8fp" Dec 01 08:42:05 crc kubenswrapper[4813]: I1201 08:42:05.621866 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mj64s\" (UID: \"a2b231fa-0ecc-44da-9259-a7dc7261127c\") " pod="openshift-image-registry/image-registry-697d97f7c8-mj64s" Dec 01 08:42:05 crc kubenswrapper[4813]: E1201 08:42:05.623606 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 08:42:06.123593791 +0000 UTC m=+110.746515377 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mj64s" (UID: "a2b231fa-0ecc-44da-9259-a7dc7261127c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:42:05 crc kubenswrapper[4813]: I1201 08:42:05.723396 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 08:42:05 crc kubenswrapper[4813]: E1201 08:42:05.723991 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 08:42:06.22397504 +0000 UTC m=+110.846896626 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:42:05 crc kubenswrapper[4813]: I1201 08:42:05.811802 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-sp4wh" Dec 01 08:42:05 crc kubenswrapper[4813]: I1201 08:42:05.817152 4813 patch_prober.go:28] interesting pod/router-default-5444994796-sp4wh container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 01 08:42:05 crc kubenswrapper[4813]: [-]has-synced failed: reason withheld Dec 01 08:42:05 crc kubenswrapper[4813]: [+]process-running ok Dec 01 08:42:05 crc kubenswrapper[4813]: healthz check failed Dec 01 08:42:05 crc kubenswrapper[4813]: I1201 08:42:05.817204 4813 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-sp4wh" podUID="a905919c-6c7e-40be-8f43-b938a60f14f6" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 01 08:42:05 crc kubenswrapper[4813]: I1201 08:42:05.821095 4813 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-qsskc container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.20:8080/healthz\": dial tcp 10.217.0.20:8080: connect: connection refused" start-of-body= Dec 01 08:42:05 crc kubenswrapper[4813]: I1201 08:42:05.821109 4813 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-qsskc container/marketplace-operator namespace/openshift-marketplace: Liveness probe status=failure output="Get \"http://10.217.0.20:8080/healthz\": dial tcp 10.217.0.20:8080: connect: connection refused" start-of-body= Dec 01 08:42:05 crc kubenswrapper[4813]: I1201 08:42:05.821124 4813 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-qsskc" podUID="c342090f-f331-484d-b287-929d8303f5d7" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.20:8080/healthz\": dial tcp 10.217.0.20:8080: connect: connection refused" Dec 01 08:42:05 crc kubenswrapper[4813]: I1201 08:42:05.821142 4813 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-marketplace/marketplace-operator-79b997595-qsskc" podUID="c342090f-f331-484d-b287-929d8303f5d7" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.20:8080/healthz\": dial tcp 10.217.0.20:8080: connect: connection refused" Dec 01 08:42:05 crc kubenswrapper[4813]: I1201 08:42:05.824865 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mj64s\" (UID: \"a2b231fa-0ecc-44da-9259-a7dc7261127c\") " pod="openshift-image-registry/image-registry-697d97f7c8-mj64s" Dec 01 08:42:05 crc kubenswrapper[4813]: E1201 08:42:05.825190 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 08:42:06.325178474 +0000 UTC m=+110.948100060 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mj64s" (UID: "a2b231fa-0ecc-44da-9259-a7dc7261127c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:42:05 crc kubenswrapper[4813]: I1201 08:42:05.834998 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Dec 01 08:42:05 crc kubenswrapper[4813]: W1201 08:42:05.846119 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod62fe4f70_1303_4787_ad08_24efee57578a.slice/crio-e932ab3f01d5c0c15049c134b104a28d6cd817157803a385385fcf74b37f5f24 WatchSource:0}: Error finding container e932ab3f01d5c0c15049c134b104a28d6cd817157803a385385fcf74b37f5f24: Status 404 returned error can't find the container with id e932ab3f01d5c0c15049c134b104a28d6cd817157803a385385fcf74b37f5f24 Dec 01 08:42:05 crc kubenswrapper[4813]: I1201 08:42:05.875479 4813 patch_prober.go:28] interesting pod/apiserver-76f77b778f-5ptsh container/openshift-apiserver namespace/openshift-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[+]ping ok Dec 01 08:42:05 crc kubenswrapper[4813]: [+]log ok Dec 01 08:42:05 crc kubenswrapper[4813]: [+]etcd ok Dec 01 08:42:05 crc kubenswrapper[4813]: [+]poststarthook/start-apiserver-admission-initializer ok Dec 01 08:42:05 crc kubenswrapper[4813]: [+]poststarthook/generic-apiserver-start-informers ok Dec 01 08:42:05 crc kubenswrapper[4813]: [+]poststarthook/max-in-flight-filter ok Dec 01 08:42:05 crc kubenswrapper[4813]: [+]poststarthook/storage-object-count-tracker-hook ok Dec 01 08:42:05 crc kubenswrapper[4813]: [+]poststarthook/image.openshift.io-apiserver-caches ok Dec 01 08:42:05 crc kubenswrapper[4813]: [-]poststarthook/authorization.openshift.io-bootstrapclusterroles failed: reason withheld Dec 01 08:42:05 crc kubenswrapper[4813]: [-]poststarthook/authorization.openshift.io-ensurenodebootstrap-sa failed: reason withheld Dec 01 08:42:05 crc kubenswrapper[4813]: [+]poststarthook/project.openshift.io-projectcache ok Dec 01 08:42:05 crc kubenswrapper[4813]: [+]poststarthook/project.openshift.io-projectauthorizationcache ok Dec 01 08:42:05 crc kubenswrapper[4813]: [+]poststarthook/openshift.io-startinformers ok Dec 01 08:42:05 crc kubenswrapper[4813]: [+]poststarthook/openshift.io-restmapperupdater ok Dec 01 08:42:05 crc kubenswrapper[4813]: [+]poststarthook/quota.openshift.io-clusterquotamapping ok Dec 01 08:42:05 crc kubenswrapper[4813]: livez check failed Dec 01 08:42:05 crc kubenswrapper[4813]: I1201 08:42:05.875543 4813 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-apiserver/apiserver-76f77b778f-5ptsh" podUID="d2aac47a-b117-42c7-afa3-2c989d9536a5" containerName="openshift-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 01 08:42:05 crc kubenswrapper[4813]: I1201 08:42:05.926465 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 08:42:05 crc kubenswrapper[4813]: E1201 08:42:05.927374 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 08:42:06.427336557 +0000 UTC m=+111.050258143 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:42:06 crc kubenswrapper[4813]: I1201 08:42:06.028788 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mj64s\" (UID: \"a2b231fa-0ecc-44da-9259-a7dc7261127c\") " pod="openshift-image-registry/image-registry-697d97f7c8-mj64s" Dec 01 08:42:06 crc kubenswrapper[4813]: E1201 08:42:06.029262 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 08:42:06.529242891 +0000 UTC m=+111.152164467 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mj64s" (UID: "a2b231fa-0ecc-44da-9259-a7dc7261127c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:42:06 crc kubenswrapper[4813]: I1201 08:42:06.037757 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-qh5hd"] Dec 01 08:42:06 crc kubenswrapper[4813]: I1201 08:42:06.039123 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-qh5hd" Dec 01 08:42:06 crc kubenswrapper[4813]: I1201 08:42:06.043079 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Dec 01 08:42:06 crc kubenswrapper[4813]: I1201 08:42:06.080317 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-qh5hd"] Dec 01 08:42:06 crc kubenswrapper[4813]: I1201 08:42:06.088868 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-f8kzc" Dec 01 08:42:06 crc kubenswrapper[4813]: I1201 08:42:06.099131 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-j4pcc" Dec 01 08:42:06 crc kubenswrapper[4813]: I1201 08:42:06.130558 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 08:42:06 crc kubenswrapper[4813]: E1201 08:42:06.130788 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 08:42:06.630747255 +0000 UTC m=+111.253668841 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:42:06 crc kubenswrapper[4813]: I1201 08:42:06.130954 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mj64s\" (UID: \"a2b231fa-0ecc-44da-9259-a7dc7261127c\") " pod="openshift-image-registry/image-registry-697d97f7c8-mj64s" Dec 01 08:42:06 crc kubenswrapper[4813]: E1201 08:42:06.131374 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 08:42:06.631367724 +0000 UTC m=+111.254289310 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mj64s" (UID: "a2b231fa-0ecc-44da-9259-a7dc7261127c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:42:06 crc kubenswrapper[4813]: I1201 08:42:06.138060 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-7fqxs"] Dec 01 08:42:06 crc kubenswrapper[4813]: I1201 08:42:06.139072 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7fqxs" Dec 01 08:42:06 crc kubenswrapper[4813]: I1201 08:42:06.147304 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Dec 01 08:42:06 crc kubenswrapper[4813]: I1201 08:42:06.164296 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-7fqxs"] Dec 01 08:42:06 crc kubenswrapper[4813]: I1201 08:42:06.255656 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 08:42:06 crc kubenswrapper[4813]: E1201 08:42:06.255820 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 08:42:06.75579045 +0000 UTC m=+111.378712036 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:42:06 crc kubenswrapper[4813]: I1201 08:42:06.255991 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/26f4804b-998f-488c-a501-d1cc4ef7f0b4-catalog-content\") pod \"certified-operators-qh5hd\" (UID: \"26f4804b-998f-488c-a501-d1cc4ef7f0b4\") " pod="openshift-marketplace/certified-operators-qh5hd" Dec 01 08:42:06 crc kubenswrapper[4813]: I1201 08:42:06.256023 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0959ee73-31e7-4d58-b124-3e971d777225-catalog-content\") pod \"community-operators-7fqxs\" (UID: \"0959ee73-31e7-4d58-b124-3e971d777225\") " pod="openshift-marketplace/community-operators-7fqxs" Dec 01 08:42:06 crc kubenswrapper[4813]: I1201 08:42:06.256044 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-srz6m\" (UniqueName: \"kubernetes.io/projected/0959ee73-31e7-4d58-b124-3e971d777225-kube-api-access-srz6m\") pod \"community-operators-7fqxs\" (UID: \"0959ee73-31e7-4d58-b124-3e971d777225\") " pod="openshift-marketplace/community-operators-7fqxs" Dec 01 08:42:06 crc kubenswrapper[4813]: I1201 08:42:06.256094 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0959ee73-31e7-4d58-b124-3e971d777225-utilities\") pod \"community-operators-7fqxs\" (UID: \"0959ee73-31e7-4d58-b124-3e971d777225\") " pod="openshift-marketplace/community-operators-7fqxs" Dec 01 08:42:06 crc kubenswrapper[4813]: I1201 08:42:06.256257 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ns6n2\" (UniqueName: \"kubernetes.io/projected/26f4804b-998f-488c-a501-d1cc4ef7f0b4-kube-api-access-ns6n2\") pod \"certified-operators-qh5hd\" (UID: \"26f4804b-998f-488c-a501-d1cc4ef7f0b4\") " pod="openshift-marketplace/certified-operators-qh5hd" Dec 01 08:42:06 crc kubenswrapper[4813]: I1201 08:42:06.256467 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/26f4804b-998f-488c-a501-d1cc4ef7f0b4-utilities\") pod \"certified-operators-qh5hd\" (UID: \"26f4804b-998f-488c-a501-d1cc4ef7f0b4\") " pod="openshift-marketplace/certified-operators-qh5hd" Dec 01 08:42:06 crc kubenswrapper[4813]: I1201 08:42:06.256735 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mj64s\" (UID: \"a2b231fa-0ecc-44da-9259-a7dc7261127c\") " pod="openshift-image-registry/image-registry-697d97f7c8-mj64s" Dec 01 08:42:06 crc kubenswrapper[4813]: E1201 08:42:06.257093 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 08:42:06.757074909 +0000 UTC m=+111.379996495 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mj64s" (UID: "a2b231fa-0ecc-44da-9259-a7dc7261127c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:42:06 crc kubenswrapper[4813]: I1201 08:42:06.307032 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-m8jlq" event={"ID":"363f808f-a265-464a-a3a7-4128e3734b14","Type":"ContainerStarted","Data":"cb14e62479099c5512f01ae3cbc94d72a3a6d6a5eec2d9f3da7cef2fdc22f4d2"} Dec 01 08:42:06 crc kubenswrapper[4813]: I1201 08:42:06.308975 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"62fe4f70-1303-4787-ad08-24efee57578a","Type":"ContainerStarted","Data":"e932ab3f01d5c0c15049c134b104a28d6cd817157803a385385fcf74b37f5f24"} Dec 01 08:42:06 crc kubenswrapper[4813]: I1201 08:42:06.341016 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-xswbf"] Dec 01 08:42:06 crc kubenswrapper[4813]: I1201 08:42:06.341476 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-m8jlq" podStartSLOduration=14.341453567 podStartE2EDuration="14.341453567s" podCreationTimestamp="2025-12-01 08:41:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:42:06.333103381 +0000 UTC m=+110.956024967" watchObservedRunningTime="2025-12-01 08:42:06.341453567 +0000 UTC m=+110.964375153" Dec 01 08:42:06 crc kubenswrapper[4813]: I1201 08:42:06.343096 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xswbf" Dec 01 08:42:06 crc kubenswrapper[4813]: I1201 08:42:06.357984 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 08:42:06 crc kubenswrapper[4813]: I1201 08:42:06.358358 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/26f4804b-998f-488c-a501-d1cc4ef7f0b4-catalog-content\") pod \"certified-operators-qh5hd\" (UID: \"26f4804b-998f-488c-a501-d1cc4ef7f0b4\") " pod="openshift-marketplace/certified-operators-qh5hd" Dec 01 08:42:06 crc kubenswrapper[4813]: I1201 08:42:06.358384 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0959ee73-31e7-4d58-b124-3e971d777225-catalog-content\") pod \"community-operators-7fqxs\" (UID: \"0959ee73-31e7-4d58-b124-3e971d777225\") " pod="openshift-marketplace/community-operators-7fqxs" Dec 01 08:42:06 crc kubenswrapper[4813]: I1201 08:42:06.358402 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-srz6m\" (UniqueName: \"kubernetes.io/projected/0959ee73-31e7-4d58-b124-3e971d777225-kube-api-access-srz6m\") pod \"community-operators-7fqxs\" (UID: \"0959ee73-31e7-4d58-b124-3e971d777225\") " pod="openshift-marketplace/community-operators-7fqxs" Dec 01 08:42:06 crc kubenswrapper[4813]: I1201 08:42:06.358448 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0959ee73-31e7-4d58-b124-3e971d777225-utilities\") pod \"community-operators-7fqxs\" (UID: \"0959ee73-31e7-4d58-b124-3e971d777225\") " pod="openshift-marketplace/community-operators-7fqxs" Dec 01 08:42:06 crc kubenswrapper[4813]: I1201 08:42:06.358487 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ns6n2\" (UniqueName: \"kubernetes.io/projected/26f4804b-998f-488c-a501-d1cc4ef7f0b4-kube-api-access-ns6n2\") pod \"certified-operators-qh5hd\" (UID: \"26f4804b-998f-488c-a501-d1cc4ef7f0b4\") " pod="openshift-marketplace/certified-operators-qh5hd" Dec 01 08:42:06 crc kubenswrapper[4813]: I1201 08:42:06.358546 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a81d6f9c-e9f7-4c98-8233-c9f214978bfa-utilities\") pod \"certified-operators-xswbf\" (UID: \"a81d6f9c-e9f7-4c98-8233-c9f214978bfa\") " pod="openshift-marketplace/certified-operators-xswbf" Dec 01 08:42:06 crc kubenswrapper[4813]: I1201 08:42:06.358589 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/26f4804b-998f-488c-a501-d1cc4ef7f0b4-utilities\") pod \"certified-operators-qh5hd\" (UID: \"26f4804b-998f-488c-a501-d1cc4ef7f0b4\") " pod="openshift-marketplace/certified-operators-qh5hd" Dec 01 08:42:06 crc kubenswrapper[4813]: I1201 08:42:06.358626 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a81d6f9c-e9f7-4c98-8233-c9f214978bfa-catalog-content\") pod \"certified-operators-xswbf\" (UID: \"a81d6f9c-e9f7-4c98-8233-c9f214978bfa\") " pod="openshift-marketplace/certified-operators-xswbf" Dec 01 08:42:06 crc kubenswrapper[4813]: I1201 08:42:06.358646 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f7vm5\" (UniqueName: \"kubernetes.io/projected/a81d6f9c-e9f7-4c98-8233-c9f214978bfa-kube-api-access-f7vm5\") pod \"certified-operators-xswbf\" (UID: \"a81d6f9c-e9f7-4c98-8233-c9f214978bfa\") " pod="openshift-marketplace/certified-operators-xswbf" Dec 01 08:42:06 crc kubenswrapper[4813]: E1201 08:42:06.358993 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 08:42:06.858951994 +0000 UTC m=+111.481873580 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:42:06 crc kubenswrapper[4813]: I1201 08:42:06.359606 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0959ee73-31e7-4d58-b124-3e971d777225-utilities\") pod \"community-operators-7fqxs\" (UID: \"0959ee73-31e7-4d58-b124-3e971d777225\") " pod="openshift-marketplace/community-operators-7fqxs" Dec 01 08:42:06 crc kubenswrapper[4813]: I1201 08:42:06.359909 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/26f4804b-998f-488c-a501-d1cc4ef7f0b4-catalog-content\") pod \"certified-operators-qh5hd\" (UID: \"26f4804b-998f-488c-a501-d1cc4ef7f0b4\") " pod="openshift-marketplace/certified-operators-qh5hd" Dec 01 08:42:06 crc kubenswrapper[4813]: I1201 08:42:06.360137 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0959ee73-31e7-4d58-b124-3e971d777225-catalog-content\") pod \"community-operators-7fqxs\" (UID: \"0959ee73-31e7-4d58-b124-3e971d777225\") " pod="openshift-marketplace/community-operators-7fqxs" Dec 01 08:42:06 crc kubenswrapper[4813]: I1201 08:42:06.360369 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/26f4804b-998f-488c-a501-d1cc4ef7f0b4-utilities\") pod \"certified-operators-qh5hd\" (UID: \"26f4804b-998f-488c-a501-d1cc4ef7f0b4\") " pod="openshift-marketplace/certified-operators-qh5hd" Dec 01 08:42:06 crc kubenswrapper[4813]: I1201 08:42:06.460367 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a81d6f9c-e9f7-4c98-8233-c9f214978bfa-utilities\") pod \"certified-operators-xswbf\" (UID: \"a81d6f9c-e9f7-4c98-8233-c9f214978bfa\") " pod="openshift-marketplace/certified-operators-xswbf" Dec 01 08:42:06 crc kubenswrapper[4813]: I1201 08:42:06.460450 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a81d6f9c-e9f7-4c98-8233-c9f214978bfa-catalog-content\") pod \"certified-operators-xswbf\" (UID: \"a81d6f9c-e9f7-4c98-8233-c9f214978bfa\") " pod="openshift-marketplace/certified-operators-xswbf" Dec 01 08:42:06 crc kubenswrapper[4813]: I1201 08:42:06.460478 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f7vm5\" (UniqueName: \"kubernetes.io/projected/a81d6f9c-e9f7-4c98-8233-c9f214978bfa-kube-api-access-f7vm5\") pod \"certified-operators-xswbf\" (UID: \"a81d6f9c-e9f7-4c98-8233-c9f214978bfa\") " pod="openshift-marketplace/certified-operators-xswbf" Dec 01 08:42:06 crc kubenswrapper[4813]: I1201 08:42:06.460518 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mj64s\" (UID: \"a2b231fa-0ecc-44da-9259-a7dc7261127c\") " pod="openshift-image-registry/image-registry-697d97f7c8-mj64s" Dec 01 08:42:06 crc kubenswrapper[4813]: E1201 08:42:06.460944 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 08:42:06.960930922 +0000 UTC m=+111.583852508 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mj64s" (UID: "a2b231fa-0ecc-44da-9259-a7dc7261127c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:42:06 crc kubenswrapper[4813]: I1201 08:42:06.461587 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a81d6f9c-e9f7-4c98-8233-c9f214978bfa-utilities\") pod \"certified-operators-xswbf\" (UID: \"a81d6f9c-e9f7-4c98-8233-c9f214978bfa\") " pod="openshift-marketplace/certified-operators-xswbf" Dec 01 08:42:06 crc kubenswrapper[4813]: I1201 08:42:06.461869 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a81d6f9c-e9f7-4c98-8233-c9f214978bfa-catalog-content\") pod \"certified-operators-xswbf\" (UID: \"a81d6f9c-e9f7-4c98-8233-c9f214978bfa\") " pod="openshift-marketplace/certified-operators-xswbf" Dec 01 08:42:06 crc kubenswrapper[4813]: I1201 08:42:06.556845 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-xswbf"] Dec 01 08:42:06 crc kubenswrapper[4813]: I1201 08:42:06.568128 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 08:42:06 crc kubenswrapper[4813]: E1201 08:42:06.568435 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 08:42:07.068416458 +0000 UTC m=+111.691338044 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:42:06 crc kubenswrapper[4813]: I1201 08:42:06.642130 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ns6n2\" (UniqueName: \"kubernetes.io/projected/26f4804b-998f-488c-a501-d1cc4ef7f0b4-kube-api-access-ns6n2\") pod \"certified-operators-qh5hd\" (UID: \"26f4804b-998f-488c-a501-d1cc4ef7f0b4\") " pod="openshift-marketplace/certified-operators-qh5hd" Dec 01 08:42:06 crc kubenswrapper[4813]: I1201 08:42:06.642730 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-srz6m\" (UniqueName: \"kubernetes.io/projected/0959ee73-31e7-4d58-b124-3e971d777225-kube-api-access-srz6m\") pod \"community-operators-7fqxs\" (UID: \"0959ee73-31e7-4d58-b124-3e971d777225\") " pod="openshift-marketplace/community-operators-7fqxs" Dec 01 08:42:06 crc kubenswrapper[4813]: I1201 08:42:06.670787 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mj64s\" (UID: \"a2b231fa-0ecc-44da-9259-a7dc7261127c\") " pod="openshift-image-registry/image-registry-697d97f7c8-mj64s" Dec 01 08:42:06 crc kubenswrapper[4813]: E1201 08:42:06.671227 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 08:42:07.171211561 +0000 UTC m=+111.794133147 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mj64s" (UID: "a2b231fa-0ecc-44da-9259-a7dc7261127c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:42:06 crc kubenswrapper[4813]: I1201 08:42:06.682889 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f7vm5\" (UniqueName: \"kubernetes.io/projected/a81d6f9c-e9f7-4c98-8233-c9f214978bfa-kube-api-access-f7vm5\") pod \"certified-operators-xswbf\" (UID: \"a81d6f9c-e9f7-4c98-8233-c9f214978bfa\") " pod="openshift-marketplace/certified-operators-xswbf" Dec 01 08:42:06 crc kubenswrapper[4813]: I1201 08:42:06.690272 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-qh5hd" Dec 01 08:42:06 crc kubenswrapper[4813]: I1201 08:42:06.694355 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-8wwz4"] Dec 01 08:42:06 crc kubenswrapper[4813]: I1201 08:42:06.753595 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7fqxs" Dec 01 08:42:06 crc kubenswrapper[4813]: I1201 08:42:06.764786 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-8wwz4"] Dec 01 08:42:06 crc kubenswrapper[4813]: I1201 08:42:06.764938 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-8wwz4" Dec 01 08:42:06 crc kubenswrapper[4813]: I1201 08:42:06.771891 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 08:42:06 crc kubenswrapper[4813]: I1201 08:42:06.772083 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/95d1024c-5b73-40ae-9460-29502da48a23-utilities\") pod \"community-operators-8wwz4\" (UID: \"95d1024c-5b73-40ae-9460-29502da48a23\") " pod="openshift-marketplace/community-operators-8wwz4" Dec 01 08:42:06 crc kubenswrapper[4813]: I1201 08:42:06.772112 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xjp26\" (UniqueName: \"kubernetes.io/projected/95d1024c-5b73-40ae-9460-29502da48a23-kube-api-access-xjp26\") pod \"community-operators-8wwz4\" (UID: \"95d1024c-5b73-40ae-9460-29502da48a23\") " pod="openshift-marketplace/community-operators-8wwz4" Dec 01 08:42:06 crc kubenswrapper[4813]: I1201 08:42:06.772223 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/95d1024c-5b73-40ae-9460-29502da48a23-catalog-content\") pod \"community-operators-8wwz4\" (UID: \"95d1024c-5b73-40ae-9460-29502da48a23\") " pod="openshift-marketplace/community-operators-8wwz4" Dec 01 08:42:06 crc kubenswrapper[4813]: E1201 08:42:06.772309 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 08:42:07.272296042 +0000 UTC m=+111.895217628 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:42:06 crc kubenswrapper[4813]: I1201 08:42:06.806090 4813 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Dec 01 08:42:06 crc kubenswrapper[4813]: I1201 08:42:06.815273 4813 patch_prober.go:28] interesting pod/router-default-5444994796-sp4wh container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 01 08:42:06 crc kubenswrapper[4813]: [-]has-synced failed: reason withheld Dec 01 08:42:06 crc kubenswrapper[4813]: [+]process-running ok Dec 01 08:42:06 crc kubenswrapper[4813]: healthz check failed Dec 01 08:42:06 crc kubenswrapper[4813]: I1201 08:42:06.815348 4813 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-sp4wh" podUID="a905919c-6c7e-40be-8f43-b938a60f14f6" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 01 08:42:06 crc kubenswrapper[4813]: I1201 08:42:06.828594 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xswbf" Dec 01 08:42:07 crc kubenswrapper[4813]: I1201 08:42:07.055672 4813 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2025-12-01T08:42:06.80614999Z","Handler":null,"Name":""} Dec 01 08:42:07 crc kubenswrapper[4813]: I1201 08:42:07.057069 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/95d1024c-5b73-40ae-9460-29502da48a23-utilities\") pod \"community-operators-8wwz4\" (UID: \"95d1024c-5b73-40ae-9460-29502da48a23\") " pod="openshift-marketplace/community-operators-8wwz4" Dec 01 08:42:07 crc kubenswrapper[4813]: I1201 08:42:07.057092 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xjp26\" (UniqueName: \"kubernetes.io/projected/95d1024c-5b73-40ae-9460-29502da48a23-kube-api-access-xjp26\") pod \"community-operators-8wwz4\" (UID: \"95d1024c-5b73-40ae-9460-29502da48a23\") " pod="openshift-marketplace/community-operators-8wwz4" Dec 01 08:42:07 crc kubenswrapper[4813]: I1201 08:42:07.057134 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mj64s\" (UID: \"a2b231fa-0ecc-44da-9259-a7dc7261127c\") " pod="openshift-image-registry/image-registry-697d97f7c8-mj64s" Dec 01 08:42:07 crc kubenswrapper[4813]: I1201 08:42:07.057196 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/95d1024c-5b73-40ae-9460-29502da48a23-catalog-content\") pod \"community-operators-8wwz4\" (UID: \"95d1024c-5b73-40ae-9460-29502da48a23\") " pod="openshift-marketplace/community-operators-8wwz4" Dec 01 08:42:07 crc kubenswrapper[4813]: I1201 08:42:07.057593 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/95d1024c-5b73-40ae-9460-29502da48a23-catalog-content\") pod \"community-operators-8wwz4\" (UID: \"95d1024c-5b73-40ae-9460-29502da48a23\") " pod="openshift-marketplace/community-operators-8wwz4" Dec 01 08:42:07 crc kubenswrapper[4813]: I1201 08:42:07.057872 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/95d1024c-5b73-40ae-9460-29502da48a23-utilities\") pod \"community-operators-8wwz4\" (UID: \"95d1024c-5b73-40ae-9460-29502da48a23\") " pod="openshift-marketplace/community-operators-8wwz4" Dec 01 08:42:07 crc kubenswrapper[4813]: E1201 08:42:07.058275 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 08:42:07.558259382 +0000 UTC m=+112.181181008 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mj64s" (UID: "a2b231fa-0ecc-44da-9259-a7dc7261127c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 08:42:07 crc kubenswrapper[4813]: I1201 08:42:07.072884 4813 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Dec 01 08:42:07 crc kubenswrapper[4813]: I1201 08:42:07.072925 4813 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Dec 01 08:42:07 crc kubenswrapper[4813]: I1201 08:42:07.154934 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xjp26\" (UniqueName: \"kubernetes.io/projected/95d1024c-5b73-40ae-9460-29502da48a23-kube-api-access-xjp26\") pod \"community-operators-8wwz4\" (UID: \"95d1024c-5b73-40ae-9460-29502da48a23\") " pod="openshift-marketplace/community-operators-8wwz4" Dec 01 08:42:07 crc kubenswrapper[4813]: I1201 08:42:07.161573 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 08:42:07 crc kubenswrapper[4813]: I1201 08:42:07.206879 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Dec 01 08:42:07 crc kubenswrapper[4813]: I1201 08:42:07.263791 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mj64s\" (UID: \"a2b231fa-0ecc-44da-9259-a7dc7261127c\") " pod="openshift-image-registry/image-registry-697d97f7c8-mj64s" Dec 01 08:42:07 crc kubenswrapper[4813]: I1201 08:42:07.303933 4813 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 01 08:42:07 crc kubenswrapper[4813]: I1201 08:42:07.304000 4813 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mj64s\" (UID: \"a2b231fa-0ecc-44da-9259-a7dc7261127c\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-mj64s" Dec 01 08:42:07 crc kubenswrapper[4813]: I1201 08:42:07.334836 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"62fe4f70-1303-4787-ad08-24efee57578a","Type":"ContainerStarted","Data":"c57b4e1e3794dff692bbfcc730be22ddbe7b5c3092c8e769981e8ecc6e152769"} Dec 01 08:42:07 crc kubenswrapper[4813]: I1201 08:42:07.361115 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-qh5hd"] Dec 01 08:42:07 crc kubenswrapper[4813]: I1201 08:42:07.367221 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/revision-pruner-9-crc" podStartSLOduration=4.367196478 podStartE2EDuration="4.367196478s" podCreationTimestamp="2025-12-01 08:42:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:42:07.364938118 +0000 UTC m=+111.987859704" watchObservedRunningTime="2025-12-01 08:42:07.367196478 +0000 UTC m=+111.990118064" Dec 01 08:42:07 crc kubenswrapper[4813]: I1201 08:42:07.409015 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-8wwz4" Dec 01 08:42:07 crc kubenswrapper[4813]: I1201 08:42:07.498908 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mj64s\" (UID: \"a2b231fa-0ecc-44da-9259-a7dc7261127c\") " pod="openshift-image-registry/image-registry-697d97f7c8-mj64s" Dec 01 08:42:07 crc kubenswrapper[4813]: I1201 08:42:07.564329 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-7fqxs"] Dec 01 08:42:07 crc kubenswrapper[4813]: I1201 08:42:07.657261 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-xswbf"] Dec 01 08:42:07 crc kubenswrapper[4813]: I1201 08:42:07.790251 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-mj64s" Dec 01 08:42:07 crc kubenswrapper[4813]: I1201 08:42:07.819522 4813 patch_prober.go:28] interesting pod/router-default-5444994796-sp4wh container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 01 08:42:07 crc kubenswrapper[4813]: [-]has-synced failed: reason withheld Dec 01 08:42:07 crc kubenswrapper[4813]: [+]process-running ok Dec 01 08:42:07 crc kubenswrapper[4813]: healthz check failed Dec 01 08:42:07 crc kubenswrapper[4813]: I1201 08:42:07.819592 4813 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-sp4wh" podUID="a905919c-6c7e-40be-8f43-b938a60f14f6" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 01 08:42:07 crc kubenswrapper[4813]: I1201 08:42:07.838024 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-8wwz4"] Dec 01 08:42:07 crc kubenswrapper[4813]: E1201 08:42:07.844143 4813 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod26f4804b_998f_488c_a501_d1cc4ef7f0b4.slice/crio-a85ebdc73ea8bec3e97712277dd9d3aa025ade7c5c066f36896adca48b049aa8.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod26f4804b_998f_488c_a501_d1cc4ef7f0b4.slice/crio-conmon-a85ebdc73ea8bec3e97712277dd9d3aa025ade7c5c066f36896adca48b049aa8.scope\": RecentStats: unable to find data in memory cache]" Dec 01 08:42:07 crc kubenswrapper[4813]: I1201 08:42:07.993505 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-lk5db"] Dec 01 08:42:07 crc kubenswrapper[4813]: I1201 08:42:07.995464 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-lk5db" Dec 01 08:42:08 crc kubenswrapper[4813]: I1201 08:42:08.001758 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Dec 01 08:42:08 crc kubenswrapper[4813]: I1201 08:42:08.025899 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-lk5db"] Dec 01 08:42:08 crc kubenswrapper[4813]: I1201 08:42:08.077953 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4f1619ff-e7ed-42ee-b7a9-484c18d6677d-catalog-content\") pod \"redhat-marketplace-lk5db\" (UID: \"4f1619ff-e7ed-42ee-b7a9-484c18d6677d\") " pod="openshift-marketplace/redhat-marketplace-lk5db" Dec 01 08:42:08 crc kubenswrapper[4813]: I1201 08:42:08.078256 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vqnmj\" (UniqueName: \"kubernetes.io/projected/4f1619ff-e7ed-42ee-b7a9-484c18d6677d-kube-api-access-vqnmj\") pod \"redhat-marketplace-lk5db\" (UID: \"4f1619ff-e7ed-42ee-b7a9-484c18d6677d\") " pod="openshift-marketplace/redhat-marketplace-lk5db" Dec 01 08:42:08 crc kubenswrapper[4813]: I1201 08:42:08.078280 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4f1619ff-e7ed-42ee-b7a9-484c18d6677d-utilities\") pod \"redhat-marketplace-lk5db\" (UID: \"4f1619ff-e7ed-42ee-b7a9-484c18d6677d\") " pod="openshift-marketplace/redhat-marketplace-lk5db" Dec 01 08:42:08 crc kubenswrapper[4813]: I1201 08:42:08.175169 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-mj64s"] Dec 01 08:42:08 crc kubenswrapper[4813]: I1201 08:42:08.179298 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4f1619ff-e7ed-42ee-b7a9-484c18d6677d-utilities\") pod \"redhat-marketplace-lk5db\" (UID: \"4f1619ff-e7ed-42ee-b7a9-484c18d6677d\") " pod="openshift-marketplace/redhat-marketplace-lk5db" Dec 01 08:42:08 crc kubenswrapper[4813]: I1201 08:42:08.179427 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4f1619ff-e7ed-42ee-b7a9-484c18d6677d-catalog-content\") pod \"redhat-marketplace-lk5db\" (UID: \"4f1619ff-e7ed-42ee-b7a9-484c18d6677d\") " pod="openshift-marketplace/redhat-marketplace-lk5db" Dec 01 08:42:08 crc kubenswrapper[4813]: I1201 08:42:08.179483 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vqnmj\" (UniqueName: \"kubernetes.io/projected/4f1619ff-e7ed-42ee-b7a9-484c18d6677d-kube-api-access-vqnmj\") pod \"redhat-marketplace-lk5db\" (UID: \"4f1619ff-e7ed-42ee-b7a9-484c18d6677d\") " pod="openshift-marketplace/redhat-marketplace-lk5db" Dec 01 08:42:08 crc kubenswrapper[4813]: I1201 08:42:08.180247 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4f1619ff-e7ed-42ee-b7a9-484c18d6677d-utilities\") pod \"redhat-marketplace-lk5db\" (UID: \"4f1619ff-e7ed-42ee-b7a9-484c18d6677d\") " pod="openshift-marketplace/redhat-marketplace-lk5db" Dec 01 08:42:08 crc kubenswrapper[4813]: I1201 08:42:08.182047 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4f1619ff-e7ed-42ee-b7a9-484c18d6677d-catalog-content\") pod \"redhat-marketplace-lk5db\" (UID: \"4f1619ff-e7ed-42ee-b7a9-484c18d6677d\") " pod="openshift-marketplace/redhat-marketplace-lk5db" Dec 01 08:42:08 crc kubenswrapper[4813]: I1201 08:42:08.202584 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vqnmj\" (UniqueName: \"kubernetes.io/projected/4f1619ff-e7ed-42ee-b7a9-484c18d6677d-kube-api-access-vqnmj\") pod \"redhat-marketplace-lk5db\" (UID: \"4f1619ff-e7ed-42ee-b7a9-484c18d6677d\") " pod="openshift-marketplace/redhat-marketplace-lk5db" Dec 01 08:42:08 crc kubenswrapper[4813]: W1201 08:42:08.305025 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda2b231fa_0ecc_44da_9259_a7dc7261127c.slice/crio-824177aa29c11338724cc5b8be3c5968e74b376b652172baad8b598d76606b57 WatchSource:0}: Error finding container 824177aa29c11338724cc5b8be3c5968e74b376b652172baad8b598d76606b57: Status 404 returned error can't find the container with id 824177aa29c11338724cc5b8be3c5968e74b376b652172baad8b598d76606b57 Dec 01 08:42:08 crc kubenswrapper[4813]: I1201 08:42:08.326978 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-bpvbz"] Dec 01 08:42:08 crc kubenswrapper[4813]: I1201 08:42:08.328057 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bpvbz" Dec 01 08:42:08 crc kubenswrapper[4813]: I1201 08:42:08.328367 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-lk5db" Dec 01 08:42:08 crc kubenswrapper[4813]: I1201 08:42:08.349283 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-bpvbz"] Dec 01 08:42:08 crc kubenswrapper[4813]: I1201 08:42:08.352670 4813 generic.go:334] "Generic (PLEG): container finished" podID="95e0dc94-8aec-42d0-8b19-7cdf12fa7dee" containerID="f6d3eb682a8ac8a11cfdaff1cc058f21bb6b5cfef699c8f6fdf59a4c88c3f826" exitCode=0 Dec 01 08:42:08 crc kubenswrapper[4813]: I1201 08:42:08.352769 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29409630-45pn5" event={"ID":"95e0dc94-8aec-42d0-8b19-7cdf12fa7dee","Type":"ContainerDied","Data":"f6d3eb682a8ac8a11cfdaff1cc058f21bb6b5cfef699c8f6fdf59a4c88c3f826"} Dec 01 08:42:08 crc kubenswrapper[4813]: I1201 08:42:08.367424 4813 generic.go:334] "Generic (PLEG): container finished" podID="62fe4f70-1303-4787-ad08-24efee57578a" containerID="c57b4e1e3794dff692bbfcc730be22ddbe7b5c3092c8e769981e8ecc6e152769" exitCode=0 Dec 01 08:42:08 crc kubenswrapper[4813]: I1201 08:42:08.367671 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"62fe4f70-1303-4787-ad08-24efee57578a","Type":"ContainerDied","Data":"c57b4e1e3794dff692bbfcc730be22ddbe7b5c3092c8e769981e8ecc6e152769"} Dec 01 08:42:08 crc kubenswrapper[4813]: I1201 08:42:08.372863 4813 generic.go:334] "Generic (PLEG): container finished" podID="a81d6f9c-e9f7-4c98-8233-c9f214978bfa" containerID="8ca7da1c941e45e90ca434d5fd0b5e0fee687124784fb0f939a7fc7d231a558a" exitCode=0 Dec 01 08:42:08 crc kubenswrapper[4813]: I1201 08:42:08.372958 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xswbf" event={"ID":"a81d6f9c-e9f7-4c98-8233-c9f214978bfa","Type":"ContainerDied","Data":"8ca7da1c941e45e90ca434d5fd0b5e0fee687124784fb0f939a7fc7d231a558a"} Dec 01 08:42:08 crc kubenswrapper[4813]: I1201 08:42:08.373025 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xswbf" event={"ID":"a81d6f9c-e9f7-4c98-8233-c9f214978bfa","Type":"ContainerStarted","Data":"9a24ad031c98407f6e9a691b08ad6410f5e6ca12c71bca1033958e076c49a5fe"} Dec 01 08:42:08 crc kubenswrapper[4813]: I1201 08:42:08.375522 4813 generic.go:334] "Generic (PLEG): container finished" podID="0959ee73-31e7-4d58-b124-3e971d777225" containerID="9940458ef29891b048dfad33fda694cea93fa1876d4ca455069e938ffa765d3a" exitCode=0 Dec 01 08:42:08 crc kubenswrapper[4813]: I1201 08:42:08.375605 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7fqxs" event={"ID":"0959ee73-31e7-4d58-b124-3e971d777225","Type":"ContainerDied","Data":"9940458ef29891b048dfad33fda694cea93fa1876d4ca455069e938ffa765d3a"} Dec 01 08:42:08 crc kubenswrapper[4813]: I1201 08:42:08.375641 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7fqxs" event={"ID":"0959ee73-31e7-4d58-b124-3e971d777225","Type":"ContainerStarted","Data":"c3f740dd049aa2768738295f2186f891e43c9627bfba804b249e2eb6758d4a95"} Dec 01 08:42:08 crc kubenswrapper[4813]: I1201 08:42:08.376777 4813 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 01 08:42:08 crc kubenswrapper[4813]: I1201 08:42:08.381533 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-mj64s" event={"ID":"a2b231fa-0ecc-44da-9259-a7dc7261127c","Type":"ContainerStarted","Data":"824177aa29c11338724cc5b8be3c5968e74b376b652172baad8b598d76606b57"} Dec 01 08:42:08 crc kubenswrapper[4813]: I1201 08:42:08.384835 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tbrl7\" (UniqueName: \"kubernetes.io/projected/bd02bd16-7f86-48ad-b554-6404f62b1b28-kube-api-access-tbrl7\") pod \"redhat-marketplace-bpvbz\" (UID: \"bd02bd16-7f86-48ad-b554-6404f62b1b28\") " pod="openshift-marketplace/redhat-marketplace-bpvbz" Dec 01 08:42:08 crc kubenswrapper[4813]: I1201 08:42:08.385131 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bd02bd16-7f86-48ad-b554-6404f62b1b28-utilities\") pod \"redhat-marketplace-bpvbz\" (UID: \"bd02bd16-7f86-48ad-b554-6404f62b1b28\") " pod="openshift-marketplace/redhat-marketplace-bpvbz" Dec 01 08:42:08 crc kubenswrapper[4813]: I1201 08:42:08.385179 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bd02bd16-7f86-48ad-b554-6404f62b1b28-catalog-content\") pod \"redhat-marketplace-bpvbz\" (UID: \"bd02bd16-7f86-48ad-b554-6404f62b1b28\") " pod="openshift-marketplace/redhat-marketplace-bpvbz" Dec 01 08:42:08 crc kubenswrapper[4813]: I1201 08:42:08.394277 4813 generic.go:334] "Generic (PLEG): container finished" podID="26f4804b-998f-488c-a501-d1cc4ef7f0b4" containerID="a85ebdc73ea8bec3e97712277dd9d3aa025ade7c5c066f36896adca48b049aa8" exitCode=0 Dec 01 08:42:08 crc kubenswrapper[4813]: I1201 08:42:08.403580 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Dec 01 08:42:08 crc kubenswrapper[4813]: I1201 08:42:08.404402 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qh5hd" event={"ID":"26f4804b-998f-488c-a501-d1cc4ef7f0b4","Type":"ContainerDied","Data":"a85ebdc73ea8bec3e97712277dd9d3aa025ade7c5c066f36896adca48b049aa8"} Dec 01 08:42:08 crc kubenswrapper[4813]: I1201 08:42:08.404433 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qh5hd" event={"ID":"26f4804b-998f-488c-a501-d1cc4ef7f0b4","Type":"ContainerStarted","Data":"3b0157e891ec5eb8ab47e9c1ce21457c24e98cfae8c5c64214cfc597f5f1b146"} Dec 01 08:42:08 crc kubenswrapper[4813]: I1201 08:42:08.406072 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8wwz4" event={"ID":"95d1024c-5b73-40ae-9460-29502da48a23","Type":"ContainerStarted","Data":"efc0b55bd67f818bfb6a18d4781f9109c8f5b61e101e6974dc09069dab49127b"} Dec 01 08:42:08 crc kubenswrapper[4813]: I1201 08:42:08.486277 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bd02bd16-7f86-48ad-b554-6404f62b1b28-utilities\") pod \"redhat-marketplace-bpvbz\" (UID: \"bd02bd16-7f86-48ad-b554-6404f62b1b28\") " pod="openshift-marketplace/redhat-marketplace-bpvbz" Dec 01 08:42:08 crc kubenswrapper[4813]: I1201 08:42:08.486323 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bd02bd16-7f86-48ad-b554-6404f62b1b28-catalog-content\") pod \"redhat-marketplace-bpvbz\" (UID: \"bd02bd16-7f86-48ad-b554-6404f62b1b28\") " pod="openshift-marketplace/redhat-marketplace-bpvbz" Dec 01 08:42:08 crc kubenswrapper[4813]: I1201 08:42:08.486406 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tbrl7\" (UniqueName: \"kubernetes.io/projected/bd02bd16-7f86-48ad-b554-6404f62b1b28-kube-api-access-tbrl7\") pod \"redhat-marketplace-bpvbz\" (UID: \"bd02bd16-7f86-48ad-b554-6404f62b1b28\") " pod="openshift-marketplace/redhat-marketplace-bpvbz" Dec 01 08:42:08 crc kubenswrapper[4813]: I1201 08:42:08.489782 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bd02bd16-7f86-48ad-b554-6404f62b1b28-utilities\") pod \"redhat-marketplace-bpvbz\" (UID: \"bd02bd16-7f86-48ad-b554-6404f62b1b28\") " pod="openshift-marketplace/redhat-marketplace-bpvbz" Dec 01 08:42:08 crc kubenswrapper[4813]: I1201 08:42:08.489943 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bd02bd16-7f86-48ad-b554-6404f62b1b28-catalog-content\") pod \"redhat-marketplace-bpvbz\" (UID: \"bd02bd16-7f86-48ad-b554-6404f62b1b28\") " pod="openshift-marketplace/redhat-marketplace-bpvbz" Dec 01 08:42:08 crc kubenswrapper[4813]: I1201 08:42:08.510058 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tbrl7\" (UniqueName: \"kubernetes.io/projected/bd02bd16-7f86-48ad-b554-6404f62b1b28-kube-api-access-tbrl7\") pod \"redhat-marketplace-bpvbz\" (UID: \"bd02bd16-7f86-48ad-b554-6404f62b1b28\") " pod="openshift-marketplace/redhat-marketplace-bpvbz" Dec 01 08:42:08 crc kubenswrapper[4813]: I1201 08:42:08.652473 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bpvbz" Dec 01 08:42:08 crc kubenswrapper[4813]: I1201 08:42:08.834204 4813 patch_prober.go:28] interesting pod/router-default-5444994796-sp4wh container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 01 08:42:08 crc kubenswrapper[4813]: [-]has-synced failed: reason withheld Dec 01 08:42:08 crc kubenswrapper[4813]: [+]process-running ok Dec 01 08:42:08 crc kubenswrapper[4813]: healthz check failed Dec 01 08:42:08 crc kubenswrapper[4813]: I1201 08:42:08.834530 4813 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-sp4wh" podUID="a905919c-6c7e-40be-8f43-b938a60f14f6" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 01 08:42:08 crc kubenswrapper[4813]: I1201 08:42:08.924070 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-lk5db"] Dec 01 08:42:08 crc kubenswrapper[4813]: W1201 08:42:08.939197 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4f1619ff_e7ed_42ee_b7a9_484c18d6677d.slice/crio-d1fd115940029c23e9d5fd263b2f54473fa1d8fa42e3f02fb128dc0def171258 WatchSource:0}: Error finding container d1fd115940029c23e9d5fd263b2f54473fa1d8fa42e3f02fb128dc0def171258: Status 404 returned error can't find the container with id d1fd115940029c23e9d5fd263b2f54473fa1d8fa42e3f02fb128dc0def171258 Dec 01 08:42:09 crc kubenswrapper[4813]: I1201 08:42:09.102725 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-5ptsh" Dec 01 08:42:09 crc kubenswrapper[4813]: I1201 08:42:09.119345 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-5ptsh" Dec 01 08:42:09 crc kubenswrapper[4813]: I1201 08:42:09.161899 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-bpvbz"] Dec 01 08:42:09 crc kubenswrapper[4813]: I1201 08:42:09.332022 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-wbvcp"] Dec 01 08:42:09 crc kubenswrapper[4813]: I1201 08:42:09.336396 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wbvcp" Dec 01 08:42:09 crc kubenswrapper[4813]: I1201 08:42:09.338873 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Dec 01 08:42:09 crc kubenswrapper[4813]: I1201 08:42:09.345483 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-wbvcp"] Dec 01 08:42:09 crc kubenswrapper[4813]: I1201 08:42:09.409644 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4rzrc\" (UniqueName: \"kubernetes.io/projected/979ea876-4dd3-4d31-9144-7a9a6e914c67-kube-api-access-4rzrc\") pod \"redhat-operators-wbvcp\" (UID: \"979ea876-4dd3-4d31-9144-7a9a6e914c67\") " pod="openshift-marketplace/redhat-operators-wbvcp" Dec 01 08:42:09 crc kubenswrapper[4813]: I1201 08:42:09.409733 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/979ea876-4dd3-4d31-9144-7a9a6e914c67-utilities\") pod \"redhat-operators-wbvcp\" (UID: \"979ea876-4dd3-4d31-9144-7a9a6e914c67\") " pod="openshift-marketplace/redhat-operators-wbvcp" Dec 01 08:42:09 crc kubenswrapper[4813]: I1201 08:42:09.409827 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/979ea876-4dd3-4d31-9144-7a9a6e914c67-catalog-content\") pod \"redhat-operators-wbvcp\" (UID: \"979ea876-4dd3-4d31-9144-7a9a6e914c67\") " pod="openshift-marketplace/redhat-operators-wbvcp" Dec 01 08:42:09 crc kubenswrapper[4813]: I1201 08:42:09.447096 4813 generic.go:334] "Generic (PLEG): container finished" podID="4f1619ff-e7ed-42ee-b7a9-484c18d6677d" containerID="be2b96d2a5411e24ef73a4a32274a3a7cf486c30a3b857f32b237216a1a4b83f" exitCode=0 Dec 01 08:42:09 crc kubenswrapper[4813]: I1201 08:42:09.447222 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lk5db" event={"ID":"4f1619ff-e7ed-42ee-b7a9-484c18d6677d","Type":"ContainerDied","Data":"be2b96d2a5411e24ef73a4a32274a3a7cf486c30a3b857f32b237216a1a4b83f"} Dec 01 08:42:09 crc kubenswrapper[4813]: I1201 08:42:09.447260 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lk5db" event={"ID":"4f1619ff-e7ed-42ee-b7a9-484c18d6677d","Type":"ContainerStarted","Data":"d1fd115940029c23e9d5fd263b2f54473fa1d8fa42e3f02fb128dc0def171258"} Dec 01 08:42:09 crc kubenswrapper[4813]: I1201 08:42:09.470406 4813 generic.go:334] "Generic (PLEG): container finished" podID="95d1024c-5b73-40ae-9460-29502da48a23" containerID="ebe4667ac0b69195462fd75dcaf7de45a48aa09d477a4f482dca8336cf604cf5" exitCode=0 Dec 01 08:42:09 crc kubenswrapper[4813]: I1201 08:42:09.470489 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8wwz4" event={"ID":"95d1024c-5b73-40ae-9460-29502da48a23","Type":"ContainerDied","Data":"ebe4667ac0b69195462fd75dcaf7de45a48aa09d477a4f482dca8336cf604cf5"} Dec 01 08:42:09 crc kubenswrapper[4813]: I1201 08:42:09.489406 4813 generic.go:334] "Generic (PLEG): container finished" podID="bd02bd16-7f86-48ad-b554-6404f62b1b28" containerID="e30f802ba7038b957dde0fd7ec61e4e78c5a773cb92cae1974cff027d52e3a2b" exitCode=0 Dec 01 08:42:09 crc kubenswrapper[4813]: I1201 08:42:09.490159 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bpvbz" event={"ID":"bd02bd16-7f86-48ad-b554-6404f62b1b28","Type":"ContainerDied","Data":"e30f802ba7038b957dde0fd7ec61e4e78c5a773cb92cae1974cff027d52e3a2b"} Dec 01 08:42:09 crc kubenswrapper[4813]: I1201 08:42:09.490192 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bpvbz" event={"ID":"bd02bd16-7f86-48ad-b554-6404f62b1b28","Type":"ContainerStarted","Data":"e99621bb15483390dda6f16ddb39f8c2ff88ef51507adc0def12ac33409b4e20"} Dec 01 08:42:09 crc kubenswrapper[4813]: I1201 08:42:09.498220 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-mj64s" event={"ID":"a2b231fa-0ecc-44da-9259-a7dc7261127c","Type":"ContainerStarted","Data":"ac60085c4717fdae779b48e0b9b874818666761bf46e67459bdf3c2cd59f37de"} Dec 01 08:42:09 crc kubenswrapper[4813]: I1201 08:42:09.498361 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-mj64s" Dec 01 08:42:09 crc kubenswrapper[4813]: I1201 08:42:09.513024 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/979ea876-4dd3-4d31-9144-7a9a6e914c67-catalog-content\") pod \"redhat-operators-wbvcp\" (UID: \"979ea876-4dd3-4d31-9144-7a9a6e914c67\") " pod="openshift-marketplace/redhat-operators-wbvcp" Dec 01 08:42:09 crc kubenswrapper[4813]: I1201 08:42:09.513105 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4rzrc\" (UniqueName: \"kubernetes.io/projected/979ea876-4dd3-4d31-9144-7a9a6e914c67-kube-api-access-4rzrc\") pod \"redhat-operators-wbvcp\" (UID: \"979ea876-4dd3-4d31-9144-7a9a6e914c67\") " pod="openshift-marketplace/redhat-operators-wbvcp" Dec 01 08:42:09 crc kubenswrapper[4813]: I1201 08:42:09.513269 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/979ea876-4dd3-4d31-9144-7a9a6e914c67-utilities\") pod \"redhat-operators-wbvcp\" (UID: \"979ea876-4dd3-4d31-9144-7a9a6e914c67\") " pod="openshift-marketplace/redhat-operators-wbvcp" Dec 01 08:42:09 crc kubenswrapper[4813]: I1201 08:42:09.513730 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/979ea876-4dd3-4d31-9144-7a9a6e914c67-catalog-content\") pod \"redhat-operators-wbvcp\" (UID: \"979ea876-4dd3-4d31-9144-7a9a6e914c67\") " pod="openshift-marketplace/redhat-operators-wbvcp" Dec 01 08:42:09 crc kubenswrapper[4813]: I1201 08:42:09.514054 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/979ea876-4dd3-4d31-9144-7a9a6e914c67-utilities\") pod \"redhat-operators-wbvcp\" (UID: \"979ea876-4dd3-4d31-9144-7a9a6e914c67\") " pod="openshift-marketplace/redhat-operators-wbvcp" Dec 01 08:42:09 crc kubenswrapper[4813]: I1201 08:42:09.536802 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4rzrc\" (UniqueName: \"kubernetes.io/projected/979ea876-4dd3-4d31-9144-7a9a6e914c67-kube-api-access-4rzrc\") pod \"redhat-operators-wbvcp\" (UID: \"979ea876-4dd3-4d31-9144-7a9a6e914c67\") " pod="openshift-marketplace/redhat-operators-wbvcp" Dec 01 08:42:09 crc kubenswrapper[4813]: I1201 08:42:09.550438 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-mj64s" podStartSLOduration=89.550408357 podStartE2EDuration="1m29.550408357s" podCreationTimestamp="2025-12-01 08:40:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:42:09.545836627 +0000 UTC m=+114.168758213" watchObservedRunningTime="2025-12-01 08:42:09.550408357 +0000 UTC m=+114.173329943" Dec 01 08:42:09 crc kubenswrapper[4813]: I1201 08:42:09.660363 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wbvcp" Dec 01 08:42:09 crc kubenswrapper[4813]: I1201 08:42:09.734919 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-58jst"] Dec 01 08:42:09 crc kubenswrapper[4813]: I1201 08:42:09.736462 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-58jst" Dec 01 08:42:09 crc kubenswrapper[4813]: I1201 08:42:09.746350 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-58jst"] Dec 01 08:42:09 crc kubenswrapper[4813]: I1201 08:42:09.817846 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3c0b52aa-86ba-4c56-b23e-108cdb4750d9-utilities\") pod \"redhat-operators-58jst\" (UID: \"3c0b52aa-86ba-4c56-b23e-108cdb4750d9\") " pod="openshift-marketplace/redhat-operators-58jst" Dec 01 08:42:09 crc kubenswrapper[4813]: I1201 08:42:09.817885 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3c0b52aa-86ba-4c56-b23e-108cdb4750d9-catalog-content\") pod \"redhat-operators-58jst\" (UID: \"3c0b52aa-86ba-4c56-b23e-108cdb4750d9\") " pod="openshift-marketplace/redhat-operators-58jst" Dec 01 08:42:09 crc kubenswrapper[4813]: I1201 08:42:09.817948 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qbk44\" (UniqueName: \"kubernetes.io/projected/3c0b52aa-86ba-4c56-b23e-108cdb4750d9-kube-api-access-qbk44\") pod \"redhat-operators-58jst\" (UID: \"3c0b52aa-86ba-4c56-b23e-108cdb4750d9\") " pod="openshift-marketplace/redhat-operators-58jst" Dec 01 08:42:09 crc kubenswrapper[4813]: I1201 08:42:09.822196 4813 patch_prober.go:28] interesting pod/router-default-5444994796-sp4wh container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 01 08:42:09 crc kubenswrapper[4813]: [-]has-synced failed: reason withheld Dec 01 08:42:09 crc kubenswrapper[4813]: [+]process-running ok Dec 01 08:42:09 crc kubenswrapper[4813]: healthz check failed Dec 01 08:42:09 crc kubenswrapper[4813]: I1201 08:42:09.822267 4813 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-sp4wh" podUID="a905919c-6c7e-40be-8f43-b938a60f14f6" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 01 08:42:09 crc kubenswrapper[4813]: I1201 08:42:09.934506 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3c0b52aa-86ba-4c56-b23e-108cdb4750d9-utilities\") pod \"redhat-operators-58jst\" (UID: \"3c0b52aa-86ba-4c56-b23e-108cdb4750d9\") " pod="openshift-marketplace/redhat-operators-58jst" Dec 01 08:42:09 crc kubenswrapper[4813]: I1201 08:42:09.933817 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3c0b52aa-86ba-4c56-b23e-108cdb4750d9-utilities\") pod \"redhat-operators-58jst\" (UID: \"3c0b52aa-86ba-4c56-b23e-108cdb4750d9\") " pod="openshift-marketplace/redhat-operators-58jst" Dec 01 08:42:09 crc kubenswrapper[4813]: I1201 08:42:09.934990 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3c0b52aa-86ba-4c56-b23e-108cdb4750d9-catalog-content\") pod \"redhat-operators-58jst\" (UID: \"3c0b52aa-86ba-4c56-b23e-108cdb4750d9\") " pod="openshift-marketplace/redhat-operators-58jst" Dec 01 08:42:09 crc kubenswrapper[4813]: I1201 08:42:09.935330 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3c0b52aa-86ba-4c56-b23e-108cdb4750d9-catalog-content\") pod \"redhat-operators-58jst\" (UID: \"3c0b52aa-86ba-4c56-b23e-108cdb4750d9\") " pod="openshift-marketplace/redhat-operators-58jst" Dec 01 08:42:09 crc kubenswrapper[4813]: I1201 08:42:09.939189 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qbk44\" (UniqueName: \"kubernetes.io/projected/3c0b52aa-86ba-4c56-b23e-108cdb4750d9-kube-api-access-qbk44\") pod \"redhat-operators-58jst\" (UID: \"3c0b52aa-86ba-4c56-b23e-108cdb4750d9\") " pod="openshift-marketplace/redhat-operators-58jst" Dec 01 08:42:09 crc kubenswrapper[4813]: I1201 08:42:09.984982 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qbk44\" (UniqueName: \"kubernetes.io/projected/3c0b52aa-86ba-4c56-b23e-108cdb4750d9-kube-api-access-qbk44\") pod \"redhat-operators-58jst\" (UID: \"3c0b52aa-86ba-4c56-b23e-108cdb4750d9\") " pod="openshift-marketplace/redhat-operators-58jst" Dec 01 08:42:10 crc kubenswrapper[4813]: I1201 08:42:10.106965 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-58jst" Dec 01 08:42:10 crc kubenswrapper[4813]: I1201 08:42:10.447115 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 01 08:42:10 crc kubenswrapper[4813]: I1201 08:42:10.479020 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-wbvcp"] Dec 01 08:42:10 crc kubenswrapper[4813]: I1201 08:42:10.512512 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29409630-45pn5" Dec 01 08:42:10 crc kubenswrapper[4813]: W1201 08:42:10.516105 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod979ea876_4dd3_4d31_9144_7a9a6e914c67.slice/crio-5cc0c87f16f0f1eb7a9a1dcd16afcdb84d301f455a3ac440e2afa8b17264b9aa WatchSource:0}: Error finding container 5cc0c87f16f0f1eb7a9a1dcd16afcdb84d301f455a3ac440e2afa8b17264b9aa: Status 404 returned error can't find the container with id 5cc0c87f16f0f1eb7a9a1dcd16afcdb84d301f455a3ac440e2afa8b17264b9aa Dec 01 08:42:10 crc kubenswrapper[4813]: I1201 08:42:10.517440 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29409630-45pn5" event={"ID":"95e0dc94-8aec-42d0-8b19-7cdf12fa7dee","Type":"ContainerDied","Data":"97c05336981e5cd1c1cddcc9c2fcc2d3a3c658e8a05351e240d02255ceaa7081"} Dec 01 08:42:10 crc kubenswrapper[4813]: I1201 08:42:10.517508 4813 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="97c05336981e5cd1c1cddcc9c2fcc2d3a3c658e8a05351e240d02255ceaa7081" Dec 01 08:42:10 crc kubenswrapper[4813]: I1201 08:42:10.525466 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"62fe4f70-1303-4787-ad08-24efee57578a","Type":"ContainerDied","Data":"e932ab3f01d5c0c15049c134b104a28d6cd817157803a385385fcf74b37f5f24"} Dec 01 08:42:10 crc kubenswrapper[4813]: I1201 08:42:10.525574 4813 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e932ab3f01d5c0c15049c134b104a28d6cd817157803a385385fcf74b37f5f24" Dec 01 08:42:10 crc kubenswrapper[4813]: I1201 08:42:10.525575 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 01 08:42:10 crc kubenswrapper[4813]: I1201 08:42:10.714120 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/62fe4f70-1303-4787-ad08-24efee57578a-kubelet-dir\") pod \"62fe4f70-1303-4787-ad08-24efee57578a\" (UID: \"62fe4f70-1303-4787-ad08-24efee57578a\") " Dec 01 08:42:10 crc kubenswrapper[4813]: I1201 08:42:10.714325 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/62fe4f70-1303-4787-ad08-24efee57578a-kube-api-access\") pod \"62fe4f70-1303-4787-ad08-24efee57578a\" (UID: \"62fe4f70-1303-4787-ad08-24efee57578a\") " Dec 01 08:42:10 crc kubenswrapper[4813]: I1201 08:42:10.714364 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/62fe4f70-1303-4787-ad08-24efee57578a-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "62fe4f70-1303-4787-ad08-24efee57578a" (UID: "62fe4f70-1303-4787-ad08-24efee57578a"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 08:42:10 crc kubenswrapper[4813]: I1201 08:42:10.714397 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gjfw8\" (UniqueName: \"kubernetes.io/projected/95e0dc94-8aec-42d0-8b19-7cdf12fa7dee-kube-api-access-gjfw8\") pod \"95e0dc94-8aec-42d0-8b19-7cdf12fa7dee\" (UID: \"95e0dc94-8aec-42d0-8b19-7cdf12fa7dee\") " Dec 01 08:42:10 crc kubenswrapper[4813]: I1201 08:42:10.714476 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/95e0dc94-8aec-42d0-8b19-7cdf12fa7dee-secret-volume\") pod \"95e0dc94-8aec-42d0-8b19-7cdf12fa7dee\" (UID: \"95e0dc94-8aec-42d0-8b19-7cdf12fa7dee\") " Dec 01 08:42:10 crc kubenswrapper[4813]: I1201 08:42:10.714554 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/95e0dc94-8aec-42d0-8b19-7cdf12fa7dee-config-volume\") pod \"95e0dc94-8aec-42d0-8b19-7cdf12fa7dee\" (UID: \"95e0dc94-8aec-42d0-8b19-7cdf12fa7dee\") " Dec 01 08:42:10 crc kubenswrapper[4813]: I1201 08:42:10.716809 4813 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/62fe4f70-1303-4787-ad08-24efee57578a-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 01 08:42:10 crc kubenswrapper[4813]: I1201 08:42:10.719891 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/95e0dc94-8aec-42d0-8b19-7cdf12fa7dee-config-volume" (OuterVolumeSpecName: "config-volume") pod "95e0dc94-8aec-42d0-8b19-7cdf12fa7dee" (UID: "95e0dc94-8aec-42d0-8b19-7cdf12fa7dee"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:42:10 crc kubenswrapper[4813]: I1201 08:42:10.723941 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/62fe4f70-1303-4787-ad08-24efee57578a-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "62fe4f70-1303-4787-ad08-24efee57578a" (UID: "62fe4f70-1303-4787-ad08-24efee57578a"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:42:10 crc kubenswrapper[4813]: I1201 08:42:10.725759 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/95e0dc94-8aec-42d0-8b19-7cdf12fa7dee-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "95e0dc94-8aec-42d0-8b19-7cdf12fa7dee" (UID: "95e0dc94-8aec-42d0-8b19-7cdf12fa7dee"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:42:10 crc kubenswrapper[4813]: I1201 08:42:10.725809 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/95e0dc94-8aec-42d0-8b19-7cdf12fa7dee-kube-api-access-gjfw8" (OuterVolumeSpecName: "kube-api-access-gjfw8") pod "95e0dc94-8aec-42d0-8b19-7cdf12fa7dee" (UID: "95e0dc94-8aec-42d0-8b19-7cdf12fa7dee"). InnerVolumeSpecName "kube-api-access-gjfw8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:42:10 crc kubenswrapper[4813]: I1201 08:42:10.768218 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-58jst"] Dec 01 08:42:10 crc kubenswrapper[4813]: W1201 08:42:10.777550 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3c0b52aa_86ba_4c56_b23e_108cdb4750d9.slice/crio-ae96f422c1483c19efa4aa9d6d31da5bb7d236ffd7a4c817cf236212a4243543 WatchSource:0}: Error finding container ae96f422c1483c19efa4aa9d6d31da5bb7d236ffd7a4c817cf236212a4243543: Status 404 returned error can't find the container with id ae96f422c1483c19efa4aa9d6d31da5bb7d236ffd7a4c817cf236212a4243543 Dec 01 08:42:10 crc kubenswrapper[4813]: I1201 08:42:10.817802 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/62fe4f70-1303-4787-ad08-24efee57578a-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 01 08:42:10 crc kubenswrapper[4813]: I1201 08:42:10.818066 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gjfw8\" (UniqueName: \"kubernetes.io/projected/95e0dc94-8aec-42d0-8b19-7cdf12fa7dee-kube-api-access-gjfw8\") on node \"crc\" DevicePath \"\"" Dec 01 08:42:10 crc kubenswrapper[4813]: I1201 08:42:10.818076 4813 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/95e0dc94-8aec-42d0-8b19-7cdf12fa7dee-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 01 08:42:10 crc kubenswrapper[4813]: I1201 08:42:10.818084 4813 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/95e0dc94-8aec-42d0-8b19-7cdf12fa7dee-config-volume\") on node \"crc\" DevicePath \"\"" Dec 01 08:42:10 crc kubenswrapper[4813]: I1201 08:42:10.818812 4813 patch_prober.go:28] interesting pod/router-default-5444994796-sp4wh container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 01 08:42:10 crc kubenswrapper[4813]: [-]has-synced failed: reason withheld Dec 01 08:42:10 crc kubenswrapper[4813]: [+]process-running ok Dec 01 08:42:10 crc kubenswrapper[4813]: healthz check failed Dec 01 08:42:10 crc kubenswrapper[4813]: I1201 08:42:10.818876 4813 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-sp4wh" podUID="a905919c-6c7e-40be-8f43-b938a60f14f6" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 01 08:42:11 crc kubenswrapper[4813]: I1201 08:42:11.690372 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-58jst" event={"ID":"3c0b52aa-86ba-4c56-b23e-108cdb4750d9","Type":"ContainerStarted","Data":"d9ea5b6b0d37675df39b7b61235961ddc4686468313e264a4c8271242376f87c"} Dec 01 08:42:11 crc kubenswrapper[4813]: I1201 08:42:11.690503 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-58jst" event={"ID":"3c0b52aa-86ba-4c56-b23e-108cdb4750d9","Type":"ContainerStarted","Data":"ae96f422c1483c19efa4aa9d6d31da5bb7d236ffd7a4c817cf236212a4243543"} Dec 01 08:42:11 crc kubenswrapper[4813]: I1201 08:42:11.695919 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Dec 01 08:42:11 crc kubenswrapper[4813]: E1201 08:42:11.696473 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="95e0dc94-8aec-42d0-8b19-7cdf12fa7dee" containerName="collect-profiles" Dec 01 08:42:11 crc kubenswrapper[4813]: I1201 08:42:11.696515 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="95e0dc94-8aec-42d0-8b19-7cdf12fa7dee" containerName="collect-profiles" Dec 01 08:42:11 crc kubenswrapper[4813]: E1201 08:42:11.696546 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="62fe4f70-1303-4787-ad08-24efee57578a" containerName="pruner" Dec 01 08:42:11 crc kubenswrapper[4813]: I1201 08:42:11.696556 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="62fe4f70-1303-4787-ad08-24efee57578a" containerName="pruner" Dec 01 08:42:11 crc kubenswrapper[4813]: I1201 08:42:11.696805 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="95e0dc94-8aec-42d0-8b19-7cdf12fa7dee" containerName="collect-profiles" Dec 01 08:42:11 crc kubenswrapper[4813]: I1201 08:42:11.696845 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="62fe4f70-1303-4787-ad08-24efee57578a" containerName="pruner" Dec 01 08:42:11 crc kubenswrapper[4813]: I1201 08:42:11.710179 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 01 08:42:11 crc kubenswrapper[4813]: I1201 08:42:11.710278 4813 generic.go:334] "Generic (PLEG): container finished" podID="979ea876-4dd3-4d31-9144-7a9a6e914c67" containerID="b59544951bce53f8c68dee91538682a664b7d5825a97dc0b5c017c5156d731cf" exitCode=0 Dec 01 08:42:11 crc kubenswrapper[4813]: I1201 08:42:11.710830 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29409630-45pn5" Dec 01 08:42:11 crc kubenswrapper[4813]: I1201 08:42:11.710412 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wbvcp" event={"ID":"979ea876-4dd3-4d31-9144-7a9a6e914c67","Type":"ContainerDied","Data":"b59544951bce53f8c68dee91538682a664b7d5825a97dc0b5c017c5156d731cf"} Dec 01 08:42:11 crc kubenswrapper[4813]: I1201 08:42:11.711514 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wbvcp" event={"ID":"979ea876-4dd3-4d31-9144-7a9a6e914c67","Type":"ContainerStarted","Data":"5cc0c87f16f0f1eb7a9a1dcd16afcdb84d301f455a3ac440e2afa8b17264b9aa"} Dec 01 08:42:11 crc kubenswrapper[4813]: I1201 08:42:11.712414 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Dec 01 08:42:11 crc kubenswrapper[4813]: I1201 08:42:11.713042 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Dec 01 08:42:11 crc kubenswrapper[4813]: I1201 08:42:11.713659 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Dec 01 08:42:11 crc kubenswrapper[4813]: I1201 08:42:11.814195 4813 patch_prober.go:28] interesting pod/router-default-5444994796-sp4wh container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 01 08:42:11 crc kubenswrapper[4813]: [-]has-synced failed: reason withheld Dec 01 08:42:11 crc kubenswrapper[4813]: [+]process-running ok Dec 01 08:42:11 crc kubenswrapper[4813]: healthz check failed Dec 01 08:42:11 crc kubenswrapper[4813]: I1201 08:42:11.814300 4813 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-sp4wh" podUID="a905919c-6c7e-40be-8f43-b938a60f14f6" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 01 08:42:11 crc kubenswrapper[4813]: I1201 08:42:11.898867 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/0f66717e-8cd9-4f69-9037-db019002c8a1-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"0f66717e-8cd9-4f69-9037-db019002c8a1\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 01 08:42:11 crc kubenswrapper[4813]: I1201 08:42:11.898956 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0f66717e-8cd9-4f69-9037-db019002c8a1-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"0f66717e-8cd9-4f69-9037-db019002c8a1\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 01 08:42:12 crc kubenswrapper[4813]: I1201 08:42:12.148822 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/0f66717e-8cd9-4f69-9037-db019002c8a1-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"0f66717e-8cd9-4f69-9037-db019002c8a1\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 01 08:42:12 crc kubenswrapper[4813]: I1201 08:42:12.148952 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0f66717e-8cd9-4f69-9037-db019002c8a1-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"0f66717e-8cd9-4f69-9037-db019002c8a1\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 01 08:42:12 crc kubenswrapper[4813]: I1201 08:42:12.150366 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/0f66717e-8cd9-4f69-9037-db019002c8a1-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"0f66717e-8cd9-4f69-9037-db019002c8a1\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 01 08:42:12 crc kubenswrapper[4813]: I1201 08:42:12.177922 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0f66717e-8cd9-4f69-9037-db019002c8a1-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"0f66717e-8cd9-4f69-9037-db019002c8a1\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 01 08:42:12 crc kubenswrapper[4813]: I1201 08:42:12.349779 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 01 08:42:12 crc kubenswrapper[4813]: I1201 08:42:12.849854 4813 patch_prober.go:28] interesting pod/router-default-5444994796-sp4wh container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 01 08:42:12 crc kubenswrapper[4813]: [-]has-synced failed: reason withheld Dec 01 08:42:12 crc kubenswrapper[4813]: [+]process-running ok Dec 01 08:42:12 crc kubenswrapper[4813]: healthz check failed Dec 01 08:42:12 crc kubenswrapper[4813]: I1201 08:42:12.850153 4813 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-sp4wh" podUID="a905919c-6c7e-40be-8f43-b938a60f14f6" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 01 08:42:12 crc kubenswrapper[4813]: I1201 08:42:12.884264 4813 generic.go:334] "Generic (PLEG): container finished" podID="3c0b52aa-86ba-4c56-b23e-108cdb4750d9" containerID="d9ea5b6b0d37675df39b7b61235961ddc4686468313e264a4c8271242376f87c" exitCode=0 Dec 01 08:42:12 crc kubenswrapper[4813]: I1201 08:42:12.884317 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-58jst" event={"ID":"3c0b52aa-86ba-4c56-b23e-108cdb4750d9","Type":"ContainerDied","Data":"d9ea5b6b0d37675df39b7b61235961ddc4686468313e264a4c8271242376f87c"} Dec 01 08:42:12 crc kubenswrapper[4813]: I1201 08:42:12.962330 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Dec 01 08:42:12 crc kubenswrapper[4813]: W1201 08:42:12.975005 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod0f66717e_8cd9_4f69_9037_db019002c8a1.slice/crio-4ab8be9ba54af1859f607786e5ece9cc2bf89989468c1bfd69b508aec431e747 WatchSource:0}: Error finding container 4ab8be9ba54af1859f607786e5ece9cc2bf89989468c1bfd69b508aec431e747: Status 404 returned error can't find the container with id 4ab8be9ba54af1859f607786e5ece9cc2bf89989468c1bfd69b508aec431e747 Dec 01 08:42:13 crc kubenswrapper[4813]: I1201 08:42:13.712987 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-j7v65" Dec 01 08:42:13 crc kubenswrapper[4813]: I1201 08:42:13.786142 4813 patch_prober.go:28] interesting pod/console-f9d7485db-9ql8j container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.10:8443/health\": dial tcp 10.217.0.10:8443: connect: connection refused" start-of-body= Dec 01 08:42:13 crc kubenswrapper[4813]: I1201 08:42:13.786223 4813 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-9ql8j" podUID="65bbc5b9-f6f0-40ad-9e47-4122bb5cabab" containerName="console" probeResult="failure" output="Get \"https://10.217.0.10:8443/health\": dial tcp 10.217.0.10:8443: connect: connection refused" Dec 01 08:42:13 crc kubenswrapper[4813]: I1201 08:42:13.813580 4813 patch_prober.go:28] interesting pod/router-default-5444994796-sp4wh container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 01 08:42:13 crc kubenswrapper[4813]: [-]has-synced failed: reason withheld Dec 01 08:42:13 crc kubenswrapper[4813]: [+]process-running ok Dec 01 08:42:13 crc kubenswrapper[4813]: healthz check failed Dec 01 08:42:13 crc kubenswrapper[4813]: I1201 08:42:13.813682 4813 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-sp4wh" podUID="a905919c-6c7e-40be-8f43-b938a60f14f6" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 01 08:42:13 crc kubenswrapper[4813]: I1201 08:42:13.925156 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"0f66717e-8cd9-4f69-9037-db019002c8a1","Type":"ContainerStarted","Data":"cff15c8c21e415126e4f0be4affd745a61934a7cc2ad1a50ef4f55bf3bdf4262"} Dec 01 08:42:13 crc kubenswrapper[4813]: I1201 08:42:13.925248 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"0f66717e-8cd9-4f69-9037-db019002c8a1","Type":"ContainerStarted","Data":"4ab8be9ba54af1859f607786e5ece9cc2bf89989468c1bfd69b508aec431e747"} Dec 01 08:42:13 crc kubenswrapper[4813]: I1201 08:42:13.997688 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-8-crc" podStartSLOduration=2.997645257 podStartE2EDuration="2.997645257s" podCreationTimestamp="2025-12-01 08:42:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:42:13.994645885 +0000 UTC m=+118.617567471" watchObservedRunningTime="2025-12-01 08:42:13.997645257 +0000 UTC m=+118.620566843" Dec 01 08:42:14 crc kubenswrapper[4813]: I1201 08:42:14.171383 4813 patch_prober.go:28] interesting pod/downloads-7954f5f757-ct7w8 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" start-of-body= Dec 01 08:42:14 crc kubenswrapper[4813]: I1201 08:42:14.171453 4813 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-ct7w8" podUID="e17c11d6-eb41-4391-b513-a817eeffcdb4" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" Dec 01 08:42:14 crc kubenswrapper[4813]: I1201 08:42:14.171904 4813 patch_prober.go:28] interesting pod/downloads-7954f5f757-ct7w8 container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" start-of-body= Dec 01 08:42:14 crc kubenswrapper[4813]: I1201 08:42:14.171933 4813 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-ct7w8" podUID="e17c11d6-eb41-4391-b513-a817eeffcdb4" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" Dec 01 08:42:14 crc kubenswrapper[4813]: I1201 08:42:14.820393 4813 patch_prober.go:28] interesting pod/router-default-5444994796-sp4wh container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 01 08:42:14 crc kubenswrapper[4813]: [-]has-synced failed: reason withheld Dec 01 08:42:14 crc kubenswrapper[4813]: [+]process-running ok Dec 01 08:42:14 crc kubenswrapper[4813]: healthz check failed Dec 01 08:42:14 crc kubenswrapper[4813]: I1201 08:42:14.820486 4813 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-sp4wh" podUID="a905919c-6c7e-40be-8f43-b938a60f14f6" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 01 08:42:14 crc kubenswrapper[4813]: I1201 08:42:14.936635 4813 generic.go:334] "Generic (PLEG): container finished" podID="0f66717e-8cd9-4f69-9037-db019002c8a1" containerID="cff15c8c21e415126e4f0be4affd745a61934a7cc2ad1a50ef4f55bf3bdf4262" exitCode=0 Dec 01 08:42:14 crc kubenswrapper[4813]: I1201 08:42:14.936685 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"0f66717e-8cd9-4f69-9037-db019002c8a1","Type":"ContainerDied","Data":"cff15c8c21e415126e4f0be4affd745a61934a7cc2ad1a50ef4f55bf3bdf4262"} Dec 01 08:42:15 crc kubenswrapper[4813]: I1201 08:42:15.817021 4813 patch_prober.go:28] interesting pod/router-default-5444994796-sp4wh container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 01 08:42:15 crc kubenswrapper[4813]: [-]has-synced failed: reason withheld Dec 01 08:42:15 crc kubenswrapper[4813]: [+]process-running ok Dec 01 08:42:15 crc kubenswrapper[4813]: healthz check failed Dec 01 08:42:15 crc kubenswrapper[4813]: I1201 08:42:15.817101 4813 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-sp4wh" podUID="a905919c-6c7e-40be-8f43-b938a60f14f6" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 01 08:42:15 crc kubenswrapper[4813]: I1201 08:42:15.824585 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-qsskc" Dec 01 08:42:16 crc kubenswrapper[4813]: I1201 08:42:16.409267 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 01 08:42:16 crc kubenswrapper[4813]: I1201 08:42:16.528189 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0f66717e-8cd9-4f69-9037-db019002c8a1-kube-api-access\") pod \"0f66717e-8cd9-4f69-9037-db019002c8a1\" (UID: \"0f66717e-8cd9-4f69-9037-db019002c8a1\") " Dec 01 08:42:16 crc kubenswrapper[4813]: I1201 08:42:16.528280 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/0f66717e-8cd9-4f69-9037-db019002c8a1-kubelet-dir\") pod \"0f66717e-8cd9-4f69-9037-db019002c8a1\" (UID: \"0f66717e-8cd9-4f69-9037-db019002c8a1\") " Dec 01 08:42:16 crc kubenswrapper[4813]: I1201 08:42:16.528448 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0f66717e-8cd9-4f69-9037-db019002c8a1-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "0f66717e-8cd9-4f69-9037-db019002c8a1" (UID: "0f66717e-8cd9-4f69-9037-db019002c8a1"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 08:42:16 crc kubenswrapper[4813]: I1201 08:42:16.528672 4813 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/0f66717e-8cd9-4f69-9037-db019002c8a1-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 01 08:42:16 crc kubenswrapper[4813]: I1201 08:42:16.536744 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0f66717e-8cd9-4f69-9037-db019002c8a1-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0f66717e-8cd9-4f69-9037-db019002c8a1" (UID: "0f66717e-8cd9-4f69-9037-db019002c8a1"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:42:16 crc kubenswrapper[4813]: I1201 08:42:16.630457 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0f66717e-8cd9-4f69-9037-db019002c8a1-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 01 08:42:16 crc kubenswrapper[4813]: I1201 08:42:16.819242 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-sp4wh" Dec 01 08:42:16 crc kubenswrapper[4813]: I1201 08:42:16.825662 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-sp4wh" Dec 01 08:42:17 crc kubenswrapper[4813]: I1201 08:42:17.066672 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 01 08:42:17 crc kubenswrapper[4813]: I1201 08:42:17.067317 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"0f66717e-8cd9-4f69-9037-db019002c8a1","Type":"ContainerDied","Data":"4ab8be9ba54af1859f607786e5ece9cc2bf89989468c1bfd69b508aec431e747"} Dec 01 08:42:17 crc kubenswrapper[4813]: I1201 08:42:17.067401 4813 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4ab8be9ba54af1859f607786e5ece9cc2bf89989468c1bfd69b508aec431e747" Dec 01 08:42:23 crc kubenswrapper[4813]: I1201 08:42:23.849929 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-9ql8j" Dec 01 08:42:23 crc kubenswrapper[4813]: I1201 08:42:23.855493 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-9ql8j" Dec 01 08:42:24 crc kubenswrapper[4813]: I1201 08:42:24.158817 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-ct7w8" Dec 01 08:42:27 crc kubenswrapper[4813]: I1201 08:42:27.799108 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-mj64s" Dec 01 08:42:35 crc kubenswrapper[4813]: I1201 08:42:35.585389 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-l5xp9" Dec 01 08:42:40 crc kubenswrapper[4813]: E1201 08:42:40.553734 4813 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Dec 01 08:42:40 crc kubenswrapper[4813]: E1201 08:42:40.554605 4813 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-ns6n2,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-qh5hd_openshift-marketplace(26f4804b-998f-488c-a501-d1cc4ef7f0b4): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 01 08:42:40 crc kubenswrapper[4813]: E1201 08:42:40.555895 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-qh5hd" podUID="26f4804b-998f-488c-a501-d1cc4ef7f0b4" Dec 01 08:42:43 crc kubenswrapper[4813]: E1201 08:42:43.949464 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-qh5hd" podUID="26f4804b-998f-488c-a501-d1cc4ef7f0b4" Dec 01 08:42:44 crc kubenswrapper[4813]: E1201 08:42:44.015579 4813 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Dec 01 08:42:44 crc kubenswrapper[4813]: E1201 08:42:44.015775 4813 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-xjp26,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-8wwz4_openshift-marketplace(95d1024c-5b73-40ae-9460-29502da48a23): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 01 08:42:44 crc kubenswrapper[4813]: E1201 08:42:44.016997 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-8wwz4" podUID="95d1024c-5b73-40ae-9460-29502da48a23" Dec 01 08:42:44 crc kubenswrapper[4813]: E1201 08:42:44.039542 4813 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Dec 01 08:42:44 crc kubenswrapper[4813]: E1201 08:42:44.039735 4813 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-vqnmj,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-lk5db_openshift-marketplace(4f1619ff-e7ed-42ee-b7a9-484c18d6677d): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 01 08:42:44 crc kubenswrapper[4813]: E1201 08:42:44.041868 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-lk5db" podUID="4f1619ff-e7ed-42ee-b7a9-484c18d6677d" Dec 01 08:42:45 crc kubenswrapper[4813]: I1201 08:42:45.284407 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Dec 01 08:42:45 crc kubenswrapper[4813]: E1201 08:42:45.284955 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0f66717e-8cd9-4f69-9037-db019002c8a1" containerName="pruner" Dec 01 08:42:45 crc kubenswrapper[4813]: I1201 08:42:45.284995 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="0f66717e-8cd9-4f69-9037-db019002c8a1" containerName="pruner" Dec 01 08:42:45 crc kubenswrapper[4813]: I1201 08:42:45.285146 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="0f66717e-8cd9-4f69-9037-db019002c8a1" containerName="pruner" Dec 01 08:42:45 crc kubenswrapper[4813]: I1201 08:42:45.286586 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 01 08:42:45 crc kubenswrapper[4813]: I1201 08:42:45.300912 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Dec 01 08:42:45 crc kubenswrapper[4813]: I1201 08:42:45.300983 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Dec 01 08:42:45 crc kubenswrapper[4813]: I1201 08:42:45.300912 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Dec 01 08:42:45 crc kubenswrapper[4813]: I1201 08:42:45.402468 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/03b218ed-5951-4b7e-8df7-b33977fd70dd-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"03b218ed-5951-4b7e-8df7-b33977fd70dd\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 01 08:42:45 crc kubenswrapper[4813]: I1201 08:42:45.402653 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/03b218ed-5951-4b7e-8df7-b33977fd70dd-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"03b218ed-5951-4b7e-8df7-b33977fd70dd\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 01 08:42:45 crc kubenswrapper[4813]: I1201 08:42:45.503246 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/03b218ed-5951-4b7e-8df7-b33977fd70dd-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"03b218ed-5951-4b7e-8df7-b33977fd70dd\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 01 08:42:45 crc kubenswrapper[4813]: I1201 08:42:45.503591 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/03b218ed-5951-4b7e-8df7-b33977fd70dd-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"03b218ed-5951-4b7e-8df7-b33977fd70dd\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 01 08:42:45 crc kubenswrapper[4813]: I1201 08:42:45.503680 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/03b218ed-5951-4b7e-8df7-b33977fd70dd-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"03b218ed-5951-4b7e-8df7-b33977fd70dd\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 01 08:42:45 crc kubenswrapper[4813]: I1201 08:42:45.536771 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/03b218ed-5951-4b7e-8df7-b33977fd70dd-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"03b218ed-5951-4b7e-8df7-b33977fd70dd\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 01 08:42:45 crc kubenswrapper[4813]: I1201 08:42:45.627548 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 01 08:42:46 crc kubenswrapper[4813]: I1201 08:42:46.313465 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 08:42:46 crc kubenswrapper[4813]: I1201 08:42:46.313876 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 08:42:46 crc kubenswrapper[4813]: I1201 08:42:46.315932 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Dec 01 08:42:46 crc kubenswrapper[4813]: I1201 08:42:46.316410 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Dec 01 08:42:46 crc kubenswrapper[4813]: I1201 08:42:46.325559 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 08:42:46 crc kubenswrapper[4813]: I1201 08:42:46.327816 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 08:42:46 crc kubenswrapper[4813]: I1201 08:42:46.415758 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 08:42:46 crc kubenswrapper[4813]: I1201 08:42:46.415814 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 08:42:46 crc kubenswrapper[4813]: I1201 08:42:46.417776 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Dec 01 08:42:46 crc kubenswrapper[4813]: I1201 08:42:46.428057 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Dec 01 08:42:46 crc kubenswrapper[4813]: I1201 08:42:46.440827 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 08:42:46 crc kubenswrapper[4813]: I1201 08:42:46.461284 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 08:42:46 crc kubenswrapper[4813]: I1201 08:42:46.471023 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 08:42:46 crc kubenswrapper[4813]: I1201 08:42:46.524081 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 08:42:46 crc kubenswrapper[4813]: I1201 08:42:46.748958 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 08:42:46 crc kubenswrapper[4813]: E1201 08:42:46.847847 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-lk5db" podUID="4f1619ff-e7ed-42ee-b7a9-484c18d6677d" Dec 01 08:42:46 crc kubenswrapper[4813]: E1201 08:42:46.847907 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-8wwz4" podUID="95d1024c-5b73-40ae-9460-29502da48a23" Dec 01 08:42:46 crc kubenswrapper[4813]: E1201 08:42:46.919600 4813 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Dec 01 08:42:46 crc kubenswrapper[4813]: E1201 08:42:46.919925 4813 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-srz6m,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-7fqxs_openshift-marketplace(0959ee73-31e7-4d58-b124-3e971d777225): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 01 08:42:46 crc kubenswrapper[4813]: E1201 08:42:46.921388 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-7fqxs" podUID="0959ee73-31e7-4d58-b124-3e971d777225" Dec 01 08:42:46 crc kubenswrapper[4813]: E1201 08:42:46.940497 4813 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Dec 01 08:42:46 crc kubenswrapper[4813]: E1201 08:42:46.940638 4813 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-4rzrc,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-wbvcp_openshift-marketplace(979ea876-4dd3-4d31-9144-7a9a6e914c67): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 01 08:42:46 crc kubenswrapper[4813]: E1201 08:42:46.941138 4813 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Dec 01 08:42:46 crc kubenswrapper[4813]: E1201 08:42:46.941389 4813 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-qbk44,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-58jst_openshift-marketplace(3c0b52aa-86ba-4c56-b23e-108cdb4750d9): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 01 08:42:46 crc kubenswrapper[4813]: E1201 08:42:46.942308 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-wbvcp" podUID="979ea876-4dd3-4d31-9144-7a9a6e914c67" Dec 01 08:42:46 crc kubenswrapper[4813]: E1201 08:42:46.942544 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-58jst" podUID="3c0b52aa-86ba-4c56-b23e-108cdb4750d9" Dec 01 08:42:47 crc kubenswrapper[4813]: I1201 08:42:47.224201 4813 patch_prober.go:28] interesting pod/machine-config-daemon-9wn4q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 08:42:47 crc kubenswrapper[4813]: I1201 08:42:47.224283 4813 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9wn4q" podUID="3d423402-e08f-42e9-b60c-6eb9bf2f6d48" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 08:42:47 crc kubenswrapper[4813]: W1201 08:42:47.441878 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9d751cbb_f2e2_430d_9754_c882a5e924a5.slice/crio-554c493b8e09f27394f291cfcdbcabce27b416f7fe5d4d68fb7036993ae79057 WatchSource:0}: Error finding container 554c493b8e09f27394f291cfcdbcabce27b416f7fe5d4d68fb7036993ae79057: Status 404 returned error can't find the container with id 554c493b8e09f27394f291cfcdbcabce27b416f7fe5d4d68fb7036993ae79057 Dec 01 08:42:47 crc kubenswrapper[4813]: I1201 08:42:47.468380 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Dec 01 08:42:47 crc kubenswrapper[4813]: I1201 08:42:47.477500 4813 generic.go:334] "Generic (PLEG): container finished" podID="a81d6f9c-e9f7-4c98-8233-c9f214978bfa" containerID="74c966ba35cf3ff64990cb6a68b3ead5a31c5eb0f3d3dd3c143682299f8f9411" exitCode=0 Dec 01 08:42:47 crc kubenswrapper[4813]: I1201 08:42:47.477603 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xswbf" event={"ID":"a81d6f9c-e9f7-4c98-8233-c9f214978bfa","Type":"ContainerDied","Data":"74c966ba35cf3ff64990cb6a68b3ead5a31c5eb0f3d3dd3c143682299f8f9411"} Dec 01 08:42:47 crc kubenswrapper[4813]: I1201 08:42:47.484039 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bpvbz" event={"ID":"bd02bd16-7f86-48ad-b554-6404f62b1b28","Type":"ContainerDied","Data":"bf5a076910ff550f7a79f4464c5a065cb7d1461cd9186aa2f1a94e1923d57ccc"} Dec 01 08:42:47 crc kubenswrapper[4813]: I1201 08:42:47.484117 4813 generic.go:334] "Generic (PLEG): container finished" podID="bd02bd16-7f86-48ad-b554-6404f62b1b28" containerID="bf5a076910ff550f7a79f4464c5a065cb7d1461cd9186aa2f1a94e1923d57ccc" exitCode=0 Dec 01 08:42:47 crc kubenswrapper[4813]: I1201 08:42:47.490235 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"554c493b8e09f27394f291cfcdbcabce27b416f7fe5d4d68fb7036993ae79057"} Dec 01 08:42:47 crc kubenswrapper[4813]: E1201 08:42:47.492126 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-7fqxs" podUID="0959ee73-31e7-4d58-b124-3e971d777225" Dec 01 08:42:47 crc kubenswrapper[4813]: E1201 08:42:47.492695 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-58jst" podUID="3c0b52aa-86ba-4c56-b23e-108cdb4750d9" Dec 01 08:42:47 crc kubenswrapper[4813]: E1201 08:42:47.494878 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-wbvcp" podUID="979ea876-4dd3-4d31-9144-7a9a6e914c67" Dec 01 08:42:47 crc kubenswrapper[4813]: W1201 08:42:47.514406 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3b6479f0_333b_4a96_9adf_2099afdc2447.slice/crio-9f3778fa2d10bd5e6a2ed370afd08414ecca88cce5e8268b74080d3af94ce0ee WatchSource:0}: Error finding container 9f3778fa2d10bd5e6a2ed370afd08414ecca88cce5e8268b74080d3af94ce0ee: Status 404 returned error can't find the container with id 9f3778fa2d10bd5e6a2ed370afd08414ecca88cce5e8268b74080d3af94ce0ee Dec 01 08:42:48 crc kubenswrapper[4813]: I1201 08:42:48.496586 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xswbf" event={"ID":"a81d6f9c-e9f7-4c98-8233-c9f214978bfa","Type":"ContainerStarted","Data":"d7d18c55dcb481dc146267eb09b218c0f48f0b8608f81c13362bb51c160ee0c9"} Dec 01 08:42:48 crc kubenswrapper[4813]: I1201 08:42:48.501247 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bpvbz" event={"ID":"bd02bd16-7f86-48ad-b554-6404f62b1b28","Type":"ContainerStarted","Data":"8e00841bcb69076ca8fea77c435203fb0d3323db5bb7e9884cd944fcd6693d71"} Dec 01 08:42:48 crc kubenswrapper[4813]: I1201 08:42:48.502654 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"c22f0986cd38b7367b53b22fefc9a5b9ab89fc4322ffa21312c6dc6ca082bd53"} Dec 01 08:42:48 crc kubenswrapper[4813]: I1201 08:42:48.504377 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"454b85d627ef58379a61f0349d9180249f268e91b7ddc63af556175db1892666"} Dec 01 08:42:48 crc kubenswrapper[4813]: I1201 08:42:48.504406 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"76bf66276c945ebc2846bac145ac3cc161b4b414a4579dcd74404e5ff719b869"} Dec 01 08:42:48 crc kubenswrapper[4813]: I1201 08:42:48.506194 4813 generic.go:334] "Generic (PLEG): container finished" podID="03b218ed-5951-4b7e-8df7-b33977fd70dd" containerID="54d0f8160523fc7e8124519c9736b5fc879a0fc00fb7f5843f2186b288f6f30a" exitCode=0 Dec 01 08:42:48 crc kubenswrapper[4813]: I1201 08:42:48.506423 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"03b218ed-5951-4b7e-8df7-b33977fd70dd","Type":"ContainerDied","Data":"54d0f8160523fc7e8124519c9736b5fc879a0fc00fb7f5843f2186b288f6f30a"} Dec 01 08:42:48 crc kubenswrapper[4813]: I1201 08:42:48.506470 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"03b218ed-5951-4b7e-8df7-b33977fd70dd","Type":"ContainerStarted","Data":"54e01ecd35a54c586de23652bcb5ef60a0cd482e88903dfb18d9f150aefe4570"} Dec 01 08:42:48 crc kubenswrapper[4813]: I1201 08:42:48.511115 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"316b7a4a17ce1e68b06adc17642456914da8b27992faaf43898bb9cda471e4d0"} Dec 01 08:42:48 crc kubenswrapper[4813]: I1201 08:42:48.511139 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"9f3778fa2d10bd5e6a2ed370afd08414ecca88cce5e8268b74080d3af94ce0ee"} Dec 01 08:42:48 crc kubenswrapper[4813]: I1201 08:42:48.511779 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 08:42:48 crc kubenswrapper[4813]: I1201 08:42:48.536988 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-bpvbz" podStartSLOduration=2.053627011 podStartE2EDuration="40.536910272s" podCreationTimestamp="2025-12-01 08:42:08 +0000 UTC" firstStartedPulling="2025-12-01 08:42:09.4907897 +0000 UTC m=+114.113711286" lastFinishedPulling="2025-12-01 08:42:47.974072951 +0000 UTC m=+152.596994547" observedRunningTime="2025-12-01 08:42:48.536327274 +0000 UTC m=+153.159248880" watchObservedRunningTime="2025-12-01 08:42:48.536910272 +0000 UTC m=+153.159831858" Dec 01 08:42:48 crc kubenswrapper[4813]: I1201 08:42:48.537675 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-xswbf" podStartSLOduration=2.8347305609999998 podStartE2EDuration="42.537669275s" podCreationTimestamp="2025-12-01 08:42:06 +0000 UTC" firstStartedPulling="2025-12-01 08:42:08.376266156 +0000 UTC m=+112.999187742" lastFinishedPulling="2025-12-01 08:42:48.07920487 +0000 UTC m=+152.702126456" observedRunningTime="2025-12-01 08:42:48.5163947 +0000 UTC m=+153.139316286" watchObservedRunningTime="2025-12-01 08:42:48.537669275 +0000 UTC m=+153.160590861" Dec 01 08:42:48 crc kubenswrapper[4813]: I1201 08:42:48.652756 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-bpvbz" Dec 01 08:42:48 crc kubenswrapper[4813]: I1201 08:42:48.653111 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-bpvbz" Dec 01 08:42:49 crc kubenswrapper[4813]: I1201 08:42:49.775208 4813 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-marketplace-bpvbz" podUID="bd02bd16-7f86-48ad-b554-6404f62b1b28" containerName="registry-server" probeResult="failure" output=< Dec 01 08:42:49 crc kubenswrapper[4813]: timeout: failed to connect service ":50051" within 1s Dec 01 08:42:49 crc kubenswrapper[4813]: > Dec 01 08:42:49 crc kubenswrapper[4813]: I1201 08:42:49.925127 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 01 08:42:49 crc kubenswrapper[4813]: I1201 08:42:49.930948 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/03b218ed-5951-4b7e-8df7-b33977fd70dd-kubelet-dir\") pod \"03b218ed-5951-4b7e-8df7-b33977fd70dd\" (UID: \"03b218ed-5951-4b7e-8df7-b33977fd70dd\") " Dec 01 08:42:49 crc kubenswrapper[4813]: I1201 08:42:49.931086 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/03b218ed-5951-4b7e-8df7-b33977fd70dd-kube-api-access\") pod \"03b218ed-5951-4b7e-8df7-b33977fd70dd\" (UID: \"03b218ed-5951-4b7e-8df7-b33977fd70dd\") " Dec 01 08:42:49 crc kubenswrapper[4813]: I1201 08:42:49.931722 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/03b218ed-5951-4b7e-8df7-b33977fd70dd-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "03b218ed-5951-4b7e-8df7-b33977fd70dd" (UID: "03b218ed-5951-4b7e-8df7-b33977fd70dd"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 08:42:49 crc kubenswrapper[4813]: I1201 08:42:49.939265 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/03b218ed-5951-4b7e-8df7-b33977fd70dd-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "03b218ed-5951-4b7e-8df7-b33977fd70dd" (UID: "03b218ed-5951-4b7e-8df7-b33977fd70dd"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:42:50 crc kubenswrapper[4813]: I1201 08:42:50.032466 4813 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/03b218ed-5951-4b7e-8df7-b33977fd70dd-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 01 08:42:50 crc kubenswrapper[4813]: I1201 08:42:50.032560 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/03b218ed-5951-4b7e-8df7-b33977fd70dd-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 01 08:42:50 crc kubenswrapper[4813]: I1201 08:42:50.284947 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Dec 01 08:42:50 crc kubenswrapper[4813]: E1201 08:42:50.285249 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="03b218ed-5951-4b7e-8df7-b33977fd70dd" containerName="pruner" Dec 01 08:42:50 crc kubenswrapper[4813]: I1201 08:42:50.285265 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="03b218ed-5951-4b7e-8df7-b33977fd70dd" containerName="pruner" Dec 01 08:42:50 crc kubenswrapper[4813]: I1201 08:42:50.285402 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="03b218ed-5951-4b7e-8df7-b33977fd70dd" containerName="pruner" Dec 01 08:42:50 crc kubenswrapper[4813]: I1201 08:42:50.285829 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 01 08:42:50 crc kubenswrapper[4813]: I1201 08:42:50.292181 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Dec 01 08:42:50 crc kubenswrapper[4813]: I1201 08:42:50.437308 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/269d86e5-dc67-48b3-a955-201b986f6c97-kube-api-access\") pod \"installer-9-crc\" (UID: \"269d86e5-dc67-48b3-a955-201b986f6c97\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 01 08:42:50 crc kubenswrapper[4813]: I1201 08:42:50.437377 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/269d86e5-dc67-48b3-a955-201b986f6c97-kubelet-dir\") pod \"installer-9-crc\" (UID: \"269d86e5-dc67-48b3-a955-201b986f6c97\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 01 08:42:50 crc kubenswrapper[4813]: I1201 08:42:50.437442 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/269d86e5-dc67-48b3-a955-201b986f6c97-var-lock\") pod \"installer-9-crc\" (UID: \"269d86e5-dc67-48b3-a955-201b986f6c97\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 01 08:42:50 crc kubenswrapper[4813]: I1201 08:42:50.521371 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"03b218ed-5951-4b7e-8df7-b33977fd70dd","Type":"ContainerDied","Data":"54e01ecd35a54c586de23652bcb5ef60a0cd482e88903dfb18d9f150aefe4570"} Dec 01 08:42:50 crc kubenswrapper[4813]: I1201 08:42:50.521423 4813 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="54e01ecd35a54c586de23652bcb5ef60a0cd482e88903dfb18d9f150aefe4570" Dec 01 08:42:50 crc kubenswrapper[4813]: I1201 08:42:50.521480 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 01 08:42:50 crc kubenswrapper[4813]: I1201 08:42:50.538512 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/269d86e5-dc67-48b3-a955-201b986f6c97-var-lock\") pod \"installer-9-crc\" (UID: \"269d86e5-dc67-48b3-a955-201b986f6c97\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 01 08:42:50 crc kubenswrapper[4813]: I1201 08:42:50.538606 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/269d86e5-dc67-48b3-a955-201b986f6c97-kube-api-access\") pod \"installer-9-crc\" (UID: \"269d86e5-dc67-48b3-a955-201b986f6c97\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 01 08:42:50 crc kubenswrapper[4813]: I1201 08:42:50.538635 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/269d86e5-dc67-48b3-a955-201b986f6c97-kubelet-dir\") pod \"installer-9-crc\" (UID: \"269d86e5-dc67-48b3-a955-201b986f6c97\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 01 08:42:50 crc kubenswrapper[4813]: I1201 08:42:50.538720 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/269d86e5-dc67-48b3-a955-201b986f6c97-kubelet-dir\") pod \"installer-9-crc\" (UID: \"269d86e5-dc67-48b3-a955-201b986f6c97\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 01 08:42:50 crc kubenswrapper[4813]: I1201 08:42:50.538907 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/269d86e5-dc67-48b3-a955-201b986f6c97-var-lock\") pod \"installer-9-crc\" (UID: \"269d86e5-dc67-48b3-a955-201b986f6c97\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 01 08:42:50 crc kubenswrapper[4813]: I1201 08:42:50.563403 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/269d86e5-dc67-48b3-a955-201b986f6c97-kube-api-access\") pod \"installer-9-crc\" (UID: \"269d86e5-dc67-48b3-a955-201b986f6c97\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 01 08:42:50 crc kubenswrapper[4813]: I1201 08:42:50.606228 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 01 08:42:50 crc kubenswrapper[4813]: I1201 08:42:50.984152 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Dec 01 08:42:51 crc kubenswrapper[4813]: I1201 08:42:51.529952 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"269d86e5-dc67-48b3-a955-201b986f6c97","Type":"ContainerStarted","Data":"7c8d443ea8564db109e481b03d907237fa30fd861875a9041e49e2dc433321f8"} Dec 01 08:42:52 crc kubenswrapper[4813]: I1201 08:42:52.536593 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"269d86e5-dc67-48b3-a955-201b986f6c97","Type":"ContainerStarted","Data":"1ed1b375825fa13897f16c218101a45d9e8e16e971dbd5c8cde715800e1ee232"} Dec 01 08:42:52 crc kubenswrapper[4813]: I1201 08:42:52.557482 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/installer-9-crc" podStartSLOduration=2.55745823 podStartE2EDuration="2.55745823s" podCreationTimestamp="2025-12-01 08:42:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:42:52.554176118 +0000 UTC m=+157.177097724" watchObservedRunningTime="2025-12-01 08:42:52.55745823 +0000 UTC m=+157.180379816" Dec 01 08:42:53 crc kubenswrapper[4813]: I1201 08:42:53.389596 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-5fgn9"] Dec 01 08:42:56 crc kubenswrapper[4813]: I1201 08:42:56.829197 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-xswbf" Dec 01 08:42:56 crc kubenswrapper[4813]: I1201 08:42:56.829597 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-xswbf" Dec 01 08:42:56 crc kubenswrapper[4813]: I1201 08:42:56.871666 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-xswbf" Dec 01 08:42:57 crc kubenswrapper[4813]: I1201 08:42:57.602480 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-xswbf" Dec 01 08:42:58 crc kubenswrapper[4813]: I1201 08:42:58.106598 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-xswbf"] Dec 01 08:42:58 crc kubenswrapper[4813]: I1201 08:42:58.573315 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lk5db" event={"ID":"4f1619ff-e7ed-42ee-b7a9-484c18d6677d","Type":"ContainerStarted","Data":"e35636586ffe20bdca2438e4633c8f87e36115bddb54a527ea25d024c2d8b53c"} Dec 01 08:42:58 crc kubenswrapper[4813]: E1201 08:42:58.658446 4813 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4f1619ff_e7ed_42ee_b7a9_484c18d6677d.slice/crio-e35636586ffe20bdca2438e4633c8f87e36115bddb54a527ea25d024c2d8b53c.scope\": RecentStats: unable to find data in memory cache]" Dec 01 08:42:58 crc kubenswrapper[4813]: I1201 08:42:58.688859 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-bpvbz" Dec 01 08:42:58 crc kubenswrapper[4813]: I1201 08:42:58.728878 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-bpvbz" Dec 01 08:42:59 crc kubenswrapper[4813]: I1201 08:42:59.580033 4813 generic.go:334] "Generic (PLEG): container finished" podID="4f1619ff-e7ed-42ee-b7a9-484c18d6677d" containerID="e35636586ffe20bdca2438e4633c8f87e36115bddb54a527ea25d024c2d8b53c" exitCode=0 Dec 01 08:42:59 crc kubenswrapper[4813]: I1201 08:42:59.580240 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-xswbf" podUID="a81d6f9c-e9f7-4c98-8233-c9f214978bfa" containerName="registry-server" containerID="cri-o://d7d18c55dcb481dc146267eb09b218c0f48f0b8608f81c13362bb51c160ee0c9" gracePeriod=2 Dec 01 08:42:59 crc kubenswrapper[4813]: I1201 08:42:59.580566 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lk5db" event={"ID":"4f1619ff-e7ed-42ee-b7a9-484c18d6677d","Type":"ContainerDied","Data":"e35636586ffe20bdca2438e4633c8f87e36115bddb54a527ea25d024c2d8b53c"} Dec 01 08:43:00 crc kubenswrapper[4813]: I1201 08:43:00.129775 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xswbf" Dec 01 08:43:00 crc kubenswrapper[4813]: I1201 08:43:00.161919 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f7vm5\" (UniqueName: \"kubernetes.io/projected/a81d6f9c-e9f7-4c98-8233-c9f214978bfa-kube-api-access-f7vm5\") pod \"a81d6f9c-e9f7-4c98-8233-c9f214978bfa\" (UID: \"a81d6f9c-e9f7-4c98-8233-c9f214978bfa\") " Dec 01 08:43:00 crc kubenswrapper[4813]: I1201 08:43:00.162019 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a81d6f9c-e9f7-4c98-8233-c9f214978bfa-catalog-content\") pod \"a81d6f9c-e9f7-4c98-8233-c9f214978bfa\" (UID: \"a81d6f9c-e9f7-4c98-8233-c9f214978bfa\") " Dec 01 08:43:00 crc kubenswrapper[4813]: I1201 08:43:00.162057 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a81d6f9c-e9f7-4c98-8233-c9f214978bfa-utilities\") pod \"a81d6f9c-e9f7-4c98-8233-c9f214978bfa\" (UID: \"a81d6f9c-e9f7-4c98-8233-c9f214978bfa\") " Dec 01 08:43:00 crc kubenswrapper[4813]: I1201 08:43:00.162860 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a81d6f9c-e9f7-4c98-8233-c9f214978bfa-utilities" (OuterVolumeSpecName: "utilities") pod "a81d6f9c-e9f7-4c98-8233-c9f214978bfa" (UID: "a81d6f9c-e9f7-4c98-8233-c9f214978bfa"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 08:43:00 crc kubenswrapper[4813]: I1201 08:43:00.168148 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a81d6f9c-e9f7-4c98-8233-c9f214978bfa-kube-api-access-f7vm5" (OuterVolumeSpecName: "kube-api-access-f7vm5") pod "a81d6f9c-e9f7-4c98-8233-c9f214978bfa" (UID: "a81d6f9c-e9f7-4c98-8233-c9f214978bfa"). InnerVolumeSpecName "kube-api-access-f7vm5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:43:00 crc kubenswrapper[4813]: I1201 08:43:00.205887 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a81d6f9c-e9f7-4c98-8233-c9f214978bfa-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a81d6f9c-e9f7-4c98-8233-c9f214978bfa" (UID: "a81d6f9c-e9f7-4c98-8233-c9f214978bfa"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 08:43:00 crc kubenswrapper[4813]: I1201 08:43:00.263109 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f7vm5\" (UniqueName: \"kubernetes.io/projected/a81d6f9c-e9f7-4c98-8233-c9f214978bfa-kube-api-access-f7vm5\") on node \"crc\" DevicePath \"\"" Dec 01 08:43:00 crc kubenswrapper[4813]: I1201 08:43:00.263156 4813 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a81d6f9c-e9f7-4c98-8233-c9f214978bfa-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 08:43:00 crc kubenswrapper[4813]: I1201 08:43:00.263169 4813 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a81d6f9c-e9f7-4c98-8233-c9f214978bfa-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 08:43:00 crc kubenswrapper[4813]: I1201 08:43:00.586593 4813 generic.go:334] "Generic (PLEG): container finished" podID="a81d6f9c-e9f7-4c98-8233-c9f214978bfa" containerID="d7d18c55dcb481dc146267eb09b218c0f48f0b8608f81c13362bb51c160ee0c9" exitCode=0 Dec 01 08:43:00 crc kubenswrapper[4813]: I1201 08:43:00.586665 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xswbf" Dec 01 08:43:00 crc kubenswrapper[4813]: I1201 08:43:00.586649 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xswbf" event={"ID":"a81d6f9c-e9f7-4c98-8233-c9f214978bfa","Type":"ContainerDied","Data":"d7d18c55dcb481dc146267eb09b218c0f48f0b8608f81c13362bb51c160ee0c9"} Dec 01 08:43:00 crc kubenswrapper[4813]: I1201 08:43:00.586817 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xswbf" event={"ID":"a81d6f9c-e9f7-4c98-8233-c9f214978bfa","Type":"ContainerDied","Data":"9a24ad031c98407f6e9a691b08ad6410f5e6ca12c71bca1033958e076c49a5fe"} Dec 01 08:43:00 crc kubenswrapper[4813]: I1201 08:43:00.586917 4813 scope.go:117] "RemoveContainer" containerID="d7d18c55dcb481dc146267eb09b218c0f48f0b8608f81c13362bb51c160ee0c9" Dec 01 08:43:00 crc kubenswrapper[4813]: I1201 08:43:00.627676 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-xswbf"] Dec 01 08:43:00 crc kubenswrapper[4813]: I1201 08:43:00.628297 4813 scope.go:117] "RemoveContainer" containerID="74c966ba35cf3ff64990cb6a68b3ead5a31c5eb0f3d3dd3c143682299f8f9411" Dec 01 08:43:00 crc kubenswrapper[4813]: I1201 08:43:00.633664 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-xswbf"] Dec 01 08:43:00 crc kubenswrapper[4813]: I1201 08:43:00.652152 4813 scope.go:117] "RemoveContainer" containerID="8ca7da1c941e45e90ca434d5fd0b5e0fee687124784fb0f939a7fc7d231a558a" Dec 01 08:43:00 crc kubenswrapper[4813]: I1201 08:43:00.672658 4813 scope.go:117] "RemoveContainer" containerID="d7d18c55dcb481dc146267eb09b218c0f48f0b8608f81c13362bb51c160ee0c9" Dec 01 08:43:00 crc kubenswrapper[4813]: E1201 08:43:00.673287 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d7d18c55dcb481dc146267eb09b218c0f48f0b8608f81c13362bb51c160ee0c9\": container with ID starting with d7d18c55dcb481dc146267eb09b218c0f48f0b8608f81c13362bb51c160ee0c9 not found: ID does not exist" containerID="d7d18c55dcb481dc146267eb09b218c0f48f0b8608f81c13362bb51c160ee0c9" Dec 01 08:43:00 crc kubenswrapper[4813]: I1201 08:43:00.673324 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d7d18c55dcb481dc146267eb09b218c0f48f0b8608f81c13362bb51c160ee0c9"} err="failed to get container status \"d7d18c55dcb481dc146267eb09b218c0f48f0b8608f81c13362bb51c160ee0c9\": rpc error: code = NotFound desc = could not find container \"d7d18c55dcb481dc146267eb09b218c0f48f0b8608f81c13362bb51c160ee0c9\": container with ID starting with d7d18c55dcb481dc146267eb09b218c0f48f0b8608f81c13362bb51c160ee0c9 not found: ID does not exist" Dec 01 08:43:00 crc kubenswrapper[4813]: I1201 08:43:00.673380 4813 scope.go:117] "RemoveContainer" containerID="74c966ba35cf3ff64990cb6a68b3ead5a31c5eb0f3d3dd3c143682299f8f9411" Dec 01 08:43:00 crc kubenswrapper[4813]: E1201 08:43:00.674012 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"74c966ba35cf3ff64990cb6a68b3ead5a31c5eb0f3d3dd3c143682299f8f9411\": container with ID starting with 74c966ba35cf3ff64990cb6a68b3ead5a31c5eb0f3d3dd3c143682299f8f9411 not found: ID does not exist" containerID="74c966ba35cf3ff64990cb6a68b3ead5a31c5eb0f3d3dd3c143682299f8f9411" Dec 01 08:43:00 crc kubenswrapper[4813]: I1201 08:43:00.674041 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"74c966ba35cf3ff64990cb6a68b3ead5a31c5eb0f3d3dd3c143682299f8f9411"} err="failed to get container status \"74c966ba35cf3ff64990cb6a68b3ead5a31c5eb0f3d3dd3c143682299f8f9411\": rpc error: code = NotFound desc = could not find container \"74c966ba35cf3ff64990cb6a68b3ead5a31c5eb0f3d3dd3c143682299f8f9411\": container with ID starting with 74c966ba35cf3ff64990cb6a68b3ead5a31c5eb0f3d3dd3c143682299f8f9411 not found: ID does not exist" Dec 01 08:43:00 crc kubenswrapper[4813]: I1201 08:43:00.674058 4813 scope.go:117] "RemoveContainer" containerID="8ca7da1c941e45e90ca434d5fd0b5e0fee687124784fb0f939a7fc7d231a558a" Dec 01 08:43:00 crc kubenswrapper[4813]: E1201 08:43:00.674427 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8ca7da1c941e45e90ca434d5fd0b5e0fee687124784fb0f939a7fc7d231a558a\": container with ID starting with 8ca7da1c941e45e90ca434d5fd0b5e0fee687124784fb0f939a7fc7d231a558a not found: ID does not exist" containerID="8ca7da1c941e45e90ca434d5fd0b5e0fee687124784fb0f939a7fc7d231a558a" Dec 01 08:43:00 crc kubenswrapper[4813]: I1201 08:43:00.674450 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8ca7da1c941e45e90ca434d5fd0b5e0fee687124784fb0f939a7fc7d231a558a"} err="failed to get container status \"8ca7da1c941e45e90ca434d5fd0b5e0fee687124784fb0f939a7fc7d231a558a\": rpc error: code = NotFound desc = could not find container \"8ca7da1c941e45e90ca434d5fd0b5e0fee687124784fb0f939a7fc7d231a558a\": container with ID starting with 8ca7da1c941e45e90ca434d5fd0b5e0fee687124784fb0f939a7fc7d231a558a not found: ID does not exist" Dec 01 08:43:00 crc kubenswrapper[4813]: I1201 08:43:00.897905 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-bpvbz"] Dec 01 08:43:00 crc kubenswrapper[4813]: I1201 08:43:00.898160 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-bpvbz" podUID="bd02bd16-7f86-48ad-b554-6404f62b1b28" containerName="registry-server" containerID="cri-o://8e00841bcb69076ca8fea77c435203fb0d3323db5bb7e9884cd944fcd6693d71" gracePeriod=2 Dec 01 08:43:01 crc kubenswrapper[4813]: I1201 08:43:01.596623 4813 generic.go:334] "Generic (PLEG): container finished" podID="bd02bd16-7f86-48ad-b554-6404f62b1b28" containerID="8e00841bcb69076ca8fea77c435203fb0d3323db5bb7e9884cd944fcd6693d71" exitCode=0 Dec 01 08:43:01 crc kubenswrapper[4813]: I1201 08:43:01.596741 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bpvbz" event={"ID":"bd02bd16-7f86-48ad-b554-6404f62b1b28","Type":"ContainerDied","Data":"8e00841bcb69076ca8fea77c435203fb0d3323db5bb7e9884cd944fcd6693d71"} Dec 01 08:43:02 crc kubenswrapper[4813]: I1201 08:43:02.402021 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a81d6f9c-e9f7-4c98-8233-c9f214978bfa" path="/var/lib/kubelet/pods/a81d6f9c-e9f7-4c98-8233-c9f214978bfa/volumes" Dec 01 08:43:04 crc kubenswrapper[4813]: I1201 08:43:04.343070 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bpvbz" Dec 01 08:43:04 crc kubenswrapper[4813]: I1201 08:43:04.518533 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bd02bd16-7f86-48ad-b554-6404f62b1b28-catalog-content\") pod \"bd02bd16-7f86-48ad-b554-6404f62b1b28\" (UID: \"bd02bd16-7f86-48ad-b554-6404f62b1b28\") " Dec 01 08:43:04 crc kubenswrapper[4813]: I1201 08:43:04.518605 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tbrl7\" (UniqueName: \"kubernetes.io/projected/bd02bd16-7f86-48ad-b554-6404f62b1b28-kube-api-access-tbrl7\") pod \"bd02bd16-7f86-48ad-b554-6404f62b1b28\" (UID: \"bd02bd16-7f86-48ad-b554-6404f62b1b28\") " Dec 01 08:43:04 crc kubenswrapper[4813]: I1201 08:43:04.518643 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bd02bd16-7f86-48ad-b554-6404f62b1b28-utilities\") pod \"bd02bd16-7f86-48ad-b554-6404f62b1b28\" (UID: \"bd02bd16-7f86-48ad-b554-6404f62b1b28\") " Dec 01 08:43:04 crc kubenswrapper[4813]: I1201 08:43:04.519675 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bd02bd16-7f86-48ad-b554-6404f62b1b28-utilities" (OuterVolumeSpecName: "utilities") pod "bd02bd16-7f86-48ad-b554-6404f62b1b28" (UID: "bd02bd16-7f86-48ad-b554-6404f62b1b28"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 08:43:04 crc kubenswrapper[4813]: I1201 08:43:04.526318 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd02bd16-7f86-48ad-b554-6404f62b1b28-kube-api-access-tbrl7" (OuterVolumeSpecName: "kube-api-access-tbrl7") pod "bd02bd16-7f86-48ad-b554-6404f62b1b28" (UID: "bd02bd16-7f86-48ad-b554-6404f62b1b28"). InnerVolumeSpecName "kube-api-access-tbrl7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:43:04 crc kubenswrapper[4813]: I1201 08:43:04.540741 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bd02bd16-7f86-48ad-b554-6404f62b1b28-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "bd02bd16-7f86-48ad-b554-6404f62b1b28" (UID: "bd02bd16-7f86-48ad-b554-6404f62b1b28"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 08:43:04 crc kubenswrapper[4813]: I1201 08:43:04.619916 4813 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bd02bd16-7f86-48ad-b554-6404f62b1b28-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 08:43:04 crc kubenswrapper[4813]: I1201 08:43:04.620428 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tbrl7\" (UniqueName: \"kubernetes.io/projected/bd02bd16-7f86-48ad-b554-6404f62b1b28-kube-api-access-tbrl7\") on node \"crc\" DevicePath \"\"" Dec 01 08:43:04 crc kubenswrapper[4813]: I1201 08:43:04.620551 4813 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bd02bd16-7f86-48ad-b554-6404f62b1b28-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 08:43:04 crc kubenswrapper[4813]: I1201 08:43:04.628127 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bpvbz" event={"ID":"bd02bd16-7f86-48ad-b554-6404f62b1b28","Type":"ContainerDied","Data":"e99621bb15483390dda6f16ddb39f8c2ff88ef51507adc0def12ac33409b4e20"} Dec 01 08:43:04 crc kubenswrapper[4813]: I1201 08:43:04.628198 4813 scope.go:117] "RemoveContainer" containerID="8e00841bcb69076ca8fea77c435203fb0d3323db5bb7e9884cd944fcd6693d71" Dec 01 08:43:04 crc kubenswrapper[4813]: I1201 08:43:04.628612 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bpvbz" Dec 01 08:43:04 crc kubenswrapper[4813]: I1201 08:43:04.634909 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lk5db" event={"ID":"4f1619ff-e7ed-42ee-b7a9-484c18d6677d","Type":"ContainerStarted","Data":"9ffc53906faf61c55a06e12e23404613fa25399b22cf92a6835b44ae83f7eb5a"} Dec 01 08:43:04 crc kubenswrapper[4813]: I1201 08:43:04.654820 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-lk5db" podStartSLOduration=2.723433934 podStartE2EDuration="57.654795047s" podCreationTimestamp="2025-12-01 08:42:07 +0000 UTC" firstStartedPulling="2025-12-01 08:42:09.449089631 +0000 UTC m=+114.072011227" lastFinishedPulling="2025-12-01 08:43:04.380450744 +0000 UTC m=+169.003372340" observedRunningTime="2025-12-01 08:43:04.651929318 +0000 UTC m=+169.274850914" watchObservedRunningTime="2025-12-01 08:43:04.654795047 +0000 UTC m=+169.277716643" Dec 01 08:43:04 crc kubenswrapper[4813]: I1201 08:43:04.669055 4813 scope.go:117] "RemoveContainer" containerID="bf5a076910ff550f7a79f4464c5a065cb7d1461cd9186aa2f1a94e1923d57ccc" Dec 01 08:43:04 crc kubenswrapper[4813]: I1201 08:43:04.693705 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-bpvbz"] Dec 01 08:43:04 crc kubenswrapper[4813]: I1201 08:43:04.696243 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-bpvbz"] Dec 01 08:43:04 crc kubenswrapper[4813]: I1201 08:43:04.706194 4813 scope.go:117] "RemoveContainer" containerID="e30f802ba7038b957dde0fd7ec61e4e78c5a773cb92cae1974cff027d52e3a2b" Dec 01 08:43:05 crc kubenswrapper[4813]: I1201 08:43:05.643805 4813 generic.go:334] "Generic (PLEG): container finished" podID="3c0b52aa-86ba-4c56-b23e-108cdb4750d9" containerID="6b60e03d0c7d656754adbb6ea0151c08f5c7d0d176006fe073fcf041604580f2" exitCode=0 Dec 01 08:43:05 crc kubenswrapper[4813]: I1201 08:43:05.643894 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-58jst" event={"ID":"3c0b52aa-86ba-4c56-b23e-108cdb4750d9","Type":"ContainerDied","Data":"6b60e03d0c7d656754adbb6ea0151c08f5c7d0d176006fe073fcf041604580f2"} Dec 01 08:43:05 crc kubenswrapper[4813]: I1201 08:43:05.648243 4813 generic.go:334] "Generic (PLEG): container finished" podID="0959ee73-31e7-4d58-b124-3e971d777225" containerID="bd47a660e132f743b3808aa5e85b3bbd8b48162275f7f73041d0564054db8d60" exitCode=0 Dec 01 08:43:05 crc kubenswrapper[4813]: I1201 08:43:05.648280 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7fqxs" event={"ID":"0959ee73-31e7-4d58-b124-3e971d777225","Type":"ContainerDied","Data":"bd47a660e132f743b3808aa5e85b3bbd8b48162275f7f73041d0564054db8d60"} Dec 01 08:43:05 crc kubenswrapper[4813]: I1201 08:43:05.653601 4813 generic.go:334] "Generic (PLEG): container finished" podID="979ea876-4dd3-4d31-9144-7a9a6e914c67" containerID="45a64db69ae0dfec8da00fa2f2586f5c5abb6678585119474fe56a698206b546" exitCode=0 Dec 01 08:43:05 crc kubenswrapper[4813]: I1201 08:43:05.653713 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wbvcp" event={"ID":"979ea876-4dd3-4d31-9144-7a9a6e914c67","Type":"ContainerDied","Data":"45a64db69ae0dfec8da00fa2f2586f5c5abb6678585119474fe56a698206b546"} Dec 01 08:43:05 crc kubenswrapper[4813]: I1201 08:43:05.655715 4813 generic.go:334] "Generic (PLEG): container finished" podID="26f4804b-998f-488c-a501-d1cc4ef7f0b4" containerID="7650b8743cec1bafe177f09abd09ceb115bff7266e4465d450482e7dab874dd4" exitCode=0 Dec 01 08:43:05 crc kubenswrapper[4813]: I1201 08:43:05.655797 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qh5hd" event={"ID":"26f4804b-998f-488c-a501-d1cc4ef7f0b4","Type":"ContainerDied","Data":"7650b8743cec1bafe177f09abd09ceb115bff7266e4465d450482e7dab874dd4"} Dec 01 08:43:05 crc kubenswrapper[4813]: I1201 08:43:05.662228 4813 generic.go:334] "Generic (PLEG): container finished" podID="95d1024c-5b73-40ae-9460-29502da48a23" containerID="45fbeac15bf94fa9f8a5b1e8f064eaa32720a0ea44cdb1c52b73bd2edfb4d1d1" exitCode=0 Dec 01 08:43:05 crc kubenswrapper[4813]: I1201 08:43:05.662295 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8wwz4" event={"ID":"95d1024c-5b73-40ae-9460-29502da48a23","Type":"ContainerDied","Data":"45fbeac15bf94fa9f8a5b1e8f064eaa32720a0ea44cdb1c52b73bd2edfb4d1d1"} Dec 01 08:43:06 crc kubenswrapper[4813]: I1201 08:43:06.411441 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd02bd16-7f86-48ad-b554-6404f62b1b28" path="/var/lib/kubelet/pods/bd02bd16-7f86-48ad-b554-6404f62b1b28/volumes" Dec 01 08:43:06 crc kubenswrapper[4813]: I1201 08:43:06.669943 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qh5hd" event={"ID":"26f4804b-998f-488c-a501-d1cc4ef7f0b4","Type":"ContainerStarted","Data":"1acc7744b9f14eab07ba59bc0c898cc8e8c944a526dd3599948ffbbaf19aabfd"} Dec 01 08:43:06 crc kubenswrapper[4813]: I1201 08:43:06.673228 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8wwz4" event={"ID":"95d1024c-5b73-40ae-9460-29502da48a23","Type":"ContainerStarted","Data":"a79b9174e0e8ddb88c32638d02d30d3c694c35e25b59d65387c7ab7ded854f1d"} Dec 01 08:43:06 crc kubenswrapper[4813]: I1201 08:43:06.675545 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-58jst" event={"ID":"3c0b52aa-86ba-4c56-b23e-108cdb4750d9","Type":"ContainerStarted","Data":"65b04048cd8d26b92bd90bdc1ca981ed3e74ff90736aa5dd8fe87e7ba637245a"} Dec 01 08:43:06 crc kubenswrapper[4813]: I1201 08:43:06.678473 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7fqxs" event={"ID":"0959ee73-31e7-4d58-b124-3e971d777225","Type":"ContainerStarted","Data":"e33294dfa539d5c4653c0b4bf83d01a72b7e2848567f6ec9f836d922d084cbd4"} Dec 01 08:43:06 crc kubenswrapper[4813]: I1201 08:43:06.681329 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wbvcp" event={"ID":"979ea876-4dd3-4d31-9144-7a9a6e914c67","Type":"ContainerStarted","Data":"b449598529a2bf89bc7d1b288f3aef72364ec65d95f4efa1630f2e0298648fcb"} Dec 01 08:43:06 crc kubenswrapper[4813]: I1201 08:43:06.691009 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-qh5hd" Dec 01 08:43:06 crc kubenswrapper[4813]: I1201 08:43:06.691089 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-qh5hd" Dec 01 08:43:06 crc kubenswrapper[4813]: I1201 08:43:06.697349 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-qh5hd" podStartSLOduration=3.984718869 podStartE2EDuration="1m1.697331436s" podCreationTimestamp="2025-12-01 08:42:05 +0000 UTC" firstStartedPulling="2025-12-01 08:42:08.398437226 +0000 UTC m=+113.021358822" lastFinishedPulling="2025-12-01 08:43:06.111049803 +0000 UTC m=+170.733971389" observedRunningTime="2025-12-01 08:43:06.697231793 +0000 UTC m=+171.320153379" watchObservedRunningTime="2025-12-01 08:43:06.697331436 +0000 UTC m=+171.320253022" Dec 01 08:43:06 crc kubenswrapper[4813]: I1201 08:43:06.724096 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-58jst" podStartSLOduration=3.342833663 podStartE2EDuration="57.72407763s" podCreationTimestamp="2025-12-01 08:42:09 +0000 UTC" firstStartedPulling="2025-12-01 08:42:11.698121 +0000 UTC m=+116.321042586" lastFinishedPulling="2025-12-01 08:43:06.079364967 +0000 UTC m=+170.702286553" observedRunningTime="2025-12-01 08:43:06.723081159 +0000 UTC m=+171.346002745" watchObservedRunningTime="2025-12-01 08:43:06.72407763 +0000 UTC m=+171.346999216" Dec 01 08:43:06 crc kubenswrapper[4813]: I1201 08:43:06.742505 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-7fqxs" podStartSLOduration=2.947565528 podStartE2EDuration="1m0.742477997s" podCreationTimestamp="2025-12-01 08:42:06 +0000 UTC" firstStartedPulling="2025-12-01 08:42:08.377429132 +0000 UTC m=+113.000350718" lastFinishedPulling="2025-12-01 08:43:06.172341601 +0000 UTC m=+170.795263187" observedRunningTime="2025-12-01 08:43:06.740997401 +0000 UTC m=+171.363918977" watchObservedRunningTime="2025-12-01 08:43:06.742477997 +0000 UTC m=+171.365399583" Dec 01 08:43:06 crc kubenswrapper[4813]: I1201 08:43:06.754955 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-7fqxs" Dec 01 08:43:06 crc kubenswrapper[4813]: I1201 08:43:06.755050 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-7fqxs" Dec 01 08:43:06 crc kubenswrapper[4813]: I1201 08:43:06.755052 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-wbvcp" podStartSLOduration=3.3752012479999998 podStartE2EDuration="57.755034194s" podCreationTimestamp="2025-12-01 08:42:09 +0000 UTC" firstStartedPulling="2025-12-01 08:42:11.713404458 +0000 UTC m=+116.336326044" lastFinishedPulling="2025-12-01 08:43:06.093237404 +0000 UTC m=+170.716158990" observedRunningTime="2025-12-01 08:43:06.75327888 +0000 UTC m=+171.376200466" watchObservedRunningTime="2025-12-01 08:43:06.755034194 +0000 UTC m=+171.377955780" Dec 01 08:43:06 crc kubenswrapper[4813]: I1201 08:43:06.768447 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-8wwz4" podStartSLOduration=3.882394557 podStartE2EDuration="1m0.768427076s" podCreationTimestamp="2025-12-01 08:42:06 +0000 UTC" firstStartedPulling="2025-12-01 08:42:09.472748086 +0000 UTC m=+114.095669672" lastFinishedPulling="2025-12-01 08:43:06.358780605 +0000 UTC m=+170.981702191" observedRunningTime="2025-12-01 08:43:06.767272461 +0000 UTC m=+171.390194057" watchObservedRunningTime="2025-12-01 08:43:06.768427076 +0000 UTC m=+171.391348662" Dec 01 08:43:07 crc kubenswrapper[4813]: I1201 08:43:07.410179 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-8wwz4" Dec 01 08:43:07 crc kubenswrapper[4813]: I1201 08:43:07.410238 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-8wwz4" Dec 01 08:43:07 crc kubenswrapper[4813]: I1201 08:43:07.956495 4813 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-qh5hd" podUID="26f4804b-998f-488c-a501-d1cc4ef7f0b4" containerName="registry-server" probeResult="failure" output=< Dec 01 08:43:07 crc kubenswrapper[4813]: timeout: failed to connect service ":50051" within 1s Dec 01 08:43:07 crc kubenswrapper[4813]: > Dec 01 08:43:07 crc kubenswrapper[4813]: I1201 08:43:07.963478 4813 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-7fqxs" podUID="0959ee73-31e7-4d58-b124-3e971d777225" containerName="registry-server" probeResult="failure" output=< Dec 01 08:43:07 crc kubenswrapper[4813]: timeout: failed to connect service ":50051" within 1s Dec 01 08:43:07 crc kubenswrapper[4813]: > Dec 01 08:43:08 crc kubenswrapper[4813]: I1201 08:43:08.329519 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-lk5db" Dec 01 08:43:08 crc kubenswrapper[4813]: I1201 08:43:08.329566 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-lk5db" Dec 01 08:43:08 crc kubenswrapper[4813]: I1201 08:43:08.372491 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-lk5db" Dec 01 08:43:08 crc kubenswrapper[4813]: I1201 08:43:08.453145 4813 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-8wwz4" podUID="95d1024c-5b73-40ae-9460-29502da48a23" containerName="registry-server" probeResult="failure" output=< Dec 01 08:43:08 crc kubenswrapper[4813]: timeout: failed to connect service ":50051" within 1s Dec 01 08:43:08 crc kubenswrapper[4813]: > Dec 01 08:43:09 crc kubenswrapper[4813]: I1201 08:43:09.661638 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-wbvcp" Dec 01 08:43:09 crc kubenswrapper[4813]: I1201 08:43:09.662003 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-wbvcp" Dec 01 08:43:10 crc kubenswrapper[4813]: I1201 08:43:10.022549 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-lk5db" Dec 01 08:43:10 crc kubenswrapper[4813]: I1201 08:43:10.108023 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-58jst" Dec 01 08:43:10 crc kubenswrapper[4813]: I1201 08:43:10.108083 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-58jst" Dec 01 08:43:10 crc kubenswrapper[4813]: I1201 08:43:10.712538 4813 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-wbvcp" podUID="979ea876-4dd3-4d31-9144-7a9a6e914c67" containerName="registry-server" probeResult="failure" output=< Dec 01 08:43:10 crc kubenswrapper[4813]: timeout: failed to connect service ":50051" within 1s Dec 01 08:43:10 crc kubenswrapper[4813]: > Dec 01 08:43:11 crc kubenswrapper[4813]: I1201 08:43:11.149091 4813 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-58jst" podUID="3c0b52aa-86ba-4c56-b23e-108cdb4750d9" containerName="registry-server" probeResult="failure" output=< Dec 01 08:43:11 crc kubenswrapper[4813]: timeout: failed to connect service ":50051" within 1s Dec 01 08:43:11 crc kubenswrapper[4813]: > Dec 01 08:43:16 crc kubenswrapper[4813]: I1201 08:43:16.738998 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-qh5hd" Dec 01 08:43:16 crc kubenswrapper[4813]: I1201 08:43:16.778665 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-qh5hd" Dec 01 08:43:16 crc kubenswrapper[4813]: I1201 08:43:16.815705 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-7fqxs" Dec 01 08:43:16 crc kubenswrapper[4813]: I1201 08:43:16.852100 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-7fqxs" Dec 01 08:43:17 crc kubenswrapper[4813]: I1201 08:43:17.223573 4813 patch_prober.go:28] interesting pod/machine-config-daemon-9wn4q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 08:43:17 crc kubenswrapper[4813]: I1201 08:43:17.223917 4813 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9wn4q" podUID="3d423402-e08f-42e9-b60c-6eb9bf2f6d48" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 08:43:17 crc kubenswrapper[4813]: I1201 08:43:17.450068 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-8wwz4" Dec 01 08:43:17 crc kubenswrapper[4813]: I1201 08:43:17.489229 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-8wwz4" Dec 01 08:43:18 crc kubenswrapper[4813]: I1201 08:43:18.006271 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-8wwz4"] Dec 01 08:43:18 crc kubenswrapper[4813]: I1201 08:43:18.443557 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-5fgn9" podUID="52120ed5-ff4f-4df5-8c60-3342df484be8" containerName="oauth-openshift" containerID="cri-o://fad7917420effa27fbb37fc9bc2b08f930b871c20e49fe40a368976ec808175f" gracePeriod=15 Dec 01 08:43:19 crc kubenswrapper[4813]: I1201 08:43:19.036059 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-8wwz4" podUID="95d1024c-5b73-40ae-9460-29502da48a23" containerName="registry-server" containerID="cri-o://a79b9174e0e8ddb88c32638d02d30d3c694c35e25b59d65387c7ab7ded854f1d" gracePeriod=2 Dec 01 08:43:19 crc kubenswrapper[4813]: I1201 08:43:19.711101 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-wbvcp" Dec 01 08:43:19 crc kubenswrapper[4813]: I1201 08:43:19.764811 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-wbvcp" Dec 01 08:43:20 crc kubenswrapper[4813]: I1201 08:43:20.155855 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-58jst" Dec 01 08:43:20 crc kubenswrapper[4813]: I1201 08:43:20.203802 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-58jst" Dec 01 08:43:20 crc kubenswrapper[4813]: I1201 08:43:20.638343 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-8wwz4" Dec 01 08:43:20 crc kubenswrapper[4813]: I1201 08:43:20.766030 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/95d1024c-5b73-40ae-9460-29502da48a23-utilities\") pod \"95d1024c-5b73-40ae-9460-29502da48a23\" (UID: \"95d1024c-5b73-40ae-9460-29502da48a23\") " Dec 01 08:43:20 crc kubenswrapper[4813]: I1201 08:43:20.766119 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/95d1024c-5b73-40ae-9460-29502da48a23-catalog-content\") pod \"95d1024c-5b73-40ae-9460-29502da48a23\" (UID: \"95d1024c-5b73-40ae-9460-29502da48a23\") " Dec 01 08:43:20 crc kubenswrapper[4813]: I1201 08:43:20.766175 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xjp26\" (UniqueName: \"kubernetes.io/projected/95d1024c-5b73-40ae-9460-29502da48a23-kube-api-access-xjp26\") pod \"95d1024c-5b73-40ae-9460-29502da48a23\" (UID: \"95d1024c-5b73-40ae-9460-29502da48a23\") " Dec 01 08:43:20 crc kubenswrapper[4813]: I1201 08:43:20.767084 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/95d1024c-5b73-40ae-9460-29502da48a23-utilities" (OuterVolumeSpecName: "utilities") pod "95d1024c-5b73-40ae-9460-29502da48a23" (UID: "95d1024c-5b73-40ae-9460-29502da48a23"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 08:43:20 crc kubenswrapper[4813]: I1201 08:43:20.771081 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/95d1024c-5b73-40ae-9460-29502da48a23-kube-api-access-xjp26" (OuterVolumeSpecName: "kube-api-access-xjp26") pod "95d1024c-5b73-40ae-9460-29502da48a23" (UID: "95d1024c-5b73-40ae-9460-29502da48a23"). InnerVolumeSpecName "kube-api-access-xjp26". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:43:20 crc kubenswrapper[4813]: I1201 08:43:20.826679 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/95d1024c-5b73-40ae-9460-29502da48a23-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "95d1024c-5b73-40ae-9460-29502da48a23" (UID: "95d1024c-5b73-40ae-9460-29502da48a23"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 08:43:20 crc kubenswrapper[4813]: I1201 08:43:20.867815 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xjp26\" (UniqueName: \"kubernetes.io/projected/95d1024c-5b73-40ae-9460-29502da48a23-kube-api-access-xjp26\") on node \"crc\" DevicePath \"\"" Dec 01 08:43:20 crc kubenswrapper[4813]: I1201 08:43:20.868067 4813 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/95d1024c-5b73-40ae-9460-29502da48a23-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 08:43:20 crc kubenswrapper[4813]: I1201 08:43:20.868220 4813 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/95d1024c-5b73-40ae-9460-29502da48a23-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 08:43:20 crc kubenswrapper[4813]: I1201 08:43:20.874665 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-5fgn9" Dec 01 08:43:21 crc kubenswrapper[4813]: I1201 08:43:21.050464 4813 generic.go:334] "Generic (PLEG): container finished" podID="95d1024c-5b73-40ae-9460-29502da48a23" containerID="a79b9174e0e8ddb88c32638d02d30d3c694c35e25b59d65387c7ab7ded854f1d" exitCode=0 Dec 01 08:43:21 crc kubenswrapper[4813]: I1201 08:43:21.050546 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-8wwz4" Dec 01 08:43:21 crc kubenswrapper[4813]: I1201 08:43:21.050532 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8wwz4" event={"ID":"95d1024c-5b73-40ae-9460-29502da48a23","Type":"ContainerDied","Data":"a79b9174e0e8ddb88c32638d02d30d3c694c35e25b59d65387c7ab7ded854f1d"} Dec 01 08:43:21 crc kubenswrapper[4813]: I1201 08:43:21.051154 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8wwz4" event={"ID":"95d1024c-5b73-40ae-9460-29502da48a23","Type":"ContainerDied","Data":"efc0b55bd67f818bfb6a18d4781f9109c8f5b61e101e6974dc09069dab49127b"} Dec 01 08:43:21 crc kubenswrapper[4813]: I1201 08:43:21.051202 4813 scope.go:117] "RemoveContainer" containerID="a79b9174e0e8ddb88c32638d02d30d3c694c35e25b59d65387c7ab7ded854f1d" Dec 01 08:43:21 crc kubenswrapper[4813]: I1201 08:43:21.052389 4813 generic.go:334] "Generic (PLEG): container finished" podID="52120ed5-ff4f-4df5-8c60-3342df484be8" containerID="fad7917420effa27fbb37fc9bc2b08f930b871c20e49fe40a368976ec808175f" exitCode=0 Dec 01 08:43:21 crc kubenswrapper[4813]: I1201 08:43:21.052436 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-5fgn9" Dec 01 08:43:21 crc kubenswrapper[4813]: I1201 08:43:21.052472 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-5fgn9" event={"ID":"52120ed5-ff4f-4df5-8c60-3342df484be8","Type":"ContainerDied","Data":"fad7917420effa27fbb37fc9bc2b08f930b871c20e49fe40a368976ec808175f"} Dec 01 08:43:21 crc kubenswrapper[4813]: I1201 08:43:21.052702 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-5fgn9" event={"ID":"52120ed5-ff4f-4df5-8c60-3342df484be8","Type":"ContainerDied","Data":"8cf41b039137a394ada809bf9390a7178c63a69753f291865d30047a66675172"} Dec 01 08:43:21 crc kubenswrapper[4813]: I1201 08:43:21.068670 4813 scope.go:117] "RemoveContainer" containerID="45fbeac15bf94fa9f8a5b1e8f064eaa32720a0ea44cdb1c52b73bd2edfb4d1d1" Dec 01 08:43:21 crc kubenswrapper[4813]: I1201 08:43:21.069740 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/52120ed5-ff4f-4df5-8c60-3342df484be8-v4-0-config-system-router-certs\") pod \"52120ed5-ff4f-4df5-8c60-3342df484be8\" (UID: \"52120ed5-ff4f-4df5-8c60-3342df484be8\") " Dec 01 08:43:21 crc kubenswrapper[4813]: I1201 08:43:21.069824 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/52120ed5-ff4f-4df5-8c60-3342df484be8-v4-0-config-system-trusted-ca-bundle\") pod \"52120ed5-ff4f-4df5-8c60-3342df484be8\" (UID: \"52120ed5-ff4f-4df5-8c60-3342df484be8\") " Dec 01 08:43:21 crc kubenswrapper[4813]: I1201 08:43:21.069873 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/52120ed5-ff4f-4df5-8c60-3342df484be8-v4-0-config-system-service-ca\") pod \"52120ed5-ff4f-4df5-8c60-3342df484be8\" (UID: \"52120ed5-ff4f-4df5-8c60-3342df484be8\") " Dec 01 08:43:21 crc kubenswrapper[4813]: I1201 08:43:21.069905 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/52120ed5-ff4f-4df5-8c60-3342df484be8-v4-0-config-system-session\") pod \"52120ed5-ff4f-4df5-8c60-3342df484be8\" (UID: \"52120ed5-ff4f-4df5-8c60-3342df484be8\") " Dec 01 08:43:21 crc kubenswrapper[4813]: I1201 08:43:21.069934 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/52120ed5-ff4f-4df5-8c60-3342df484be8-v4-0-config-system-serving-cert\") pod \"52120ed5-ff4f-4df5-8c60-3342df484be8\" (UID: \"52120ed5-ff4f-4df5-8c60-3342df484be8\") " Dec 01 08:43:21 crc kubenswrapper[4813]: I1201 08:43:21.069993 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lmpxp\" (UniqueName: \"kubernetes.io/projected/52120ed5-ff4f-4df5-8c60-3342df484be8-kube-api-access-lmpxp\") pod \"52120ed5-ff4f-4df5-8c60-3342df484be8\" (UID: \"52120ed5-ff4f-4df5-8c60-3342df484be8\") " Dec 01 08:43:21 crc kubenswrapper[4813]: I1201 08:43:21.070026 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/52120ed5-ff4f-4df5-8c60-3342df484be8-v4-0-config-user-idp-0-file-data\") pod \"52120ed5-ff4f-4df5-8c60-3342df484be8\" (UID: \"52120ed5-ff4f-4df5-8c60-3342df484be8\") " Dec 01 08:43:21 crc kubenswrapper[4813]: I1201 08:43:21.070052 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/52120ed5-ff4f-4df5-8c60-3342df484be8-audit-dir\") pod \"52120ed5-ff4f-4df5-8c60-3342df484be8\" (UID: \"52120ed5-ff4f-4df5-8c60-3342df484be8\") " Dec 01 08:43:21 crc kubenswrapper[4813]: I1201 08:43:21.070079 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/52120ed5-ff4f-4df5-8c60-3342df484be8-v4-0-config-user-template-provider-selection\") pod \"52120ed5-ff4f-4df5-8c60-3342df484be8\" (UID: \"52120ed5-ff4f-4df5-8c60-3342df484be8\") " Dec 01 08:43:21 crc kubenswrapper[4813]: I1201 08:43:21.070122 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/52120ed5-ff4f-4df5-8c60-3342df484be8-v4-0-config-user-template-error\") pod \"52120ed5-ff4f-4df5-8c60-3342df484be8\" (UID: \"52120ed5-ff4f-4df5-8c60-3342df484be8\") " Dec 01 08:43:21 crc kubenswrapper[4813]: I1201 08:43:21.070173 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/52120ed5-ff4f-4df5-8c60-3342df484be8-v4-0-config-system-cliconfig\") pod \"52120ed5-ff4f-4df5-8c60-3342df484be8\" (UID: \"52120ed5-ff4f-4df5-8c60-3342df484be8\") " Dec 01 08:43:21 crc kubenswrapper[4813]: I1201 08:43:21.070202 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/52120ed5-ff4f-4df5-8c60-3342df484be8-v4-0-config-system-ocp-branding-template\") pod \"52120ed5-ff4f-4df5-8c60-3342df484be8\" (UID: \"52120ed5-ff4f-4df5-8c60-3342df484be8\") " Dec 01 08:43:21 crc kubenswrapper[4813]: I1201 08:43:21.070243 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/52120ed5-ff4f-4df5-8c60-3342df484be8-v4-0-config-user-template-login\") pod \"52120ed5-ff4f-4df5-8c60-3342df484be8\" (UID: \"52120ed5-ff4f-4df5-8c60-3342df484be8\") " Dec 01 08:43:21 crc kubenswrapper[4813]: I1201 08:43:21.070285 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/52120ed5-ff4f-4df5-8c60-3342df484be8-audit-policies\") pod \"52120ed5-ff4f-4df5-8c60-3342df484be8\" (UID: \"52120ed5-ff4f-4df5-8c60-3342df484be8\") " Dec 01 08:43:21 crc kubenswrapper[4813]: I1201 08:43:21.071332 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/52120ed5-ff4f-4df5-8c60-3342df484be8-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "52120ed5-ff4f-4df5-8c60-3342df484be8" (UID: "52120ed5-ff4f-4df5-8c60-3342df484be8"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:43:21 crc kubenswrapper[4813]: I1201 08:43:21.071381 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/52120ed5-ff4f-4df5-8c60-3342df484be8-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "52120ed5-ff4f-4df5-8c60-3342df484be8" (UID: "52120ed5-ff4f-4df5-8c60-3342df484be8"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:43:21 crc kubenswrapper[4813]: I1201 08:43:21.071452 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/52120ed5-ff4f-4df5-8c60-3342df484be8-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "52120ed5-ff4f-4df5-8c60-3342df484be8" (UID: "52120ed5-ff4f-4df5-8c60-3342df484be8"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:43:21 crc kubenswrapper[4813]: I1201 08:43:21.071492 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/52120ed5-ff4f-4df5-8c60-3342df484be8-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "52120ed5-ff4f-4df5-8c60-3342df484be8" (UID: "52120ed5-ff4f-4df5-8c60-3342df484be8"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 08:43:21 crc kubenswrapper[4813]: I1201 08:43:21.072797 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/52120ed5-ff4f-4df5-8c60-3342df484be8-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "52120ed5-ff4f-4df5-8c60-3342df484be8" (UID: "52120ed5-ff4f-4df5-8c60-3342df484be8"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:43:21 crc kubenswrapper[4813]: I1201 08:43:21.073847 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/52120ed5-ff4f-4df5-8c60-3342df484be8-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "52120ed5-ff4f-4df5-8c60-3342df484be8" (UID: "52120ed5-ff4f-4df5-8c60-3342df484be8"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:43:21 crc kubenswrapper[4813]: I1201 08:43:21.076032 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/52120ed5-ff4f-4df5-8c60-3342df484be8-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "52120ed5-ff4f-4df5-8c60-3342df484be8" (UID: "52120ed5-ff4f-4df5-8c60-3342df484be8"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:43:21 crc kubenswrapper[4813]: I1201 08:43:21.076441 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/52120ed5-ff4f-4df5-8c60-3342df484be8-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "52120ed5-ff4f-4df5-8c60-3342df484be8" (UID: "52120ed5-ff4f-4df5-8c60-3342df484be8"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:43:21 crc kubenswrapper[4813]: I1201 08:43:21.077106 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/52120ed5-ff4f-4df5-8c60-3342df484be8-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "52120ed5-ff4f-4df5-8c60-3342df484be8" (UID: "52120ed5-ff4f-4df5-8c60-3342df484be8"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:43:21 crc kubenswrapper[4813]: I1201 08:43:21.077584 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/52120ed5-ff4f-4df5-8c60-3342df484be8-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "52120ed5-ff4f-4df5-8c60-3342df484be8" (UID: "52120ed5-ff4f-4df5-8c60-3342df484be8"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:43:21 crc kubenswrapper[4813]: I1201 08:43:21.082853 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-8wwz4"] Dec 01 08:43:21 crc kubenswrapper[4813]: I1201 08:43:21.089659 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-8wwz4"] Dec 01 08:43:21 crc kubenswrapper[4813]: I1201 08:43:21.092274 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/52120ed5-ff4f-4df5-8c60-3342df484be8-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "52120ed5-ff4f-4df5-8c60-3342df484be8" (UID: "52120ed5-ff4f-4df5-8c60-3342df484be8"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:43:21 crc kubenswrapper[4813]: I1201 08:43:21.109151 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/52120ed5-ff4f-4df5-8c60-3342df484be8-kube-api-access-lmpxp" (OuterVolumeSpecName: "kube-api-access-lmpxp") pod "52120ed5-ff4f-4df5-8c60-3342df484be8" (UID: "52120ed5-ff4f-4df5-8c60-3342df484be8"). InnerVolumeSpecName "kube-api-access-lmpxp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:43:21 crc kubenswrapper[4813]: I1201 08:43:21.110187 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/52120ed5-ff4f-4df5-8c60-3342df484be8-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "52120ed5-ff4f-4df5-8c60-3342df484be8" (UID: "52120ed5-ff4f-4df5-8c60-3342df484be8"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:43:21 crc kubenswrapper[4813]: I1201 08:43:21.110233 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/52120ed5-ff4f-4df5-8c60-3342df484be8-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "52120ed5-ff4f-4df5-8c60-3342df484be8" (UID: "52120ed5-ff4f-4df5-8c60-3342df484be8"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:43:21 crc kubenswrapper[4813]: I1201 08:43:21.115354 4813 scope.go:117] "RemoveContainer" containerID="ebe4667ac0b69195462fd75dcaf7de45a48aa09d477a4f482dca8336cf604cf5" Dec 01 08:43:21 crc kubenswrapper[4813]: I1201 08:43:21.128827 4813 scope.go:117] "RemoveContainer" containerID="a79b9174e0e8ddb88c32638d02d30d3c694c35e25b59d65387c7ab7ded854f1d" Dec 01 08:43:21 crc kubenswrapper[4813]: E1201 08:43:21.129470 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a79b9174e0e8ddb88c32638d02d30d3c694c35e25b59d65387c7ab7ded854f1d\": container with ID starting with a79b9174e0e8ddb88c32638d02d30d3c694c35e25b59d65387c7ab7ded854f1d not found: ID does not exist" containerID="a79b9174e0e8ddb88c32638d02d30d3c694c35e25b59d65387c7ab7ded854f1d" Dec 01 08:43:21 crc kubenswrapper[4813]: I1201 08:43:21.129509 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a79b9174e0e8ddb88c32638d02d30d3c694c35e25b59d65387c7ab7ded854f1d"} err="failed to get container status \"a79b9174e0e8ddb88c32638d02d30d3c694c35e25b59d65387c7ab7ded854f1d\": rpc error: code = NotFound desc = could not find container \"a79b9174e0e8ddb88c32638d02d30d3c694c35e25b59d65387c7ab7ded854f1d\": container with ID starting with a79b9174e0e8ddb88c32638d02d30d3c694c35e25b59d65387c7ab7ded854f1d not found: ID does not exist" Dec 01 08:43:21 crc kubenswrapper[4813]: I1201 08:43:21.129532 4813 scope.go:117] "RemoveContainer" containerID="45fbeac15bf94fa9f8a5b1e8f064eaa32720a0ea44cdb1c52b73bd2edfb4d1d1" Dec 01 08:43:21 crc kubenswrapper[4813]: E1201 08:43:21.129797 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"45fbeac15bf94fa9f8a5b1e8f064eaa32720a0ea44cdb1c52b73bd2edfb4d1d1\": container with ID starting with 45fbeac15bf94fa9f8a5b1e8f064eaa32720a0ea44cdb1c52b73bd2edfb4d1d1 not found: ID does not exist" containerID="45fbeac15bf94fa9f8a5b1e8f064eaa32720a0ea44cdb1c52b73bd2edfb4d1d1" Dec 01 08:43:21 crc kubenswrapper[4813]: I1201 08:43:21.129913 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"45fbeac15bf94fa9f8a5b1e8f064eaa32720a0ea44cdb1c52b73bd2edfb4d1d1"} err="failed to get container status \"45fbeac15bf94fa9f8a5b1e8f064eaa32720a0ea44cdb1c52b73bd2edfb4d1d1\": rpc error: code = NotFound desc = could not find container \"45fbeac15bf94fa9f8a5b1e8f064eaa32720a0ea44cdb1c52b73bd2edfb4d1d1\": container with ID starting with 45fbeac15bf94fa9f8a5b1e8f064eaa32720a0ea44cdb1c52b73bd2edfb4d1d1 not found: ID does not exist" Dec 01 08:43:21 crc kubenswrapper[4813]: I1201 08:43:21.130084 4813 scope.go:117] "RemoveContainer" containerID="ebe4667ac0b69195462fd75dcaf7de45a48aa09d477a4f482dca8336cf604cf5" Dec 01 08:43:21 crc kubenswrapper[4813]: E1201 08:43:21.130472 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ebe4667ac0b69195462fd75dcaf7de45a48aa09d477a4f482dca8336cf604cf5\": container with ID starting with ebe4667ac0b69195462fd75dcaf7de45a48aa09d477a4f482dca8336cf604cf5 not found: ID does not exist" containerID="ebe4667ac0b69195462fd75dcaf7de45a48aa09d477a4f482dca8336cf604cf5" Dec 01 08:43:21 crc kubenswrapper[4813]: I1201 08:43:21.130493 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ebe4667ac0b69195462fd75dcaf7de45a48aa09d477a4f482dca8336cf604cf5"} err="failed to get container status \"ebe4667ac0b69195462fd75dcaf7de45a48aa09d477a4f482dca8336cf604cf5\": rpc error: code = NotFound desc = could not find container \"ebe4667ac0b69195462fd75dcaf7de45a48aa09d477a4f482dca8336cf604cf5\": container with ID starting with ebe4667ac0b69195462fd75dcaf7de45a48aa09d477a4f482dca8336cf604cf5 not found: ID does not exist" Dec 01 08:43:21 crc kubenswrapper[4813]: I1201 08:43:21.130507 4813 scope.go:117] "RemoveContainer" containerID="fad7917420effa27fbb37fc9bc2b08f930b871c20e49fe40a368976ec808175f" Dec 01 08:43:21 crc kubenswrapper[4813]: I1201 08:43:21.150516 4813 scope.go:117] "RemoveContainer" containerID="fad7917420effa27fbb37fc9bc2b08f930b871c20e49fe40a368976ec808175f" Dec 01 08:43:21 crc kubenswrapper[4813]: E1201 08:43:21.151148 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fad7917420effa27fbb37fc9bc2b08f930b871c20e49fe40a368976ec808175f\": container with ID starting with fad7917420effa27fbb37fc9bc2b08f930b871c20e49fe40a368976ec808175f not found: ID does not exist" containerID="fad7917420effa27fbb37fc9bc2b08f930b871c20e49fe40a368976ec808175f" Dec 01 08:43:21 crc kubenswrapper[4813]: I1201 08:43:21.151194 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fad7917420effa27fbb37fc9bc2b08f930b871c20e49fe40a368976ec808175f"} err="failed to get container status \"fad7917420effa27fbb37fc9bc2b08f930b871c20e49fe40a368976ec808175f\": rpc error: code = NotFound desc = could not find container \"fad7917420effa27fbb37fc9bc2b08f930b871c20e49fe40a368976ec808175f\": container with ID starting with fad7917420effa27fbb37fc9bc2b08f930b871c20e49fe40a368976ec808175f not found: ID does not exist" Dec 01 08:43:21 crc kubenswrapper[4813]: I1201 08:43:21.171444 4813 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/52120ed5-ff4f-4df5-8c60-3342df484be8-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 08:43:21 crc kubenswrapper[4813]: I1201 08:43:21.172021 4813 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/52120ed5-ff4f-4df5-8c60-3342df484be8-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Dec 01 08:43:21 crc kubenswrapper[4813]: I1201 08:43:21.172142 4813 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/52120ed5-ff4f-4df5-8c60-3342df484be8-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Dec 01 08:43:21 crc kubenswrapper[4813]: I1201 08:43:21.172230 4813 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/52120ed5-ff4f-4df5-8c60-3342df484be8-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 01 08:43:21 crc kubenswrapper[4813]: I1201 08:43:21.172312 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lmpxp\" (UniqueName: \"kubernetes.io/projected/52120ed5-ff4f-4df5-8c60-3342df484be8-kube-api-access-lmpxp\") on node \"crc\" DevicePath \"\"" Dec 01 08:43:21 crc kubenswrapper[4813]: I1201 08:43:21.172396 4813 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/52120ed5-ff4f-4df5-8c60-3342df484be8-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Dec 01 08:43:21 crc kubenswrapper[4813]: I1201 08:43:21.172482 4813 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/52120ed5-ff4f-4df5-8c60-3342df484be8-audit-dir\") on node \"crc\" DevicePath \"\"" Dec 01 08:43:21 crc kubenswrapper[4813]: I1201 08:43:21.172558 4813 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/52120ed5-ff4f-4df5-8c60-3342df484be8-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Dec 01 08:43:21 crc kubenswrapper[4813]: I1201 08:43:21.172643 4813 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/52120ed5-ff4f-4df5-8c60-3342df484be8-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Dec 01 08:43:21 crc kubenswrapper[4813]: I1201 08:43:21.172720 4813 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/52120ed5-ff4f-4df5-8c60-3342df484be8-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Dec 01 08:43:21 crc kubenswrapper[4813]: I1201 08:43:21.172829 4813 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/52120ed5-ff4f-4df5-8c60-3342df484be8-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Dec 01 08:43:21 crc kubenswrapper[4813]: I1201 08:43:21.172917 4813 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/52120ed5-ff4f-4df5-8c60-3342df484be8-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Dec 01 08:43:21 crc kubenswrapper[4813]: I1201 08:43:21.173039 4813 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/52120ed5-ff4f-4df5-8c60-3342df484be8-audit-policies\") on node \"crc\" DevicePath \"\"" Dec 01 08:43:21 crc kubenswrapper[4813]: I1201 08:43:21.173119 4813 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/52120ed5-ff4f-4df5-8c60-3342df484be8-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Dec 01 08:43:21 crc kubenswrapper[4813]: I1201 08:43:21.381657 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-5fgn9"] Dec 01 08:43:21 crc kubenswrapper[4813]: I1201 08:43:21.384116 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-5fgn9"] Dec 01 08:43:21 crc kubenswrapper[4813]: I1201 08:43:21.972737 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-58jst"] Dec 01 08:43:22 crc kubenswrapper[4813]: I1201 08:43:22.060924 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-58jst" podUID="3c0b52aa-86ba-4c56-b23e-108cdb4750d9" containerName="registry-server" containerID="cri-o://65b04048cd8d26b92bd90bdc1ca981ed3e74ff90736aa5dd8fe87e7ba637245a" gracePeriod=2 Dec 01 08:43:22 crc kubenswrapper[4813]: I1201 08:43:22.399598 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="52120ed5-ff4f-4df5-8c60-3342df484be8" path="/var/lib/kubelet/pods/52120ed5-ff4f-4df5-8c60-3342df484be8/volumes" Dec 01 08:43:22 crc kubenswrapper[4813]: I1201 08:43:22.400084 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="95d1024c-5b73-40ae-9460-29502da48a23" path="/var/lib/kubelet/pods/95d1024c-5b73-40ae-9460-29502da48a23/volumes" Dec 01 08:43:22 crc kubenswrapper[4813]: I1201 08:43:22.424150 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-58jst" Dec 01 08:43:22 crc kubenswrapper[4813]: I1201 08:43:22.593805 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3c0b52aa-86ba-4c56-b23e-108cdb4750d9-catalog-content\") pod \"3c0b52aa-86ba-4c56-b23e-108cdb4750d9\" (UID: \"3c0b52aa-86ba-4c56-b23e-108cdb4750d9\") " Dec 01 08:43:22 crc kubenswrapper[4813]: I1201 08:43:22.596175 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qbk44\" (UniqueName: \"kubernetes.io/projected/3c0b52aa-86ba-4c56-b23e-108cdb4750d9-kube-api-access-qbk44\") pod \"3c0b52aa-86ba-4c56-b23e-108cdb4750d9\" (UID: \"3c0b52aa-86ba-4c56-b23e-108cdb4750d9\") " Dec 01 08:43:22 crc kubenswrapper[4813]: I1201 08:43:22.596257 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3c0b52aa-86ba-4c56-b23e-108cdb4750d9-utilities\") pod \"3c0b52aa-86ba-4c56-b23e-108cdb4750d9\" (UID: \"3c0b52aa-86ba-4c56-b23e-108cdb4750d9\") " Dec 01 08:43:22 crc kubenswrapper[4813]: I1201 08:43:22.597034 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3c0b52aa-86ba-4c56-b23e-108cdb4750d9-utilities" (OuterVolumeSpecName: "utilities") pod "3c0b52aa-86ba-4c56-b23e-108cdb4750d9" (UID: "3c0b52aa-86ba-4c56-b23e-108cdb4750d9"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 08:43:22 crc kubenswrapper[4813]: I1201 08:43:22.603597 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3c0b52aa-86ba-4c56-b23e-108cdb4750d9-kube-api-access-qbk44" (OuterVolumeSpecName: "kube-api-access-qbk44") pod "3c0b52aa-86ba-4c56-b23e-108cdb4750d9" (UID: "3c0b52aa-86ba-4c56-b23e-108cdb4750d9"). InnerVolumeSpecName "kube-api-access-qbk44". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:43:22 crc kubenswrapper[4813]: I1201 08:43:22.698069 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qbk44\" (UniqueName: \"kubernetes.io/projected/3c0b52aa-86ba-4c56-b23e-108cdb4750d9-kube-api-access-qbk44\") on node \"crc\" DevicePath \"\"" Dec 01 08:43:22 crc kubenswrapper[4813]: I1201 08:43:22.698107 4813 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3c0b52aa-86ba-4c56-b23e-108cdb4750d9-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 08:43:22 crc kubenswrapper[4813]: I1201 08:43:22.727856 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3c0b52aa-86ba-4c56-b23e-108cdb4750d9-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3c0b52aa-86ba-4c56-b23e-108cdb4750d9" (UID: "3c0b52aa-86ba-4c56-b23e-108cdb4750d9"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 08:43:22 crc kubenswrapper[4813]: I1201 08:43:22.799185 4813 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3c0b52aa-86ba-4c56-b23e-108cdb4750d9-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 08:43:23 crc kubenswrapper[4813]: I1201 08:43:23.070033 4813 generic.go:334] "Generic (PLEG): container finished" podID="3c0b52aa-86ba-4c56-b23e-108cdb4750d9" containerID="65b04048cd8d26b92bd90bdc1ca981ed3e74ff90736aa5dd8fe87e7ba637245a" exitCode=0 Dec 01 08:43:23 crc kubenswrapper[4813]: I1201 08:43:23.070098 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-58jst" event={"ID":"3c0b52aa-86ba-4c56-b23e-108cdb4750d9","Type":"ContainerDied","Data":"65b04048cd8d26b92bd90bdc1ca981ed3e74ff90736aa5dd8fe87e7ba637245a"} Dec 01 08:43:23 crc kubenswrapper[4813]: I1201 08:43:23.070149 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-58jst" event={"ID":"3c0b52aa-86ba-4c56-b23e-108cdb4750d9","Type":"ContainerDied","Data":"ae96f422c1483c19efa4aa9d6d31da5bb7d236ffd7a4c817cf236212a4243543"} Dec 01 08:43:23 crc kubenswrapper[4813]: I1201 08:43:23.070183 4813 scope.go:117] "RemoveContainer" containerID="65b04048cd8d26b92bd90bdc1ca981ed3e74ff90736aa5dd8fe87e7ba637245a" Dec 01 08:43:23 crc kubenswrapper[4813]: I1201 08:43:23.070749 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-58jst" Dec 01 08:43:23 crc kubenswrapper[4813]: I1201 08:43:23.093212 4813 scope.go:117] "RemoveContainer" containerID="6b60e03d0c7d656754adbb6ea0151c08f5c7d0d176006fe073fcf041604580f2" Dec 01 08:43:23 crc kubenswrapper[4813]: I1201 08:43:23.115956 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-58jst"] Dec 01 08:43:23 crc kubenswrapper[4813]: I1201 08:43:23.120322 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-58jst"] Dec 01 08:43:23 crc kubenswrapper[4813]: I1201 08:43:23.125981 4813 scope.go:117] "RemoveContainer" containerID="d9ea5b6b0d37675df39b7b61235961ddc4686468313e264a4c8271242376f87c" Dec 01 08:43:23 crc kubenswrapper[4813]: I1201 08:43:23.145220 4813 scope.go:117] "RemoveContainer" containerID="65b04048cd8d26b92bd90bdc1ca981ed3e74ff90736aa5dd8fe87e7ba637245a" Dec 01 08:43:23 crc kubenswrapper[4813]: E1201 08:43:23.146563 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"65b04048cd8d26b92bd90bdc1ca981ed3e74ff90736aa5dd8fe87e7ba637245a\": container with ID starting with 65b04048cd8d26b92bd90bdc1ca981ed3e74ff90736aa5dd8fe87e7ba637245a not found: ID does not exist" containerID="65b04048cd8d26b92bd90bdc1ca981ed3e74ff90736aa5dd8fe87e7ba637245a" Dec 01 08:43:23 crc kubenswrapper[4813]: I1201 08:43:23.146632 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"65b04048cd8d26b92bd90bdc1ca981ed3e74ff90736aa5dd8fe87e7ba637245a"} err="failed to get container status \"65b04048cd8d26b92bd90bdc1ca981ed3e74ff90736aa5dd8fe87e7ba637245a\": rpc error: code = NotFound desc = could not find container \"65b04048cd8d26b92bd90bdc1ca981ed3e74ff90736aa5dd8fe87e7ba637245a\": container with ID starting with 65b04048cd8d26b92bd90bdc1ca981ed3e74ff90736aa5dd8fe87e7ba637245a not found: ID does not exist" Dec 01 08:43:23 crc kubenswrapper[4813]: I1201 08:43:23.146681 4813 scope.go:117] "RemoveContainer" containerID="6b60e03d0c7d656754adbb6ea0151c08f5c7d0d176006fe073fcf041604580f2" Dec 01 08:43:23 crc kubenswrapper[4813]: E1201 08:43:23.148406 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6b60e03d0c7d656754adbb6ea0151c08f5c7d0d176006fe073fcf041604580f2\": container with ID starting with 6b60e03d0c7d656754adbb6ea0151c08f5c7d0d176006fe073fcf041604580f2 not found: ID does not exist" containerID="6b60e03d0c7d656754adbb6ea0151c08f5c7d0d176006fe073fcf041604580f2" Dec 01 08:43:23 crc kubenswrapper[4813]: I1201 08:43:23.148536 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6b60e03d0c7d656754adbb6ea0151c08f5c7d0d176006fe073fcf041604580f2"} err="failed to get container status \"6b60e03d0c7d656754adbb6ea0151c08f5c7d0d176006fe073fcf041604580f2\": rpc error: code = NotFound desc = could not find container \"6b60e03d0c7d656754adbb6ea0151c08f5c7d0d176006fe073fcf041604580f2\": container with ID starting with 6b60e03d0c7d656754adbb6ea0151c08f5c7d0d176006fe073fcf041604580f2 not found: ID does not exist" Dec 01 08:43:23 crc kubenswrapper[4813]: I1201 08:43:23.148662 4813 scope.go:117] "RemoveContainer" containerID="d9ea5b6b0d37675df39b7b61235961ddc4686468313e264a4c8271242376f87c" Dec 01 08:43:23 crc kubenswrapper[4813]: E1201 08:43:23.149169 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d9ea5b6b0d37675df39b7b61235961ddc4686468313e264a4c8271242376f87c\": container with ID starting with d9ea5b6b0d37675df39b7b61235961ddc4686468313e264a4c8271242376f87c not found: ID does not exist" containerID="d9ea5b6b0d37675df39b7b61235961ddc4686468313e264a4c8271242376f87c" Dec 01 08:43:23 crc kubenswrapper[4813]: I1201 08:43:23.149294 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d9ea5b6b0d37675df39b7b61235961ddc4686468313e264a4c8271242376f87c"} err="failed to get container status \"d9ea5b6b0d37675df39b7b61235961ddc4686468313e264a4c8271242376f87c\": rpc error: code = NotFound desc = could not find container \"d9ea5b6b0d37675df39b7b61235961ddc4686468313e264a4c8271242376f87c\": container with ID starting with d9ea5b6b0d37675df39b7b61235961ddc4686468313e264a4c8271242376f87c not found: ID does not exist" Dec 01 08:43:24 crc kubenswrapper[4813]: I1201 08:43:24.400244 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3c0b52aa-86ba-4c56-b23e-108cdb4750d9" path="/var/lib/kubelet/pods/3c0b52aa-86ba-4c56-b23e-108cdb4750d9/volumes" Dec 01 08:43:27 crc kubenswrapper[4813]: I1201 08:43:27.037324 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 08:43:29 crc kubenswrapper[4813]: I1201 08:43:29.206934 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-65c4c84884-2vrp6"] Dec 01 08:43:29 crc kubenswrapper[4813]: E1201 08:43:29.207587 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="52120ed5-ff4f-4df5-8c60-3342df484be8" containerName="oauth-openshift" Dec 01 08:43:29 crc kubenswrapper[4813]: I1201 08:43:29.207609 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="52120ed5-ff4f-4df5-8c60-3342df484be8" containerName="oauth-openshift" Dec 01 08:43:29 crc kubenswrapper[4813]: E1201 08:43:29.207625 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3c0b52aa-86ba-4c56-b23e-108cdb4750d9" containerName="extract-utilities" Dec 01 08:43:29 crc kubenswrapper[4813]: I1201 08:43:29.207635 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="3c0b52aa-86ba-4c56-b23e-108cdb4750d9" containerName="extract-utilities" Dec 01 08:43:29 crc kubenswrapper[4813]: E1201 08:43:29.207654 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bd02bd16-7f86-48ad-b554-6404f62b1b28" containerName="extract-content" Dec 01 08:43:29 crc kubenswrapper[4813]: I1201 08:43:29.207662 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="bd02bd16-7f86-48ad-b554-6404f62b1b28" containerName="extract-content" Dec 01 08:43:29 crc kubenswrapper[4813]: E1201 08:43:29.207675 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="95d1024c-5b73-40ae-9460-29502da48a23" containerName="registry-server" Dec 01 08:43:29 crc kubenswrapper[4813]: I1201 08:43:29.207683 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="95d1024c-5b73-40ae-9460-29502da48a23" containerName="registry-server" Dec 01 08:43:29 crc kubenswrapper[4813]: E1201 08:43:29.207695 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a81d6f9c-e9f7-4c98-8233-c9f214978bfa" containerName="extract-content" Dec 01 08:43:29 crc kubenswrapper[4813]: I1201 08:43:29.207703 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="a81d6f9c-e9f7-4c98-8233-c9f214978bfa" containerName="extract-content" Dec 01 08:43:29 crc kubenswrapper[4813]: E1201 08:43:29.207717 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3c0b52aa-86ba-4c56-b23e-108cdb4750d9" containerName="extract-content" Dec 01 08:43:29 crc kubenswrapper[4813]: I1201 08:43:29.207725 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="3c0b52aa-86ba-4c56-b23e-108cdb4750d9" containerName="extract-content" Dec 01 08:43:29 crc kubenswrapper[4813]: E1201 08:43:29.207735 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="95d1024c-5b73-40ae-9460-29502da48a23" containerName="extract-utilities" Dec 01 08:43:29 crc kubenswrapper[4813]: I1201 08:43:29.207742 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="95d1024c-5b73-40ae-9460-29502da48a23" containerName="extract-utilities" Dec 01 08:43:29 crc kubenswrapper[4813]: E1201 08:43:29.207753 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bd02bd16-7f86-48ad-b554-6404f62b1b28" containerName="extract-utilities" Dec 01 08:43:29 crc kubenswrapper[4813]: I1201 08:43:29.207760 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="bd02bd16-7f86-48ad-b554-6404f62b1b28" containerName="extract-utilities" Dec 01 08:43:29 crc kubenswrapper[4813]: E1201 08:43:29.207771 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a81d6f9c-e9f7-4c98-8233-c9f214978bfa" containerName="registry-server" Dec 01 08:43:29 crc kubenswrapper[4813]: I1201 08:43:29.207778 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="a81d6f9c-e9f7-4c98-8233-c9f214978bfa" containerName="registry-server" Dec 01 08:43:29 crc kubenswrapper[4813]: E1201 08:43:29.207795 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bd02bd16-7f86-48ad-b554-6404f62b1b28" containerName="registry-server" Dec 01 08:43:29 crc kubenswrapper[4813]: I1201 08:43:29.207803 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="bd02bd16-7f86-48ad-b554-6404f62b1b28" containerName="registry-server" Dec 01 08:43:29 crc kubenswrapper[4813]: E1201 08:43:29.207813 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3c0b52aa-86ba-4c56-b23e-108cdb4750d9" containerName="registry-server" Dec 01 08:43:29 crc kubenswrapper[4813]: I1201 08:43:29.207821 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="3c0b52aa-86ba-4c56-b23e-108cdb4750d9" containerName="registry-server" Dec 01 08:43:29 crc kubenswrapper[4813]: E1201 08:43:29.207833 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="95d1024c-5b73-40ae-9460-29502da48a23" containerName="extract-content" Dec 01 08:43:29 crc kubenswrapper[4813]: I1201 08:43:29.207840 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="95d1024c-5b73-40ae-9460-29502da48a23" containerName="extract-content" Dec 01 08:43:29 crc kubenswrapper[4813]: E1201 08:43:29.207850 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a81d6f9c-e9f7-4c98-8233-c9f214978bfa" containerName="extract-utilities" Dec 01 08:43:29 crc kubenswrapper[4813]: I1201 08:43:29.207858 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="a81d6f9c-e9f7-4c98-8233-c9f214978bfa" containerName="extract-utilities" Dec 01 08:43:29 crc kubenswrapper[4813]: I1201 08:43:29.208031 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="3c0b52aa-86ba-4c56-b23e-108cdb4750d9" containerName="registry-server" Dec 01 08:43:29 crc kubenswrapper[4813]: I1201 08:43:29.208064 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="a81d6f9c-e9f7-4c98-8233-c9f214978bfa" containerName="registry-server" Dec 01 08:43:29 crc kubenswrapper[4813]: I1201 08:43:29.208083 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="95d1024c-5b73-40ae-9460-29502da48a23" containerName="registry-server" Dec 01 08:43:29 crc kubenswrapper[4813]: I1201 08:43:29.208095 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="52120ed5-ff4f-4df5-8c60-3342df484be8" containerName="oauth-openshift" Dec 01 08:43:29 crc kubenswrapper[4813]: I1201 08:43:29.208110 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="bd02bd16-7f86-48ad-b554-6404f62b1b28" containerName="registry-server" Dec 01 08:43:29 crc kubenswrapper[4813]: I1201 08:43:29.208623 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-65c4c84884-2vrp6" Dec 01 08:43:29 crc kubenswrapper[4813]: I1201 08:43:29.211328 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Dec 01 08:43:29 crc kubenswrapper[4813]: I1201 08:43:29.211726 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Dec 01 08:43:29 crc kubenswrapper[4813]: I1201 08:43:29.212662 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Dec 01 08:43:29 crc kubenswrapper[4813]: I1201 08:43:29.212742 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Dec 01 08:43:29 crc kubenswrapper[4813]: I1201 08:43:29.213446 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Dec 01 08:43:29 crc kubenswrapper[4813]: I1201 08:43:29.216490 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Dec 01 08:43:29 crc kubenswrapper[4813]: I1201 08:43:29.216894 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-65c4c84884-2vrp6"] Dec 01 08:43:29 crc kubenswrapper[4813]: I1201 08:43:29.216915 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Dec 01 08:43:29 crc kubenswrapper[4813]: I1201 08:43:29.216984 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Dec 01 08:43:29 crc kubenswrapper[4813]: I1201 08:43:29.217094 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Dec 01 08:43:29 crc kubenswrapper[4813]: I1201 08:43:29.217164 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Dec 01 08:43:29 crc kubenswrapper[4813]: I1201 08:43:29.217543 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Dec 01 08:43:29 crc kubenswrapper[4813]: I1201 08:43:29.217613 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Dec 01 08:43:29 crc kubenswrapper[4813]: I1201 08:43:29.217801 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Dec 01 08:43:29 crc kubenswrapper[4813]: I1201 08:43:29.224446 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Dec 01 08:43:29 crc kubenswrapper[4813]: I1201 08:43:29.228328 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Dec 01 08:43:29 crc kubenswrapper[4813]: I1201 08:43:29.233650 4813 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Dec 01 08:43:29 crc kubenswrapper[4813]: I1201 08:43:29.238091 4813 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 01 08:43:29 crc kubenswrapper[4813]: I1201 08:43:29.239730 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" containerID="cri-o://d441a348087ae3cadda1ad8a7d1c4dff02cf3f6e1d346c294b278fdd9131edcc" gracePeriod=15 Dec 01 08:43:29 crc kubenswrapper[4813]: I1201 08:43:29.240186 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 01 08:43:29 crc kubenswrapper[4813]: I1201 08:43:29.241371 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" containerID="cri-o://8dc6feb9817f4f356b8e36af43ec3070774f312fbbb4909fa05b8b3e8dd53582" gracePeriod=15 Dec 01 08:43:29 crc kubenswrapper[4813]: I1201 08:43:29.241530 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" containerID="cri-o://7354f9337d42115bb1d974bf3772f2d03365acba57f2155135c7a0bbab39879e" gracePeriod=15 Dec 01 08:43:29 crc kubenswrapper[4813]: I1201 08:43:29.241743 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" containerID="cri-o://da67635daf6d9e32937f4d5d4ceeee3f8f09ae4fd2dd26c3e85be1286c5949de" gracePeriod=15 Dec 01 08:43:29 crc kubenswrapper[4813]: I1201 08:43:29.241811 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" containerID="cri-o://6e97356ee274ea78d9f237ffbc756abb6ac8df0dbac244bf0a86ed215b8cbf43" gracePeriod=15 Dec 01 08:43:29 crc kubenswrapper[4813]: I1201 08:43:29.246955 4813 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 01 08:43:29 crc kubenswrapper[4813]: E1201 08:43:29.247721 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Dec 01 08:43:29 crc kubenswrapper[4813]: I1201 08:43:29.247742 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Dec 01 08:43:29 crc kubenswrapper[4813]: E1201 08:43:29.247757 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 01 08:43:29 crc kubenswrapper[4813]: I1201 08:43:29.247763 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 01 08:43:29 crc kubenswrapper[4813]: E1201 08:43:29.247774 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Dec 01 08:43:29 crc kubenswrapper[4813]: I1201 08:43:29.247781 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Dec 01 08:43:29 crc kubenswrapper[4813]: E1201 08:43:29.247793 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Dec 01 08:43:29 crc kubenswrapper[4813]: I1201 08:43:29.247799 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Dec 01 08:43:29 crc kubenswrapper[4813]: E1201 08:43:29.247807 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 01 08:43:29 crc kubenswrapper[4813]: I1201 08:43:29.247813 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 01 08:43:29 crc kubenswrapper[4813]: E1201 08:43:29.247824 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Dec 01 08:43:29 crc kubenswrapper[4813]: I1201 08:43:29.247830 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Dec 01 08:43:29 crc kubenswrapper[4813]: E1201 08:43:29.247845 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Dec 01 08:43:29 crc kubenswrapper[4813]: I1201 08:43:29.247851 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Dec 01 08:43:29 crc kubenswrapper[4813]: I1201 08:43:29.248056 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Dec 01 08:43:29 crc kubenswrapper[4813]: I1201 08:43:29.248074 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 01 08:43:29 crc kubenswrapper[4813]: I1201 08:43:29.248085 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Dec 01 08:43:29 crc kubenswrapper[4813]: I1201 08:43:29.248092 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Dec 01 08:43:29 crc kubenswrapper[4813]: I1201 08:43:29.248102 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 01 08:43:29 crc kubenswrapper[4813]: I1201 08:43:29.248109 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 01 08:43:29 crc kubenswrapper[4813]: I1201 08:43:29.248123 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Dec 01 08:43:29 crc kubenswrapper[4813]: E1201 08:43:29.248323 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 01 08:43:29 crc kubenswrapper[4813]: I1201 08:43:29.248334 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 01 08:43:29 crc kubenswrapper[4813]: I1201 08:43:29.373581 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/9781d8cf-9f56-4d33-b05d-0877b490f389-audit-dir\") pod \"oauth-openshift-65c4c84884-2vrp6\" (UID: \"9781d8cf-9f56-4d33-b05d-0877b490f389\") " pod="openshift-authentication/oauth-openshift-65c4c84884-2vrp6" Dec 01 08:43:29 crc kubenswrapper[4813]: I1201 08:43:29.373813 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 01 08:43:29 crc kubenswrapper[4813]: I1201 08:43:29.373838 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/9781d8cf-9f56-4d33-b05d-0877b490f389-v4-0-config-user-template-error\") pod \"oauth-openshift-65c4c84884-2vrp6\" (UID: \"9781d8cf-9f56-4d33-b05d-0877b490f389\") " pod="openshift-authentication/oauth-openshift-65c4c84884-2vrp6" Dec 01 08:43:29 crc kubenswrapper[4813]: I1201 08:43:29.373862 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 01 08:43:29 crc kubenswrapper[4813]: I1201 08:43:29.373880 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/9781d8cf-9f56-4d33-b05d-0877b490f389-audit-policies\") pod \"oauth-openshift-65c4c84884-2vrp6\" (UID: \"9781d8cf-9f56-4d33-b05d-0877b490f389\") " pod="openshift-authentication/oauth-openshift-65c4c84884-2vrp6" Dec 01 08:43:29 crc kubenswrapper[4813]: I1201 08:43:29.373898 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x6qcw\" (UniqueName: \"kubernetes.io/projected/9781d8cf-9f56-4d33-b05d-0877b490f389-kube-api-access-x6qcw\") pod \"oauth-openshift-65c4c84884-2vrp6\" (UID: \"9781d8cf-9f56-4d33-b05d-0877b490f389\") " pod="openshift-authentication/oauth-openshift-65c4c84884-2vrp6" Dec 01 08:43:29 crc kubenswrapper[4813]: I1201 08:43:29.373916 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 01 08:43:29 crc kubenswrapper[4813]: I1201 08:43:29.373955 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/9781d8cf-9f56-4d33-b05d-0877b490f389-v4-0-config-system-serving-cert\") pod \"oauth-openshift-65c4c84884-2vrp6\" (UID: \"9781d8cf-9f56-4d33-b05d-0877b490f389\") " pod="openshift-authentication/oauth-openshift-65c4c84884-2vrp6" Dec 01 08:43:29 crc kubenswrapper[4813]: I1201 08:43:29.373996 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 08:43:29 crc kubenswrapper[4813]: I1201 08:43:29.374009 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 01 08:43:29 crc kubenswrapper[4813]: I1201 08:43:29.374026 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 01 08:43:29 crc kubenswrapper[4813]: I1201 08:43:29.374050 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/9781d8cf-9f56-4d33-b05d-0877b490f389-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-65c4c84884-2vrp6\" (UID: \"9781d8cf-9f56-4d33-b05d-0877b490f389\") " pod="openshift-authentication/oauth-openshift-65c4c84884-2vrp6" Dec 01 08:43:29 crc kubenswrapper[4813]: I1201 08:43:29.374069 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/9781d8cf-9f56-4d33-b05d-0877b490f389-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-65c4c84884-2vrp6\" (UID: \"9781d8cf-9f56-4d33-b05d-0877b490f389\") " pod="openshift-authentication/oauth-openshift-65c4c84884-2vrp6" Dec 01 08:43:29 crc kubenswrapper[4813]: I1201 08:43:29.374091 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/9781d8cf-9f56-4d33-b05d-0877b490f389-v4-0-config-user-template-login\") pod \"oauth-openshift-65c4c84884-2vrp6\" (UID: \"9781d8cf-9f56-4d33-b05d-0877b490f389\") " pod="openshift-authentication/oauth-openshift-65c4c84884-2vrp6" Dec 01 08:43:29 crc kubenswrapper[4813]: I1201 08:43:29.374107 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9781d8cf-9f56-4d33-b05d-0877b490f389-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-65c4c84884-2vrp6\" (UID: \"9781d8cf-9f56-4d33-b05d-0877b490f389\") " pod="openshift-authentication/oauth-openshift-65c4c84884-2vrp6" Dec 01 08:43:29 crc kubenswrapper[4813]: I1201 08:43:29.374192 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/9781d8cf-9f56-4d33-b05d-0877b490f389-v4-0-config-system-session\") pod \"oauth-openshift-65c4c84884-2vrp6\" (UID: \"9781d8cf-9f56-4d33-b05d-0877b490f389\") " pod="openshift-authentication/oauth-openshift-65c4c84884-2vrp6" Dec 01 08:43:29 crc kubenswrapper[4813]: I1201 08:43:29.374274 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/9781d8cf-9f56-4d33-b05d-0877b490f389-v4-0-config-system-router-certs\") pod \"oauth-openshift-65c4c84884-2vrp6\" (UID: \"9781d8cf-9f56-4d33-b05d-0877b490f389\") " pod="openshift-authentication/oauth-openshift-65c4c84884-2vrp6" Dec 01 08:43:29 crc kubenswrapper[4813]: I1201 08:43:29.374316 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/9781d8cf-9f56-4d33-b05d-0877b490f389-v4-0-config-system-cliconfig\") pod \"oauth-openshift-65c4c84884-2vrp6\" (UID: \"9781d8cf-9f56-4d33-b05d-0877b490f389\") " pod="openshift-authentication/oauth-openshift-65c4c84884-2vrp6" Dec 01 08:43:29 crc kubenswrapper[4813]: I1201 08:43:29.374346 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 08:43:29 crc kubenswrapper[4813]: I1201 08:43:29.374383 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/9781d8cf-9f56-4d33-b05d-0877b490f389-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-65c4c84884-2vrp6\" (UID: \"9781d8cf-9f56-4d33-b05d-0877b490f389\") " pod="openshift-authentication/oauth-openshift-65c4c84884-2vrp6" Dec 01 08:43:29 crc kubenswrapper[4813]: I1201 08:43:29.374405 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/9781d8cf-9f56-4d33-b05d-0877b490f389-v4-0-config-system-service-ca\") pod \"oauth-openshift-65c4c84884-2vrp6\" (UID: \"9781d8cf-9f56-4d33-b05d-0877b490f389\") " pod="openshift-authentication/oauth-openshift-65c4c84884-2vrp6" Dec 01 08:43:29 crc kubenswrapper[4813]: I1201 08:43:29.374423 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 08:43:29 crc kubenswrapper[4813]: I1201 08:43:29.474662 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/9781d8cf-9f56-4d33-b05d-0877b490f389-v4-0-config-system-serving-cert\") pod \"oauth-openshift-65c4c84884-2vrp6\" (UID: \"9781d8cf-9f56-4d33-b05d-0877b490f389\") " pod="openshift-authentication/oauth-openshift-65c4c84884-2vrp6" Dec 01 08:43:29 crc kubenswrapper[4813]: I1201 08:43:29.474726 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 08:43:29 crc kubenswrapper[4813]: I1201 08:43:29.474748 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 01 08:43:29 crc kubenswrapper[4813]: I1201 08:43:29.474781 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 01 08:43:29 crc kubenswrapper[4813]: I1201 08:43:29.474821 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/9781d8cf-9f56-4d33-b05d-0877b490f389-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-65c4c84884-2vrp6\" (UID: \"9781d8cf-9f56-4d33-b05d-0877b490f389\") " pod="openshift-authentication/oauth-openshift-65c4c84884-2vrp6" Dec 01 08:43:29 crc kubenswrapper[4813]: I1201 08:43:29.474846 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/9781d8cf-9f56-4d33-b05d-0877b490f389-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-65c4c84884-2vrp6\" (UID: \"9781d8cf-9f56-4d33-b05d-0877b490f389\") " pod="openshift-authentication/oauth-openshift-65c4c84884-2vrp6" Dec 01 08:43:29 crc kubenswrapper[4813]: I1201 08:43:29.474856 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 08:43:29 crc kubenswrapper[4813]: I1201 08:43:29.474867 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/9781d8cf-9f56-4d33-b05d-0877b490f389-v4-0-config-user-template-login\") pod \"oauth-openshift-65c4c84884-2vrp6\" (UID: \"9781d8cf-9f56-4d33-b05d-0877b490f389\") " pod="openshift-authentication/oauth-openshift-65c4c84884-2vrp6" Dec 01 08:43:29 crc kubenswrapper[4813]: I1201 08:43:29.474948 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9781d8cf-9f56-4d33-b05d-0877b490f389-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-65c4c84884-2vrp6\" (UID: \"9781d8cf-9f56-4d33-b05d-0877b490f389\") " pod="openshift-authentication/oauth-openshift-65c4c84884-2vrp6" Dec 01 08:43:29 crc kubenswrapper[4813]: I1201 08:43:29.474995 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/9781d8cf-9f56-4d33-b05d-0877b490f389-v4-0-config-system-session\") pod \"oauth-openshift-65c4c84884-2vrp6\" (UID: \"9781d8cf-9f56-4d33-b05d-0877b490f389\") " pod="openshift-authentication/oauth-openshift-65c4c84884-2vrp6" Dec 01 08:43:29 crc kubenswrapper[4813]: I1201 08:43:29.475019 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/9781d8cf-9f56-4d33-b05d-0877b490f389-v4-0-config-system-router-certs\") pod \"oauth-openshift-65c4c84884-2vrp6\" (UID: \"9781d8cf-9f56-4d33-b05d-0877b490f389\") " pod="openshift-authentication/oauth-openshift-65c4c84884-2vrp6" Dec 01 08:43:29 crc kubenswrapper[4813]: I1201 08:43:29.475035 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/9781d8cf-9f56-4d33-b05d-0877b490f389-v4-0-config-system-cliconfig\") pod \"oauth-openshift-65c4c84884-2vrp6\" (UID: \"9781d8cf-9f56-4d33-b05d-0877b490f389\") " pod="openshift-authentication/oauth-openshift-65c4c84884-2vrp6" Dec 01 08:43:29 crc kubenswrapper[4813]: I1201 08:43:29.475065 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 08:43:29 crc kubenswrapper[4813]: I1201 08:43:29.475123 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/9781d8cf-9f56-4d33-b05d-0877b490f389-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-65c4c84884-2vrp6\" (UID: \"9781d8cf-9f56-4d33-b05d-0877b490f389\") " pod="openshift-authentication/oauth-openshift-65c4c84884-2vrp6" Dec 01 08:43:29 crc kubenswrapper[4813]: I1201 08:43:29.475159 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/9781d8cf-9f56-4d33-b05d-0877b490f389-v4-0-config-system-service-ca\") pod \"oauth-openshift-65c4c84884-2vrp6\" (UID: \"9781d8cf-9f56-4d33-b05d-0877b490f389\") " pod="openshift-authentication/oauth-openshift-65c4c84884-2vrp6" Dec 01 08:43:29 crc kubenswrapper[4813]: I1201 08:43:29.475189 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 08:43:29 crc kubenswrapper[4813]: I1201 08:43:29.475227 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/9781d8cf-9f56-4d33-b05d-0877b490f389-audit-dir\") pod \"oauth-openshift-65c4c84884-2vrp6\" (UID: \"9781d8cf-9f56-4d33-b05d-0877b490f389\") " pod="openshift-authentication/oauth-openshift-65c4c84884-2vrp6" Dec 01 08:43:29 crc kubenswrapper[4813]: I1201 08:43:29.475248 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 01 08:43:29 crc kubenswrapper[4813]: I1201 08:43:29.475270 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/9781d8cf-9f56-4d33-b05d-0877b490f389-v4-0-config-user-template-error\") pod \"oauth-openshift-65c4c84884-2vrp6\" (UID: \"9781d8cf-9f56-4d33-b05d-0877b490f389\") " pod="openshift-authentication/oauth-openshift-65c4c84884-2vrp6" Dec 01 08:43:29 crc kubenswrapper[4813]: I1201 08:43:29.475336 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 01 08:43:29 crc kubenswrapper[4813]: I1201 08:43:29.475372 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/9781d8cf-9f56-4d33-b05d-0877b490f389-audit-policies\") pod \"oauth-openshift-65c4c84884-2vrp6\" (UID: \"9781d8cf-9f56-4d33-b05d-0877b490f389\") " pod="openshift-authentication/oauth-openshift-65c4c84884-2vrp6" Dec 01 08:43:29 crc kubenswrapper[4813]: I1201 08:43:29.475393 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x6qcw\" (UniqueName: \"kubernetes.io/projected/9781d8cf-9f56-4d33-b05d-0877b490f389-kube-api-access-x6qcw\") pod \"oauth-openshift-65c4c84884-2vrp6\" (UID: \"9781d8cf-9f56-4d33-b05d-0877b490f389\") " pod="openshift-authentication/oauth-openshift-65c4c84884-2vrp6" Dec 01 08:43:29 crc kubenswrapper[4813]: I1201 08:43:29.475422 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 01 08:43:29 crc kubenswrapper[4813]: I1201 08:43:29.475503 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 01 08:43:29 crc kubenswrapper[4813]: I1201 08:43:29.475530 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 01 08:43:29 crc kubenswrapper[4813]: I1201 08:43:29.477112 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9781d8cf-9f56-4d33-b05d-0877b490f389-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-65c4c84884-2vrp6\" (UID: \"9781d8cf-9f56-4d33-b05d-0877b490f389\") " pod="openshift-authentication/oauth-openshift-65c4c84884-2vrp6" Dec 01 08:43:29 crc kubenswrapper[4813]: I1201 08:43:29.477216 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 01 08:43:29 crc kubenswrapper[4813]: I1201 08:43:29.477267 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/9781d8cf-9f56-4d33-b05d-0877b490f389-audit-dir\") pod \"oauth-openshift-65c4c84884-2vrp6\" (UID: \"9781d8cf-9f56-4d33-b05d-0877b490f389\") " pod="openshift-authentication/oauth-openshift-65c4c84884-2vrp6" Dec 01 08:43:29 crc kubenswrapper[4813]: I1201 08:43:29.477851 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 01 08:43:29 crc kubenswrapper[4813]: I1201 08:43:29.478012 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 01 08:43:29 crc kubenswrapper[4813]: I1201 08:43:29.478509 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/9781d8cf-9f56-4d33-b05d-0877b490f389-v4-0-config-system-cliconfig\") pod \"oauth-openshift-65c4c84884-2vrp6\" (UID: \"9781d8cf-9f56-4d33-b05d-0877b490f389\") " pod="openshift-authentication/oauth-openshift-65c4c84884-2vrp6" Dec 01 08:43:29 crc kubenswrapper[4813]: I1201 08:43:29.478537 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/9781d8cf-9f56-4d33-b05d-0877b490f389-audit-policies\") pod \"oauth-openshift-65c4c84884-2vrp6\" (UID: \"9781d8cf-9f56-4d33-b05d-0877b490f389\") " pod="openshift-authentication/oauth-openshift-65c4c84884-2vrp6" Dec 01 08:43:29 crc kubenswrapper[4813]: I1201 08:43:29.478874 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 08:43:29 crc kubenswrapper[4813]: I1201 08:43:29.479279 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/9781d8cf-9f56-4d33-b05d-0877b490f389-v4-0-config-system-service-ca\") pod \"oauth-openshift-65c4c84884-2vrp6\" (UID: \"9781d8cf-9f56-4d33-b05d-0877b490f389\") " pod="openshift-authentication/oauth-openshift-65c4c84884-2vrp6" Dec 01 08:43:29 crc kubenswrapper[4813]: I1201 08:43:29.479325 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 08:43:29 crc kubenswrapper[4813]: E1201 08:43:29.479342 4813 projected.go:194] Error preparing data for projected volume kube-api-access-x6qcw for pod openshift-authentication/oauth-openshift-65c4c84884-2vrp6: failed to fetch token: Post "https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/serviceaccounts/oauth-openshift/token": dial tcp 38.102.83.150:6443: connect: connection refused Dec 01 08:43:29 crc kubenswrapper[4813]: E1201 08:43:29.479438 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9781d8cf-9f56-4d33-b05d-0877b490f389-kube-api-access-x6qcw podName:9781d8cf-9f56-4d33-b05d-0877b490f389 nodeName:}" failed. No retries permitted until 2025-12-01 08:43:29.979407208 +0000 UTC m=+194.602328794 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-x6qcw" (UniqueName: "kubernetes.io/projected/9781d8cf-9f56-4d33-b05d-0877b490f389-kube-api-access-x6qcw") pod "oauth-openshift-65c4c84884-2vrp6" (UID: "9781d8cf-9f56-4d33-b05d-0877b490f389") : failed to fetch token: Post "https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/serviceaccounts/oauth-openshift/token": dial tcp 38.102.83.150:6443: connect: connection refused Dec 01 08:43:29 crc kubenswrapper[4813]: E1201 08:43:29.486108 4813 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/events\": dial tcp 38.102.83.150:6443: connect: connection refused" event="&Event{ObjectMeta:{oauth-openshift-65c4c84884-2vrp6.187d0ae56c45a0ee openshift-authentication 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-authentication,Name:oauth-openshift-65c4c84884-2vrp6,UID:9781d8cf-9f56-4d33-b05d-0877b490f389,APIVersion:v1,ResourceVersion:29305,FieldPath:,},Reason:FailedMount,Message:MountVolume.SetUp failed for volume \"kube-api-access-x6qcw\" : failed to fetch token: Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/serviceaccounts/oauth-openshift/token\": dial tcp 38.102.83.150:6443: connect: connection refused,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-01 08:43:29.479393518 +0000 UTC m=+194.102315104,LastTimestamp:2025-12-01 08:43:29.479393518 +0000 UTC m=+194.102315104,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 01 08:43:29 crc kubenswrapper[4813]: I1201 08:43:29.486877 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/9781d8cf-9f56-4d33-b05d-0877b490f389-v4-0-config-system-serving-cert\") pod \"oauth-openshift-65c4c84884-2vrp6\" (UID: \"9781d8cf-9f56-4d33-b05d-0877b490f389\") " pod="openshift-authentication/oauth-openshift-65c4c84884-2vrp6" Dec 01 08:43:29 crc kubenswrapper[4813]: I1201 08:43:29.488302 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/9781d8cf-9f56-4d33-b05d-0877b490f389-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-65c4c84884-2vrp6\" (UID: \"9781d8cf-9f56-4d33-b05d-0877b490f389\") " pod="openshift-authentication/oauth-openshift-65c4c84884-2vrp6" Dec 01 08:43:29 crc kubenswrapper[4813]: I1201 08:43:29.488419 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/9781d8cf-9f56-4d33-b05d-0877b490f389-v4-0-config-system-session\") pod \"oauth-openshift-65c4c84884-2vrp6\" (UID: \"9781d8cf-9f56-4d33-b05d-0877b490f389\") " pod="openshift-authentication/oauth-openshift-65c4c84884-2vrp6" Dec 01 08:43:29 crc kubenswrapper[4813]: I1201 08:43:29.488595 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/9781d8cf-9f56-4d33-b05d-0877b490f389-v4-0-config-user-template-login\") pod \"oauth-openshift-65c4c84884-2vrp6\" (UID: \"9781d8cf-9f56-4d33-b05d-0877b490f389\") " pod="openshift-authentication/oauth-openshift-65c4c84884-2vrp6" Dec 01 08:43:29 crc kubenswrapper[4813]: I1201 08:43:29.488793 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/9781d8cf-9f56-4d33-b05d-0877b490f389-v4-0-config-system-router-certs\") pod \"oauth-openshift-65c4c84884-2vrp6\" (UID: \"9781d8cf-9f56-4d33-b05d-0877b490f389\") " pod="openshift-authentication/oauth-openshift-65c4c84884-2vrp6" Dec 01 08:43:29 crc kubenswrapper[4813]: I1201 08:43:29.489326 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/9781d8cf-9f56-4d33-b05d-0877b490f389-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-65c4c84884-2vrp6\" (UID: \"9781d8cf-9f56-4d33-b05d-0877b490f389\") " pod="openshift-authentication/oauth-openshift-65c4c84884-2vrp6" Dec 01 08:43:29 crc kubenswrapper[4813]: I1201 08:43:29.489470 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/9781d8cf-9f56-4d33-b05d-0877b490f389-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-65c4c84884-2vrp6\" (UID: \"9781d8cf-9f56-4d33-b05d-0877b490f389\") " pod="openshift-authentication/oauth-openshift-65c4c84884-2vrp6" Dec 01 08:43:29 crc kubenswrapper[4813]: I1201 08:43:29.489653 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/9781d8cf-9f56-4d33-b05d-0877b490f389-v4-0-config-user-template-error\") pod \"oauth-openshift-65c4c84884-2vrp6\" (UID: \"9781d8cf-9f56-4d33-b05d-0877b490f389\") " pod="openshift-authentication/oauth-openshift-65c4c84884-2vrp6" Dec 01 08:43:29 crc kubenswrapper[4813]: I1201 08:43:29.644836 4813 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:6443/readyz\": dial tcp 192.168.126.11:6443: connect: connection refused" start-of-body= Dec 01 08:43:29 crc kubenswrapper[4813]: I1201 08:43:29.644905 4813 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="Get \"https://192.168.126.11:6443/readyz\": dial tcp 192.168.126.11:6443: connect: connection refused" Dec 01 08:43:29 crc kubenswrapper[4813]: I1201 08:43:29.983811 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x6qcw\" (UniqueName: \"kubernetes.io/projected/9781d8cf-9f56-4d33-b05d-0877b490f389-kube-api-access-x6qcw\") pod \"oauth-openshift-65c4c84884-2vrp6\" (UID: \"9781d8cf-9f56-4d33-b05d-0877b490f389\") " pod="openshift-authentication/oauth-openshift-65c4c84884-2vrp6" Dec 01 08:43:29 crc kubenswrapper[4813]: E1201 08:43:29.984895 4813 projected.go:194] Error preparing data for projected volume kube-api-access-x6qcw for pod openshift-authentication/oauth-openshift-65c4c84884-2vrp6: failed to fetch token: Post "https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/serviceaccounts/oauth-openshift/token": dial tcp 38.102.83.150:6443: connect: connection refused Dec 01 08:43:29 crc kubenswrapper[4813]: E1201 08:43:29.985056 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9781d8cf-9f56-4d33-b05d-0877b490f389-kube-api-access-x6qcw podName:9781d8cf-9f56-4d33-b05d-0877b490f389 nodeName:}" failed. No retries permitted until 2025-12-01 08:43:30.985017226 +0000 UTC m=+195.607938842 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-x6qcw" (UniqueName: "kubernetes.io/projected/9781d8cf-9f56-4d33-b05d-0877b490f389-kube-api-access-x6qcw") pod "oauth-openshift-65c4c84884-2vrp6" (UID: "9781d8cf-9f56-4d33-b05d-0877b490f389") : failed to fetch token: Post "https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/serviceaccounts/oauth-openshift/token": dial tcp 38.102.83.150:6443: connect: connection refused Dec 01 08:43:30 crc kubenswrapper[4813]: I1201 08:43:30.111774 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Dec 01 08:43:30 crc kubenswrapper[4813]: I1201 08:43:30.114008 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 01 08:43:30 crc kubenswrapper[4813]: I1201 08:43:30.115104 4813 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="8dc6feb9817f4f356b8e36af43ec3070774f312fbbb4909fa05b8b3e8dd53582" exitCode=0 Dec 01 08:43:30 crc kubenswrapper[4813]: I1201 08:43:30.115139 4813 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="7354f9337d42115bb1d974bf3772f2d03365acba57f2155135c7a0bbab39879e" exitCode=0 Dec 01 08:43:30 crc kubenswrapper[4813]: I1201 08:43:30.115147 4813 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="da67635daf6d9e32937f4d5d4ceeee3f8f09ae4fd2dd26c3e85be1286c5949de" exitCode=0 Dec 01 08:43:30 crc kubenswrapper[4813]: I1201 08:43:30.115154 4813 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="6e97356ee274ea78d9f237ffbc756abb6ac8df0dbac244bf0a86ed215b8cbf43" exitCode=2 Dec 01 08:43:30 crc kubenswrapper[4813]: I1201 08:43:30.115658 4813 scope.go:117] "RemoveContainer" containerID="a0404f5a5ed5a9750d4c479b608e68e8960b448f455d5af9d219aa5edaeb789b" Dec 01 08:43:30 crc kubenswrapper[4813]: I1201 08:43:30.117444 4813 generic.go:334] "Generic (PLEG): container finished" podID="269d86e5-dc67-48b3-a955-201b986f6c97" containerID="1ed1b375825fa13897f16c218101a45d9e8e16e971dbd5c8cde715800e1ee232" exitCode=0 Dec 01 08:43:30 crc kubenswrapper[4813]: I1201 08:43:30.117511 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"269d86e5-dc67-48b3-a955-201b986f6c97","Type":"ContainerDied","Data":"1ed1b375825fa13897f16c218101a45d9e8e16e971dbd5c8cde715800e1ee232"} Dec 01 08:43:30 crc kubenswrapper[4813]: I1201 08:43:30.118725 4813 status_manager.go:851] "Failed to get status for pod" podUID="269d86e5-dc67-48b3-a955-201b986f6c97" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.150:6443: connect: connection refused" Dec 01 08:43:30 crc kubenswrapper[4813]: I1201 08:43:30.119297 4813 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.150:6443: connect: connection refused" Dec 01 08:43:30 crc kubenswrapper[4813]: I1201 08:43:30.998836 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x6qcw\" (UniqueName: \"kubernetes.io/projected/9781d8cf-9f56-4d33-b05d-0877b490f389-kube-api-access-x6qcw\") pod \"oauth-openshift-65c4c84884-2vrp6\" (UID: \"9781d8cf-9f56-4d33-b05d-0877b490f389\") " pod="openshift-authentication/oauth-openshift-65c4c84884-2vrp6" Dec 01 08:43:31 crc kubenswrapper[4813]: E1201 08:43:30.999928 4813 projected.go:194] Error preparing data for projected volume kube-api-access-x6qcw for pod openshift-authentication/oauth-openshift-65c4c84884-2vrp6: failed to fetch token: Post "https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/serviceaccounts/oauth-openshift/token": dial tcp 38.102.83.150:6443: connect: connection refused Dec 01 08:43:31 crc kubenswrapper[4813]: E1201 08:43:31.000028 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9781d8cf-9f56-4d33-b05d-0877b490f389-kube-api-access-x6qcw podName:9781d8cf-9f56-4d33-b05d-0877b490f389 nodeName:}" failed. No retries permitted until 2025-12-01 08:43:33.000006797 +0000 UTC m=+197.622928423 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-x6qcw" (UniqueName: "kubernetes.io/projected/9781d8cf-9f56-4d33-b05d-0877b490f389-kube-api-access-x6qcw") pod "oauth-openshift-65c4c84884-2vrp6" (UID: "9781d8cf-9f56-4d33-b05d-0877b490f389") : failed to fetch token: Post "https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/serviceaccounts/oauth-openshift/token": dial tcp 38.102.83.150:6443: connect: connection refused Dec 01 08:43:31 crc kubenswrapper[4813]: I1201 08:43:31.125202 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 01 08:43:31 crc kubenswrapper[4813]: I1201 08:43:31.398927 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 01 08:43:31 crc kubenswrapper[4813]: I1201 08:43:31.399888 4813 status_manager.go:851] "Failed to get status for pod" podUID="269d86e5-dc67-48b3-a955-201b986f6c97" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.150:6443: connect: connection refused" Dec 01 08:43:31 crc kubenswrapper[4813]: I1201 08:43:31.507801 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/269d86e5-dc67-48b3-a955-201b986f6c97-var-lock\") pod \"269d86e5-dc67-48b3-a955-201b986f6c97\" (UID: \"269d86e5-dc67-48b3-a955-201b986f6c97\") " Dec 01 08:43:31 crc kubenswrapper[4813]: I1201 08:43:31.507866 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/269d86e5-dc67-48b3-a955-201b986f6c97-kube-api-access\") pod \"269d86e5-dc67-48b3-a955-201b986f6c97\" (UID: \"269d86e5-dc67-48b3-a955-201b986f6c97\") " Dec 01 08:43:31 crc kubenswrapper[4813]: I1201 08:43:31.507924 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/269d86e5-dc67-48b3-a955-201b986f6c97-kubelet-dir\") pod \"269d86e5-dc67-48b3-a955-201b986f6c97\" (UID: \"269d86e5-dc67-48b3-a955-201b986f6c97\") " Dec 01 08:43:31 crc kubenswrapper[4813]: I1201 08:43:31.508326 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/269d86e5-dc67-48b3-a955-201b986f6c97-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "269d86e5-dc67-48b3-a955-201b986f6c97" (UID: "269d86e5-dc67-48b3-a955-201b986f6c97"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 08:43:31 crc kubenswrapper[4813]: I1201 08:43:31.508403 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/269d86e5-dc67-48b3-a955-201b986f6c97-var-lock" (OuterVolumeSpecName: "var-lock") pod "269d86e5-dc67-48b3-a955-201b986f6c97" (UID: "269d86e5-dc67-48b3-a955-201b986f6c97"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 08:43:31 crc kubenswrapper[4813]: I1201 08:43:31.515209 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/269d86e5-dc67-48b3-a955-201b986f6c97-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "269d86e5-dc67-48b3-a955-201b986f6c97" (UID: "269d86e5-dc67-48b3-a955-201b986f6c97"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:43:31 crc kubenswrapper[4813]: I1201 08:43:31.610086 4813 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/269d86e5-dc67-48b3-a955-201b986f6c97-var-lock\") on node \"crc\" DevicePath \"\"" Dec 01 08:43:31 crc kubenswrapper[4813]: I1201 08:43:31.610273 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/269d86e5-dc67-48b3-a955-201b986f6c97-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 01 08:43:31 crc kubenswrapper[4813]: I1201 08:43:31.610291 4813 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/269d86e5-dc67-48b3-a955-201b986f6c97-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 01 08:43:32 crc kubenswrapper[4813]: I1201 08:43:32.108224 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 01 08:43:32 crc kubenswrapper[4813]: I1201 08:43:32.109261 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 08:43:32 crc kubenswrapper[4813]: I1201 08:43:32.110059 4813 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.150:6443: connect: connection refused" Dec 01 08:43:32 crc kubenswrapper[4813]: I1201 08:43:32.110615 4813 status_manager.go:851] "Failed to get status for pod" podUID="269d86e5-dc67-48b3-a955-201b986f6c97" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.150:6443: connect: connection refused" Dec 01 08:43:32 crc kubenswrapper[4813]: E1201 08:43:32.133676 4813 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.150:6443: connect: connection refused" Dec 01 08:43:32 crc kubenswrapper[4813]: E1201 08:43:32.134006 4813 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.150:6443: connect: connection refused" Dec 01 08:43:32 crc kubenswrapper[4813]: E1201 08:43:32.134520 4813 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.150:6443: connect: connection refused" Dec 01 08:43:32 crc kubenswrapper[4813]: E1201 08:43:32.134924 4813 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.150:6443: connect: connection refused" Dec 01 08:43:32 crc kubenswrapper[4813]: E1201 08:43:32.135426 4813 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.150:6443: connect: connection refused" Dec 01 08:43:32 crc kubenswrapper[4813]: I1201 08:43:32.135482 4813 controller.go:115] "failed to update lease using latest lease, fallback to ensure lease" err="failed 5 attempts to update lease" Dec 01 08:43:32 crc kubenswrapper[4813]: I1201 08:43:32.135684 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 01 08:43:32 crc kubenswrapper[4813]: E1201 08:43:32.135841 4813 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.150:6443: connect: connection refused" interval="200ms" Dec 01 08:43:32 crc kubenswrapper[4813]: I1201 08:43:32.137197 4813 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="d441a348087ae3cadda1ad8a7d1c4dff02cf3f6e1d346c294b278fdd9131edcc" exitCode=0 Dec 01 08:43:32 crc kubenswrapper[4813]: I1201 08:43:32.137459 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 08:43:32 crc kubenswrapper[4813]: I1201 08:43:32.137546 4813 scope.go:117] "RemoveContainer" containerID="8dc6feb9817f4f356b8e36af43ec3070774f312fbbb4909fa05b8b3e8dd53582" Dec 01 08:43:32 crc kubenswrapper[4813]: I1201 08:43:32.141648 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"269d86e5-dc67-48b3-a955-201b986f6c97","Type":"ContainerDied","Data":"7c8d443ea8564db109e481b03d907237fa30fd861875a9041e49e2dc433321f8"} Dec 01 08:43:32 crc kubenswrapper[4813]: I1201 08:43:32.141717 4813 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7c8d443ea8564db109e481b03d907237fa30fd861875a9041e49e2dc433321f8" Dec 01 08:43:32 crc kubenswrapper[4813]: I1201 08:43:32.141792 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 01 08:43:32 crc kubenswrapper[4813]: I1201 08:43:32.160940 4813 scope.go:117] "RemoveContainer" containerID="7354f9337d42115bb1d974bf3772f2d03365acba57f2155135c7a0bbab39879e" Dec 01 08:43:32 crc kubenswrapper[4813]: I1201 08:43:32.164458 4813 status_manager.go:851] "Failed to get status for pod" podUID="269d86e5-dc67-48b3-a955-201b986f6c97" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.150:6443: connect: connection refused" Dec 01 08:43:32 crc kubenswrapper[4813]: I1201 08:43:32.164932 4813 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.150:6443: connect: connection refused" Dec 01 08:43:32 crc kubenswrapper[4813]: I1201 08:43:32.177506 4813 scope.go:117] "RemoveContainer" containerID="da67635daf6d9e32937f4d5d4ceeee3f8f09ae4fd2dd26c3e85be1286c5949de" Dec 01 08:43:32 crc kubenswrapper[4813]: I1201 08:43:32.198355 4813 scope.go:117] "RemoveContainer" containerID="6e97356ee274ea78d9f237ffbc756abb6ac8df0dbac244bf0a86ed215b8cbf43" Dec 01 08:43:32 crc kubenswrapper[4813]: I1201 08:43:32.213401 4813 scope.go:117] "RemoveContainer" containerID="d441a348087ae3cadda1ad8a7d1c4dff02cf3f6e1d346c294b278fdd9131edcc" Dec 01 08:43:32 crc kubenswrapper[4813]: I1201 08:43:32.219413 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Dec 01 08:43:32 crc kubenswrapper[4813]: I1201 08:43:32.219498 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Dec 01 08:43:32 crc kubenswrapper[4813]: I1201 08:43:32.219557 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Dec 01 08:43:32 crc kubenswrapper[4813]: I1201 08:43:32.219933 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir" (OuterVolumeSpecName: "cert-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "cert-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 08:43:32 crc kubenswrapper[4813]: I1201 08:43:32.220010 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 08:43:32 crc kubenswrapper[4813]: I1201 08:43:32.220053 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 08:43:32 crc kubenswrapper[4813]: I1201 08:43:32.220610 4813 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") on node \"crc\" DevicePath \"\"" Dec 01 08:43:32 crc kubenswrapper[4813]: I1201 08:43:32.220644 4813 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") on node \"crc\" DevicePath \"\"" Dec 01 08:43:32 crc kubenswrapper[4813]: I1201 08:43:32.220662 4813 reconciler_common.go:293] "Volume detached for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") on node \"crc\" DevicePath \"\"" Dec 01 08:43:32 crc kubenswrapper[4813]: I1201 08:43:32.233771 4813 scope.go:117] "RemoveContainer" containerID="05492df5703392cdf0932d6bf66025b18e53bd05fa0857a7cb2ad70979cbdc3e" Dec 01 08:43:32 crc kubenswrapper[4813]: I1201 08:43:32.256231 4813 scope.go:117] "RemoveContainer" containerID="8dc6feb9817f4f356b8e36af43ec3070774f312fbbb4909fa05b8b3e8dd53582" Dec 01 08:43:32 crc kubenswrapper[4813]: E1201 08:43:32.256868 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8dc6feb9817f4f356b8e36af43ec3070774f312fbbb4909fa05b8b3e8dd53582\": container with ID starting with 8dc6feb9817f4f356b8e36af43ec3070774f312fbbb4909fa05b8b3e8dd53582 not found: ID does not exist" containerID="8dc6feb9817f4f356b8e36af43ec3070774f312fbbb4909fa05b8b3e8dd53582" Dec 01 08:43:32 crc kubenswrapper[4813]: I1201 08:43:32.256930 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8dc6feb9817f4f356b8e36af43ec3070774f312fbbb4909fa05b8b3e8dd53582"} err="failed to get container status \"8dc6feb9817f4f356b8e36af43ec3070774f312fbbb4909fa05b8b3e8dd53582\": rpc error: code = NotFound desc = could not find container \"8dc6feb9817f4f356b8e36af43ec3070774f312fbbb4909fa05b8b3e8dd53582\": container with ID starting with 8dc6feb9817f4f356b8e36af43ec3070774f312fbbb4909fa05b8b3e8dd53582 not found: ID does not exist" Dec 01 08:43:32 crc kubenswrapper[4813]: I1201 08:43:32.257003 4813 scope.go:117] "RemoveContainer" containerID="7354f9337d42115bb1d974bf3772f2d03365acba57f2155135c7a0bbab39879e" Dec 01 08:43:32 crc kubenswrapper[4813]: E1201 08:43:32.257475 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7354f9337d42115bb1d974bf3772f2d03365acba57f2155135c7a0bbab39879e\": container with ID starting with 7354f9337d42115bb1d974bf3772f2d03365acba57f2155135c7a0bbab39879e not found: ID does not exist" containerID="7354f9337d42115bb1d974bf3772f2d03365acba57f2155135c7a0bbab39879e" Dec 01 08:43:32 crc kubenswrapper[4813]: I1201 08:43:32.257530 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7354f9337d42115bb1d974bf3772f2d03365acba57f2155135c7a0bbab39879e"} err="failed to get container status \"7354f9337d42115bb1d974bf3772f2d03365acba57f2155135c7a0bbab39879e\": rpc error: code = NotFound desc = could not find container \"7354f9337d42115bb1d974bf3772f2d03365acba57f2155135c7a0bbab39879e\": container with ID starting with 7354f9337d42115bb1d974bf3772f2d03365acba57f2155135c7a0bbab39879e not found: ID does not exist" Dec 01 08:43:32 crc kubenswrapper[4813]: I1201 08:43:32.257567 4813 scope.go:117] "RemoveContainer" containerID="da67635daf6d9e32937f4d5d4ceeee3f8f09ae4fd2dd26c3e85be1286c5949de" Dec 01 08:43:32 crc kubenswrapper[4813]: E1201 08:43:32.258045 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"da67635daf6d9e32937f4d5d4ceeee3f8f09ae4fd2dd26c3e85be1286c5949de\": container with ID starting with da67635daf6d9e32937f4d5d4ceeee3f8f09ae4fd2dd26c3e85be1286c5949de not found: ID does not exist" containerID="da67635daf6d9e32937f4d5d4ceeee3f8f09ae4fd2dd26c3e85be1286c5949de" Dec 01 08:43:32 crc kubenswrapper[4813]: I1201 08:43:32.258089 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"da67635daf6d9e32937f4d5d4ceeee3f8f09ae4fd2dd26c3e85be1286c5949de"} err="failed to get container status \"da67635daf6d9e32937f4d5d4ceeee3f8f09ae4fd2dd26c3e85be1286c5949de\": rpc error: code = NotFound desc = could not find container \"da67635daf6d9e32937f4d5d4ceeee3f8f09ae4fd2dd26c3e85be1286c5949de\": container with ID starting with da67635daf6d9e32937f4d5d4ceeee3f8f09ae4fd2dd26c3e85be1286c5949de not found: ID does not exist" Dec 01 08:43:32 crc kubenswrapper[4813]: I1201 08:43:32.258122 4813 scope.go:117] "RemoveContainer" containerID="6e97356ee274ea78d9f237ffbc756abb6ac8df0dbac244bf0a86ed215b8cbf43" Dec 01 08:43:32 crc kubenswrapper[4813]: E1201 08:43:32.258554 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6e97356ee274ea78d9f237ffbc756abb6ac8df0dbac244bf0a86ed215b8cbf43\": container with ID starting with 6e97356ee274ea78d9f237ffbc756abb6ac8df0dbac244bf0a86ed215b8cbf43 not found: ID does not exist" containerID="6e97356ee274ea78d9f237ffbc756abb6ac8df0dbac244bf0a86ed215b8cbf43" Dec 01 08:43:32 crc kubenswrapper[4813]: I1201 08:43:32.258623 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6e97356ee274ea78d9f237ffbc756abb6ac8df0dbac244bf0a86ed215b8cbf43"} err="failed to get container status \"6e97356ee274ea78d9f237ffbc756abb6ac8df0dbac244bf0a86ed215b8cbf43\": rpc error: code = NotFound desc = could not find container \"6e97356ee274ea78d9f237ffbc756abb6ac8df0dbac244bf0a86ed215b8cbf43\": container with ID starting with 6e97356ee274ea78d9f237ffbc756abb6ac8df0dbac244bf0a86ed215b8cbf43 not found: ID does not exist" Dec 01 08:43:32 crc kubenswrapper[4813]: I1201 08:43:32.258680 4813 scope.go:117] "RemoveContainer" containerID="d441a348087ae3cadda1ad8a7d1c4dff02cf3f6e1d346c294b278fdd9131edcc" Dec 01 08:43:32 crc kubenswrapper[4813]: E1201 08:43:32.259072 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d441a348087ae3cadda1ad8a7d1c4dff02cf3f6e1d346c294b278fdd9131edcc\": container with ID starting with d441a348087ae3cadda1ad8a7d1c4dff02cf3f6e1d346c294b278fdd9131edcc not found: ID does not exist" containerID="d441a348087ae3cadda1ad8a7d1c4dff02cf3f6e1d346c294b278fdd9131edcc" Dec 01 08:43:32 crc kubenswrapper[4813]: I1201 08:43:32.259132 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d441a348087ae3cadda1ad8a7d1c4dff02cf3f6e1d346c294b278fdd9131edcc"} err="failed to get container status \"d441a348087ae3cadda1ad8a7d1c4dff02cf3f6e1d346c294b278fdd9131edcc\": rpc error: code = NotFound desc = could not find container \"d441a348087ae3cadda1ad8a7d1c4dff02cf3f6e1d346c294b278fdd9131edcc\": container with ID starting with d441a348087ae3cadda1ad8a7d1c4dff02cf3f6e1d346c294b278fdd9131edcc not found: ID does not exist" Dec 01 08:43:32 crc kubenswrapper[4813]: I1201 08:43:32.259190 4813 scope.go:117] "RemoveContainer" containerID="05492df5703392cdf0932d6bf66025b18e53bd05fa0857a7cb2ad70979cbdc3e" Dec 01 08:43:32 crc kubenswrapper[4813]: E1201 08:43:32.259597 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"05492df5703392cdf0932d6bf66025b18e53bd05fa0857a7cb2ad70979cbdc3e\": container with ID starting with 05492df5703392cdf0932d6bf66025b18e53bd05fa0857a7cb2ad70979cbdc3e not found: ID does not exist" containerID="05492df5703392cdf0932d6bf66025b18e53bd05fa0857a7cb2ad70979cbdc3e" Dec 01 08:43:32 crc kubenswrapper[4813]: I1201 08:43:32.259640 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"05492df5703392cdf0932d6bf66025b18e53bd05fa0857a7cb2ad70979cbdc3e"} err="failed to get container status \"05492df5703392cdf0932d6bf66025b18e53bd05fa0857a7cb2ad70979cbdc3e\": rpc error: code = NotFound desc = could not find container \"05492df5703392cdf0932d6bf66025b18e53bd05fa0857a7cb2ad70979cbdc3e\": container with ID starting with 05492df5703392cdf0932d6bf66025b18e53bd05fa0857a7cb2ad70979cbdc3e not found: ID does not exist" Dec 01 08:43:32 crc kubenswrapper[4813]: E1201 08:43:32.336665 4813 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.150:6443: connect: connection refused" interval="400ms" Dec 01 08:43:32 crc kubenswrapper[4813]: I1201 08:43:32.402085 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f4b27818a5e8e43d0dc095d08835c792" path="/var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/volumes" Dec 01 08:43:32 crc kubenswrapper[4813]: I1201 08:43:32.441618 4813 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.150:6443: connect: connection refused" Dec 01 08:43:32 crc kubenswrapper[4813]: I1201 08:43:32.441874 4813 status_manager.go:851] "Failed to get status for pod" podUID="269d86e5-dc67-48b3-a955-201b986f6c97" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.150:6443: connect: connection refused" Dec 01 08:43:32 crc kubenswrapper[4813]: E1201 08:43:32.737385 4813 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.150:6443: connect: connection refused" interval="800ms" Dec 01 08:43:33 crc kubenswrapper[4813]: I1201 08:43:33.031110 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x6qcw\" (UniqueName: \"kubernetes.io/projected/9781d8cf-9f56-4d33-b05d-0877b490f389-kube-api-access-x6qcw\") pod \"oauth-openshift-65c4c84884-2vrp6\" (UID: \"9781d8cf-9f56-4d33-b05d-0877b490f389\") " pod="openshift-authentication/oauth-openshift-65c4c84884-2vrp6" Dec 01 08:43:33 crc kubenswrapper[4813]: E1201 08:43:33.031821 4813 projected.go:194] Error preparing data for projected volume kube-api-access-x6qcw for pod openshift-authentication/oauth-openshift-65c4c84884-2vrp6: failed to fetch token: Post "https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/serviceaccounts/oauth-openshift/token": dial tcp 38.102.83.150:6443: connect: connection refused Dec 01 08:43:33 crc kubenswrapper[4813]: E1201 08:43:33.031901 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9781d8cf-9f56-4d33-b05d-0877b490f389-kube-api-access-x6qcw podName:9781d8cf-9f56-4d33-b05d-0877b490f389 nodeName:}" failed. No retries permitted until 2025-12-01 08:43:37.031879187 +0000 UTC m=+201.654800773 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-x6qcw" (UniqueName: "kubernetes.io/projected/9781d8cf-9f56-4d33-b05d-0877b490f389-kube-api-access-x6qcw") pod "oauth-openshift-65c4c84884-2vrp6" (UID: "9781d8cf-9f56-4d33-b05d-0877b490f389") : failed to fetch token: Post "https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/serviceaccounts/oauth-openshift/token": dial tcp 38.102.83.150:6443: connect: connection refused Dec 01 08:43:33 crc kubenswrapper[4813]: E1201 08:43:33.538960 4813 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.150:6443: connect: connection refused" interval="1.6s" Dec 01 08:43:34 crc kubenswrapper[4813]: E1201 08:43:34.166387 4813 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:43:34Z\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:43:34Z\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:43:34Z\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:43:34Z\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Patch \"https://api-int.crc.testing:6443/api/v1/nodes/crc/status?timeout=10s\": dial tcp 38.102.83.150:6443: connect: connection refused" Dec 01 08:43:34 crc kubenswrapper[4813]: E1201 08:43:34.167064 4813 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.150:6443: connect: connection refused" Dec 01 08:43:34 crc kubenswrapper[4813]: E1201 08:43:34.167658 4813 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.150:6443: connect: connection refused" Dec 01 08:43:34 crc kubenswrapper[4813]: E1201 08:43:34.168095 4813 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.150:6443: connect: connection refused" Dec 01 08:43:34 crc kubenswrapper[4813]: E1201 08:43:34.168519 4813 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.150:6443: connect: connection refused" Dec 01 08:43:34 crc kubenswrapper[4813]: E1201 08:43:34.168556 4813 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 01 08:43:34 crc kubenswrapper[4813]: E1201 08:43:34.278818 4813 kubelet.go:1929] "Failed creating a mirror pod for" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods\": dial tcp 38.102.83.150:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 01 08:43:34 crc kubenswrapper[4813]: I1201 08:43:34.279582 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 01 08:43:34 crc kubenswrapper[4813]: E1201 08:43:34.481814 4813 desired_state_of_world_populator.go:312] "Error processing volume" err="error processing PVC openshift-image-registry/crc-image-registry-storage: failed to fetch PVC from API server: Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-image-registry/persistentvolumeclaims/crc-image-registry-storage\": dial tcp 38.102.83.150:6443: connect: connection refused" pod="openshift-image-registry/image-registry-697d97f7c8-mj64s" volumeName="registry-storage" Dec 01 08:43:35 crc kubenswrapper[4813]: E1201 08:43:35.140164 4813 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.150:6443: connect: connection refused" interval="3.2s" Dec 01 08:43:35 crc kubenswrapper[4813]: I1201 08:43:35.168048 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"c8e6209fa56fb03f0f1bf4859a3d4994d5a1b490494d427c97727ff1cf8a6fdd"} Dec 01 08:43:35 crc kubenswrapper[4813]: I1201 08:43:35.168133 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"05290bbc42329c40d6a5e0b146d892f36e9637bcb7a1452993133c31adb00dd3"} Dec 01 08:43:35 crc kubenswrapper[4813]: E1201 08:43:35.169146 4813 kubelet.go:1929] "Failed creating a mirror pod for" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods\": dial tcp 38.102.83.150:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 01 08:43:35 crc kubenswrapper[4813]: I1201 08:43:35.169292 4813 status_manager.go:851] "Failed to get status for pod" podUID="269d86e5-dc67-48b3-a955-201b986f6c97" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.150:6443: connect: connection refused" Dec 01 08:43:36 crc kubenswrapper[4813]: I1201 08:43:36.399955 4813 status_manager.go:851] "Failed to get status for pod" podUID="269d86e5-dc67-48b3-a955-201b986f6c97" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.150:6443: connect: connection refused" Dec 01 08:43:37 crc kubenswrapper[4813]: I1201 08:43:37.092487 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x6qcw\" (UniqueName: \"kubernetes.io/projected/9781d8cf-9f56-4d33-b05d-0877b490f389-kube-api-access-x6qcw\") pod \"oauth-openshift-65c4c84884-2vrp6\" (UID: \"9781d8cf-9f56-4d33-b05d-0877b490f389\") " pod="openshift-authentication/oauth-openshift-65c4c84884-2vrp6" Dec 01 08:43:37 crc kubenswrapper[4813]: E1201 08:43:37.093286 4813 projected.go:194] Error preparing data for projected volume kube-api-access-x6qcw for pod openshift-authentication/oauth-openshift-65c4c84884-2vrp6: failed to fetch token: Post "https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/serviceaccounts/oauth-openshift/token": dial tcp 38.102.83.150:6443: connect: connection refused Dec 01 08:43:37 crc kubenswrapper[4813]: E1201 08:43:37.093412 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9781d8cf-9f56-4d33-b05d-0877b490f389-kube-api-access-x6qcw podName:9781d8cf-9f56-4d33-b05d-0877b490f389 nodeName:}" failed. No retries permitted until 2025-12-01 08:43:45.093376756 +0000 UTC m=+209.716298382 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-x6qcw" (UniqueName: "kubernetes.io/projected/9781d8cf-9f56-4d33-b05d-0877b490f389-kube-api-access-x6qcw") pod "oauth-openshift-65c4c84884-2vrp6" (UID: "9781d8cf-9f56-4d33-b05d-0877b490f389") : failed to fetch token: Post "https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/serviceaccounts/oauth-openshift/token": dial tcp 38.102.83.150:6443: connect: connection refused Dec 01 08:43:38 crc kubenswrapper[4813]: E1201 08:43:38.342056 4813 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.150:6443: connect: connection refused" interval="6.4s" Dec 01 08:43:39 crc kubenswrapper[4813]: E1201 08:43:39.487271 4813 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/events\": dial tcp 38.102.83.150:6443: connect: connection refused" event="&Event{ObjectMeta:{oauth-openshift-65c4c84884-2vrp6.187d0ae56c45a0ee openshift-authentication 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-authentication,Name:oauth-openshift-65c4c84884-2vrp6,UID:9781d8cf-9f56-4d33-b05d-0877b490f389,APIVersion:v1,ResourceVersion:29305,FieldPath:,},Reason:FailedMount,Message:MountVolume.SetUp failed for volume \"kube-api-access-x6qcw\" : failed to fetch token: Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/serviceaccounts/oauth-openshift/token\": dial tcp 38.102.83.150:6443: connect: connection refused,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-01 08:43:29.479393518 +0000 UTC m=+194.102315104,LastTimestamp:2025-12-01 08:43:29.479393518 +0000 UTC m=+194.102315104,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 01 08:43:43 crc kubenswrapper[4813]: I1201 08:43:43.185668 4813 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/kube-controller-manager namespace/openshift-kube-controller-manager: Readiness probe status=failure output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" start-of-body= Dec 01 08:43:43 crc kubenswrapper[4813]: I1201 08:43:43.186189 4813 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="kube-controller-manager" probeResult="failure" output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" Dec 01 08:43:43 crc kubenswrapper[4813]: I1201 08:43:43.392837 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 08:43:43 crc kubenswrapper[4813]: I1201 08:43:43.394264 4813 status_manager.go:851] "Failed to get status for pod" podUID="269d86e5-dc67-48b3-a955-201b986f6c97" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.150:6443: connect: connection refused" Dec 01 08:43:43 crc kubenswrapper[4813]: I1201 08:43:43.420504 4813 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="349332bd-6b4d-4306-84ae-07019ca3860f" Dec 01 08:43:43 crc kubenswrapper[4813]: I1201 08:43:43.420612 4813 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="349332bd-6b4d-4306-84ae-07019ca3860f" Dec 01 08:43:43 crc kubenswrapper[4813]: E1201 08:43:43.421777 4813 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.150:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 08:43:43 crc kubenswrapper[4813]: I1201 08:43:43.422746 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 08:43:44 crc kubenswrapper[4813]: I1201 08:43:44.256506 4813 generic.go:334] "Generic (PLEG): container finished" podID="71bb4a3aecc4ba5b26c4b7318770ce13" containerID="ccf96e45c89fd12bdebadb9801e71a0a7ec68bd336848b878cddee324ea1aec1" exitCode=0 Dec 01 08:43:44 crc kubenswrapper[4813]: I1201 08:43:44.256636 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerDied","Data":"ccf96e45c89fd12bdebadb9801e71a0a7ec68bd336848b878cddee324ea1aec1"} Dec 01 08:43:44 crc kubenswrapper[4813]: I1201 08:43:44.257182 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"c1c31e51b0b70b999b9984db80573d4f50d297d3ae92ee8d0c8a104f464b1e39"} Dec 01 08:43:44 crc kubenswrapper[4813]: I1201 08:43:44.257701 4813 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="349332bd-6b4d-4306-84ae-07019ca3860f" Dec 01 08:43:44 crc kubenswrapper[4813]: I1201 08:43:44.257731 4813 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="349332bd-6b4d-4306-84ae-07019ca3860f" Dec 01 08:43:44 crc kubenswrapper[4813]: I1201 08:43:44.258398 4813 status_manager.go:851] "Failed to get status for pod" podUID="269d86e5-dc67-48b3-a955-201b986f6c97" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.150:6443: connect: connection refused" Dec 01 08:43:44 crc kubenswrapper[4813]: E1201 08:43:44.258426 4813 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.150:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 08:43:44 crc kubenswrapper[4813]: I1201 08:43:44.263660 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Dec 01 08:43:44 crc kubenswrapper[4813]: I1201 08:43:44.263738 4813 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="33898118bbf639c8e6eae5aafa10386b95795512444ed5f535db3c663f169df2" exitCode=1 Dec 01 08:43:44 crc kubenswrapper[4813]: I1201 08:43:44.263794 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"33898118bbf639c8e6eae5aafa10386b95795512444ed5f535db3c663f169df2"} Dec 01 08:43:44 crc kubenswrapper[4813]: I1201 08:43:44.264731 4813 scope.go:117] "RemoveContainer" containerID="33898118bbf639c8e6eae5aafa10386b95795512444ed5f535db3c663f169df2" Dec 01 08:43:44 crc kubenswrapper[4813]: I1201 08:43:44.265061 4813 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.150:6443: connect: connection refused" Dec 01 08:43:44 crc kubenswrapper[4813]: I1201 08:43:44.265833 4813 status_manager.go:851] "Failed to get status for pod" podUID="269d86e5-dc67-48b3-a955-201b986f6c97" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.150:6443: connect: connection refused" Dec 01 08:43:44 crc kubenswrapper[4813]: E1201 08:43:44.269400 4813 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:43:44Z\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:43:44Z\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:43:44Z\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T08:43:44Z\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Patch \"https://api-int.crc.testing:6443/api/v1/nodes/crc/status?timeout=10s\": dial tcp 38.102.83.150:6443: connect: connection refused" Dec 01 08:43:44 crc kubenswrapper[4813]: E1201 08:43:44.269895 4813 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.150:6443: connect: connection refused" Dec 01 08:43:44 crc kubenswrapper[4813]: E1201 08:43:44.270475 4813 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.150:6443: connect: connection refused" Dec 01 08:43:44 crc kubenswrapper[4813]: E1201 08:43:44.270844 4813 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.150:6443: connect: connection refused" Dec 01 08:43:44 crc kubenswrapper[4813]: E1201 08:43:44.271318 4813 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.150:6443: connect: connection refused" Dec 01 08:43:44 crc kubenswrapper[4813]: E1201 08:43:44.271356 4813 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 01 08:43:44 crc kubenswrapper[4813]: I1201 08:43:44.374478 4813 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 01 08:43:45 crc kubenswrapper[4813]: I1201 08:43:45.118380 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x6qcw\" (UniqueName: \"kubernetes.io/projected/9781d8cf-9f56-4d33-b05d-0877b490f389-kube-api-access-x6qcw\") pod \"oauth-openshift-65c4c84884-2vrp6\" (UID: \"9781d8cf-9f56-4d33-b05d-0877b490f389\") " pod="openshift-authentication/oauth-openshift-65c4c84884-2vrp6" Dec 01 08:43:45 crc kubenswrapper[4813]: I1201 08:43:45.274118 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"3c92c09634dffae6795f62c53deae70de18d1927f1745107b60e6b7df7ab56ce"} Dec 01 08:43:45 crc kubenswrapper[4813]: I1201 08:43:45.274166 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"b76fa045b467210cd52df765936f1c6697c49a5351e2159ecc5d8ef1cea38515"} Dec 01 08:43:45 crc kubenswrapper[4813]: I1201 08:43:45.274178 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"2d49d98ae8db588edfd8f3d8a49171dfc25e127ceecbd567451c2a349bb090d0"} Dec 01 08:43:45 crc kubenswrapper[4813]: I1201 08:43:45.281257 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Dec 01 08:43:45 crc kubenswrapper[4813]: I1201 08:43:45.281319 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"8acd65173833b3b53cd0efc3f17123c8e951c4d654cdb6b65be080732e387715"} Dec 01 08:43:46 crc kubenswrapper[4813]: I1201 08:43:46.291147 4813 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="349332bd-6b4d-4306-84ae-07019ca3860f" Dec 01 08:43:46 crc kubenswrapper[4813]: I1201 08:43:46.291175 4813 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="349332bd-6b4d-4306-84ae-07019ca3860f" Dec 01 08:43:46 crc kubenswrapper[4813]: I1201 08:43:46.291332 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"f0160b84b9b3f3675372ba710b37b77689dcad3aad3ffa0e70208eae84a2696a"} Dec 01 08:43:46 crc kubenswrapper[4813]: I1201 08:43:46.291353 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"1d87869b061c64aa1fdb31be09120eff6d9a5d99b539029becfb13bfc5b2641f"} Dec 01 08:43:46 crc kubenswrapper[4813]: I1201 08:43:46.291393 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 08:43:47 crc kubenswrapper[4813]: I1201 08:43:47.243411 4813 patch_prober.go:28] interesting pod/machine-config-daemon-9wn4q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 08:43:47 crc kubenswrapper[4813]: I1201 08:43:47.243778 4813 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9wn4q" podUID="3d423402-e08f-42e9-b60c-6eb9bf2f6d48" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 08:43:47 crc kubenswrapper[4813]: I1201 08:43:47.243838 4813 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-9wn4q" Dec 01 08:43:47 crc kubenswrapper[4813]: I1201 08:43:47.244398 4813 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"1c915733072b32a63237e2401cbf84b059aca396c2c0db884f0f439d599a35d2"} pod="openshift-machine-config-operator/machine-config-daemon-9wn4q" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 01 08:43:47 crc kubenswrapper[4813]: I1201 08:43:47.244495 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-9wn4q" podUID="3d423402-e08f-42e9-b60c-6eb9bf2f6d48" containerName="machine-config-daemon" containerID="cri-o://1c915733072b32a63237e2401cbf84b059aca396c2c0db884f0f439d599a35d2" gracePeriod=600 Dec 01 08:43:48 crc kubenswrapper[4813]: I1201 08:43:48.313657 4813 generic.go:334] "Generic (PLEG): container finished" podID="3d423402-e08f-42e9-b60c-6eb9bf2f6d48" containerID="1c915733072b32a63237e2401cbf84b059aca396c2c0db884f0f439d599a35d2" exitCode=0 Dec 01 08:43:48 crc kubenswrapper[4813]: I1201 08:43:48.313791 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-9wn4q" event={"ID":"3d423402-e08f-42e9-b60c-6eb9bf2f6d48","Type":"ContainerDied","Data":"1c915733072b32a63237e2401cbf84b059aca396c2c0db884f0f439d599a35d2"} Dec 01 08:43:48 crc kubenswrapper[4813]: I1201 08:43:48.314078 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-9wn4q" event={"ID":"3d423402-e08f-42e9-b60c-6eb9bf2f6d48","Type":"ContainerStarted","Data":"5899e07de4e292f45ea6ae9799d87cd4fae3236f41b153e0795a33f5f1ed93ff"} Dec 01 08:43:48 crc kubenswrapper[4813]: I1201 08:43:48.423830 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 08:43:48 crc kubenswrapper[4813]: I1201 08:43:48.424029 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 08:43:48 crc kubenswrapper[4813]: I1201 08:43:48.432342 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 08:43:51 crc kubenswrapper[4813]: I1201 08:43:51.232797 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x6qcw\" (UniqueName: \"kubernetes.io/projected/9781d8cf-9f56-4d33-b05d-0877b490f389-kube-api-access-x6qcw\") pod \"oauth-openshift-65c4c84884-2vrp6\" (UID: \"9781d8cf-9f56-4d33-b05d-0877b490f389\") " pod="openshift-authentication/oauth-openshift-65c4c84884-2vrp6" Dec 01 08:43:51 crc kubenswrapper[4813]: I1201 08:43:51.345914 4813 kubelet.go:1914] "Deleted mirror pod because it is outdated" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 08:43:51 crc kubenswrapper[4813]: I1201 08:43:51.388948 4813 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="0d8da9da-be01-4894-8610-e8c01ce383f2" Dec 01 08:43:51 crc kubenswrapper[4813]: I1201 08:43:51.425914 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-65c4c84884-2vrp6" Dec 01 08:43:52 crc kubenswrapper[4813]: I1201 08:43:52.355554 4813 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="349332bd-6b4d-4306-84ae-07019ca3860f" Dec 01 08:43:52 crc kubenswrapper[4813]: I1201 08:43:52.355955 4813 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="349332bd-6b4d-4306-84ae-07019ca3860f" Dec 01 08:43:52 crc kubenswrapper[4813]: I1201 08:43:52.355805 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-65c4c84884-2vrp6" event={"ID":"9781d8cf-9f56-4d33-b05d-0877b490f389","Type":"ContainerStarted","Data":"ead0d60d0409d152270ec3205ef5965217503e249edadc87b6ab7dc56e97821b"} Dec 01 08:43:52 crc kubenswrapper[4813]: I1201 08:43:52.356173 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-65c4c84884-2vrp6" event={"ID":"9781d8cf-9f56-4d33-b05d-0877b490f389","Type":"ContainerStarted","Data":"e3e1771ae7ededb970643f964452e013945d0d08960c3f5a373cbb6f1900f93c"} Dec 01 08:43:52 crc kubenswrapper[4813]: I1201 08:43:52.356240 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-65c4c84884-2vrp6" Dec 01 08:43:52 crc kubenswrapper[4813]: I1201 08:43:52.360833 4813 patch_prober.go:28] interesting pod/oauth-openshift-65c4c84884-2vrp6 container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.56:6443/healthz\": dial tcp 10.217.0.56:6443: connect: connection refused" start-of-body= Dec 01 08:43:52 crc kubenswrapper[4813]: I1201 08:43:52.361006 4813 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-65c4c84884-2vrp6" podUID="9781d8cf-9f56-4d33-b05d-0877b490f389" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.56:6443/healthz\": dial tcp 10.217.0.56:6443: connect: connection refused" Dec 01 08:43:52 crc kubenswrapper[4813]: I1201 08:43:52.361171 4813 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="0d8da9da-be01-4894-8610-e8c01ce383f2" Dec 01 08:43:52 crc kubenswrapper[4813]: I1201 08:43:52.361454 4813 status_manager.go:308] "Container readiness changed before pod has synced" pod="openshift-kube-apiserver/kube-apiserver-crc" containerID="cri-o://2d49d98ae8db588edfd8f3d8a49171dfc25e127ceecbd567451c2a349bb090d0" Dec 01 08:43:52 crc kubenswrapper[4813]: I1201 08:43:52.361488 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 08:43:53 crc kubenswrapper[4813]: I1201 08:43:53.184355 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 01 08:43:53 crc kubenswrapper[4813]: I1201 08:43:53.364029 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-authentication_oauth-openshift-65c4c84884-2vrp6_9781d8cf-9f56-4d33-b05d-0877b490f389/oauth-openshift/0.log" Dec 01 08:43:53 crc kubenswrapper[4813]: I1201 08:43:53.364084 4813 generic.go:334] "Generic (PLEG): container finished" podID="9781d8cf-9f56-4d33-b05d-0877b490f389" containerID="ead0d60d0409d152270ec3205ef5965217503e249edadc87b6ab7dc56e97821b" exitCode=255 Dec 01 08:43:53 crc kubenswrapper[4813]: I1201 08:43:53.364525 4813 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="349332bd-6b4d-4306-84ae-07019ca3860f" Dec 01 08:43:53 crc kubenswrapper[4813]: I1201 08:43:53.364539 4813 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="349332bd-6b4d-4306-84ae-07019ca3860f" Dec 01 08:43:53 crc kubenswrapper[4813]: I1201 08:43:53.364274 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-65c4c84884-2vrp6" event={"ID":"9781d8cf-9f56-4d33-b05d-0877b490f389","Type":"ContainerDied","Data":"ead0d60d0409d152270ec3205ef5965217503e249edadc87b6ab7dc56e97821b"} Dec 01 08:43:53 crc kubenswrapper[4813]: I1201 08:43:53.364750 4813 scope.go:117] "RemoveContainer" containerID="ead0d60d0409d152270ec3205ef5965217503e249edadc87b6ab7dc56e97821b" Dec 01 08:43:53 crc kubenswrapper[4813]: I1201 08:43:53.386637 4813 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="0d8da9da-be01-4894-8610-e8c01ce383f2" Dec 01 08:43:53 crc kubenswrapper[4813]: I1201 08:43:53.694532 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 01 08:43:53 crc kubenswrapper[4813]: I1201 08:43:53.698137 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 01 08:43:54 crc kubenswrapper[4813]: I1201 08:43:54.372387 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-authentication_oauth-openshift-65c4c84884-2vrp6_9781d8cf-9f56-4d33-b05d-0877b490f389/oauth-openshift/1.log" Dec 01 08:43:54 crc kubenswrapper[4813]: I1201 08:43:54.373010 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-authentication_oauth-openshift-65c4c84884-2vrp6_9781d8cf-9f56-4d33-b05d-0877b490f389/oauth-openshift/0.log" Dec 01 08:43:54 crc kubenswrapper[4813]: I1201 08:43:54.373051 4813 generic.go:334] "Generic (PLEG): container finished" podID="9781d8cf-9f56-4d33-b05d-0877b490f389" containerID="cb26e0972ffc5a5896767a7b439651cebc260ceb18183d3a16b4cf0d87dc81ff" exitCode=255 Dec 01 08:43:54 crc kubenswrapper[4813]: I1201 08:43:54.373227 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-65c4c84884-2vrp6" event={"ID":"9781d8cf-9f56-4d33-b05d-0877b490f389","Type":"ContainerDied","Data":"cb26e0972ffc5a5896767a7b439651cebc260ceb18183d3a16b4cf0d87dc81ff"} Dec 01 08:43:54 crc kubenswrapper[4813]: I1201 08:43:54.373314 4813 scope.go:117] "RemoveContainer" containerID="ead0d60d0409d152270ec3205ef5965217503e249edadc87b6ab7dc56e97821b" Dec 01 08:43:54 crc kubenswrapper[4813]: I1201 08:43:54.373890 4813 scope.go:117] "RemoveContainer" containerID="cb26e0972ffc5a5896767a7b439651cebc260ceb18183d3a16b4cf0d87dc81ff" Dec 01 08:43:54 crc kubenswrapper[4813]: E1201 08:43:54.374238 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oauth-openshift\" with CrashLoopBackOff: \"back-off 10s restarting failed container=oauth-openshift pod=oauth-openshift-65c4c84884-2vrp6_openshift-authentication(9781d8cf-9f56-4d33-b05d-0877b490f389)\"" pod="openshift-authentication/oauth-openshift-65c4c84884-2vrp6" podUID="9781d8cf-9f56-4d33-b05d-0877b490f389" Dec 01 08:43:54 crc kubenswrapper[4813]: I1201 08:43:54.378794 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 01 08:43:55 crc kubenswrapper[4813]: I1201 08:43:55.383524 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-authentication_oauth-openshift-65c4c84884-2vrp6_9781d8cf-9f56-4d33-b05d-0877b490f389/oauth-openshift/1.log" Dec 01 08:43:55 crc kubenswrapper[4813]: I1201 08:43:55.384917 4813 scope.go:117] "RemoveContainer" containerID="cb26e0972ffc5a5896767a7b439651cebc260ceb18183d3a16b4cf0d87dc81ff" Dec 01 08:43:55 crc kubenswrapper[4813]: E1201 08:43:55.385581 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oauth-openshift\" with CrashLoopBackOff: \"back-off 10s restarting failed container=oauth-openshift pod=oauth-openshift-65c4c84884-2vrp6_openshift-authentication(9781d8cf-9f56-4d33-b05d-0877b490f389)\"" pod="openshift-authentication/oauth-openshift-65c4c84884-2vrp6" podUID="9781d8cf-9f56-4d33-b05d-0877b490f389" Dec 01 08:44:01 crc kubenswrapper[4813]: I1201 08:44:01.427442 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-65c4c84884-2vrp6" Dec 01 08:44:01 crc kubenswrapper[4813]: I1201 08:44:01.428288 4813 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-authentication/oauth-openshift-65c4c84884-2vrp6" Dec 01 08:44:01 crc kubenswrapper[4813]: I1201 08:44:01.429067 4813 scope.go:117] "RemoveContainer" containerID="cb26e0972ffc5a5896767a7b439651cebc260ceb18183d3a16b4cf0d87dc81ff" Dec 01 08:44:01 crc kubenswrapper[4813]: E1201 08:44:01.429483 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oauth-openshift\" with CrashLoopBackOff: \"back-off 10s restarting failed container=oauth-openshift pod=oauth-openshift-65c4c84884-2vrp6_openshift-authentication(9781d8cf-9f56-4d33-b05d-0877b490f389)\"" pod="openshift-authentication/oauth-openshift-65c4c84884-2vrp6" podUID="9781d8cf-9f56-4d33-b05d-0877b490f389" Dec 01 08:44:01 crc kubenswrapper[4813]: I1201 08:44:01.718421 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Dec 01 08:44:01 crc kubenswrapper[4813]: I1201 08:44:01.751792 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Dec 01 08:44:01 crc kubenswrapper[4813]: I1201 08:44:01.919813 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Dec 01 08:44:01 crc kubenswrapper[4813]: I1201 08:44:01.971350 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Dec 01 08:44:02 crc kubenswrapper[4813]: I1201 08:44:02.246101 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Dec 01 08:44:02 crc kubenswrapper[4813]: I1201 08:44:02.429484 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Dec 01 08:44:02 crc kubenswrapper[4813]: I1201 08:44:02.469344 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Dec 01 08:44:02 crc kubenswrapper[4813]: I1201 08:44:02.602777 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Dec 01 08:44:02 crc kubenswrapper[4813]: I1201 08:44:02.646509 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Dec 01 08:44:02 crc kubenswrapper[4813]: I1201 08:44:02.729448 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Dec 01 08:44:03 crc kubenswrapper[4813]: I1201 08:44:03.035688 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Dec 01 08:44:03 crc kubenswrapper[4813]: I1201 08:44:03.037759 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Dec 01 08:44:03 crc kubenswrapper[4813]: I1201 08:44:03.083309 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Dec 01 08:44:03 crc kubenswrapper[4813]: I1201 08:44:03.121934 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Dec 01 08:44:03 crc kubenswrapper[4813]: I1201 08:44:03.248772 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Dec 01 08:44:03 crc kubenswrapper[4813]: I1201 08:44:03.398188 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 01 08:44:03 crc kubenswrapper[4813]: I1201 08:44:03.590169 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Dec 01 08:44:03 crc kubenswrapper[4813]: I1201 08:44:03.614267 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Dec 01 08:44:03 crc kubenswrapper[4813]: I1201 08:44:03.855090 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Dec 01 08:44:03 crc kubenswrapper[4813]: I1201 08:44:03.927804 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Dec 01 08:44:04 crc kubenswrapper[4813]: I1201 08:44:04.389660 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Dec 01 08:44:04 crc kubenswrapper[4813]: I1201 08:44:04.412043 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Dec 01 08:44:04 crc kubenswrapper[4813]: I1201 08:44:04.476843 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Dec 01 08:44:04 crc kubenswrapper[4813]: I1201 08:44:04.581339 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Dec 01 08:44:05 crc kubenswrapper[4813]: I1201 08:44:05.021762 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Dec 01 08:44:05 crc kubenswrapper[4813]: I1201 08:44:05.030454 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Dec 01 08:44:05 crc kubenswrapper[4813]: I1201 08:44:05.161116 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Dec 01 08:44:05 crc kubenswrapper[4813]: I1201 08:44:05.198765 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Dec 01 08:44:05 crc kubenswrapper[4813]: I1201 08:44:05.286787 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Dec 01 08:44:05 crc kubenswrapper[4813]: I1201 08:44:05.373110 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Dec 01 08:44:05 crc kubenswrapper[4813]: I1201 08:44:05.447151 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Dec 01 08:44:05 crc kubenswrapper[4813]: I1201 08:44:05.686531 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Dec 01 08:44:05 crc kubenswrapper[4813]: I1201 08:44:05.688343 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Dec 01 08:44:05 crc kubenswrapper[4813]: I1201 08:44:05.729876 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Dec 01 08:44:05 crc kubenswrapper[4813]: I1201 08:44:05.766342 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Dec 01 08:44:05 crc kubenswrapper[4813]: I1201 08:44:05.773565 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Dec 01 08:44:05 crc kubenswrapper[4813]: I1201 08:44:05.822070 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Dec 01 08:44:05 crc kubenswrapper[4813]: I1201 08:44:05.836836 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Dec 01 08:44:05 crc kubenswrapper[4813]: I1201 08:44:05.907065 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Dec 01 08:44:05 crc kubenswrapper[4813]: I1201 08:44:05.910239 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Dec 01 08:44:05 crc kubenswrapper[4813]: I1201 08:44:05.966801 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Dec 01 08:44:05 crc kubenswrapper[4813]: I1201 08:44:05.993691 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Dec 01 08:44:05 crc kubenswrapper[4813]: I1201 08:44:05.995064 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Dec 01 08:44:06 crc kubenswrapper[4813]: I1201 08:44:06.056077 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Dec 01 08:44:06 crc kubenswrapper[4813]: I1201 08:44:06.078064 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Dec 01 08:44:06 crc kubenswrapper[4813]: I1201 08:44:06.133407 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Dec 01 08:44:06 crc kubenswrapper[4813]: I1201 08:44:06.211913 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Dec 01 08:44:06 crc kubenswrapper[4813]: I1201 08:44:06.241055 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 01 08:44:06 crc kubenswrapper[4813]: I1201 08:44:06.397521 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Dec 01 08:44:06 crc kubenswrapper[4813]: I1201 08:44:06.440191 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Dec 01 08:44:06 crc kubenswrapper[4813]: I1201 08:44:06.451025 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Dec 01 08:44:06 crc kubenswrapper[4813]: I1201 08:44:06.530218 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Dec 01 08:44:06 crc kubenswrapper[4813]: I1201 08:44:06.567499 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Dec 01 08:44:06 crc kubenswrapper[4813]: I1201 08:44:06.574861 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Dec 01 08:44:06 crc kubenswrapper[4813]: I1201 08:44:06.588196 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Dec 01 08:44:06 crc kubenswrapper[4813]: I1201 08:44:06.601191 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Dec 01 08:44:06 crc kubenswrapper[4813]: I1201 08:44:06.677468 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Dec 01 08:44:06 crc kubenswrapper[4813]: I1201 08:44:06.679033 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Dec 01 08:44:06 crc kubenswrapper[4813]: I1201 08:44:06.769057 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Dec 01 08:44:06 crc kubenswrapper[4813]: I1201 08:44:06.852577 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Dec 01 08:44:06 crc kubenswrapper[4813]: I1201 08:44:06.882783 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Dec 01 08:44:06 crc kubenswrapper[4813]: I1201 08:44:06.883218 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Dec 01 08:44:06 crc kubenswrapper[4813]: I1201 08:44:06.913427 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Dec 01 08:44:07 crc kubenswrapper[4813]: I1201 08:44:07.003529 4813 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Dec 01 08:44:07 crc kubenswrapper[4813]: I1201 08:44:07.098377 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Dec 01 08:44:07 crc kubenswrapper[4813]: I1201 08:44:07.135668 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Dec 01 08:44:07 crc kubenswrapper[4813]: I1201 08:44:07.151659 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Dec 01 08:44:07 crc kubenswrapper[4813]: I1201 08:44:07.190266 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Dec 01 08:44:07 crc kubenswrapper[4813]: I1201 08:44:07.232794 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Dec 01 08:44:07 crc kubenswrapper[4813]: I1201 08:44:07.255107 4813 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Dec 01 08:44:07 crc kubenswrapper[4813]: I1201 08:44:07.310114 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Dec 01 08:44:07 crc kubenswrapper[4813]: I1201 08:44:07.338077 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 01 08:44:07 crc kubenswrapper[4813]: I1201 08:44:07.348151 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Dec 01 08:44:07 crc kubenswrapper[4813]: I1201 08:44:07.395199 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Dec 01 08:44:07 crc kubenswrapper[4813]: I1201 08:44:07.489878 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Dec 01 08:44:07 crc kubenswrapper[4813]: I1201 08:44:07.540917 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Dec 01 08:44:07 crc kubenswrapper[4813]: I1201 08:44:07.550547 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Dec 01 08:44:07 crc kubenswrapper[4813]: I1201 08:44:07.591611 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Dec 01 08:44:07 crc kubenswrapper[4813]: I1201 08:44:07.727665 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Dec 01 08:44:07 crc kubenswrapper[4813]: I1201 08:44:07.759631 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Dec 01 08:44:07 crc kubenswrapper[4813]: I1201 08:44:07.805989 4813 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Dec 01 08:44:07 crc kubenswrapper[4813]: I1201 08:44:07.881959 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Dec 01 08:44:07 crc kubenswrapper[4813]: I1201 08:44:07.940958 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Dec 01 08:44:07 crc kubenswrapper[4813]: I1201 08:44:07.950368 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Dec 01 08:44:08 crc kubenswrapper[4813]: I1201 08:44:08.182081 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Dec 01 08:44:08 crc kubenswrapper[4813]: I1201 08:44:08.203128 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Dec 01 08:44:08 crc kubenswrapper[4813]: I1201 08:44:08.317880 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Dec 01 08:44:08 crc kubenswrapper[4813]: I1201 08:44:08.318956 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Dec 01 08:44:08 crc kubenswrapper[4813]: I1201 08:44:08.335541 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Dec 01 08:44:08 crc kubenswrapper[4813]: I1201 08:44:08.342483 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Dec 01 08:44:08 crc kubenswrapper[4813]: I1201 08:44:08.454649 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Dec 01 08:44:08 crc kubenswrapper[4813]: I1201 08:44:08.549021 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Dec 01 08:44:08 crc kubenswrapper[4813]: I1201 08:44:08.550642 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Dec 01 08:44:08 crc kubenswrapper[4813]: I1201 08:44:08.554587 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Dec 01 08:44:08 crc kubenswrapper[4813]: I1201 08:44:08.573409 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Dec 01 08:44:08 crc kubenswrapper[4813]: I1201 08:44:08.648734 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Dec 01 08:44:08 crc kubenswrapper[4813]: I1201 08:44:08.650687 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Dec 01 08:44:08 crc kubenswrapper[4813]: I1201 08:44:08.656924 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Dec 01 08:44:08 crc kubenswrapper[4813]: I1201 08:44:08.662077 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Dec 01 08:44:08 crc kubenswrapper[4813]: I1201 08:44:08.708459 4813 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Dec 01 08:44:08 crc kubenswrapper[4813]: I1201 08:44:08.729356 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Dec 01 08:44:08 crc kubenswrapper[4813]: I1201 08:44:08.729417 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Dec 01 08:44:08 crc kubenswrapper[4813]: I1201 08:44:08.737479 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Dec 01 08:44:08 crc kubenswrapper[4813]: I1201 08:44:08.791124 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Dec 01 08:44:08 crc kubenswrapper[4813]: I1201 08:44:08.811773 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Dec 01 08:44:08 crc kubenswrapper[4813]: I1201 08:44:08.922449 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Dec 01 08:44:08 crc kubenswrapper[4813]: I1201 08:44:08.945220 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Dec 01 08:44:08 crc kubenswrapper[4813]: I1201 08:44:08.967024 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Dec 01 08:44:08 crc kubenswrapper[4813]: I1201 08:44:08.969552 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Dec 01 08:44:08 crc kubenswrapper[4813]: I1201 08:44:08.980508 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Dec 01 08:44:08 crc kubenswrapper[4813]: I1201 08:44:08.998744 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Dec 01 08:44:09 crc kubenswrapper[4813]: I1201 08:44:09.008824 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Dec 01 08:44:09 crc kubenswrapper[4813]: I1201 08:44:09.075435 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Dec 01 08:44:09 crc kubenswrapper[4813]: I1201 08:44:09.088860 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Dec 01 08:44:09 crc kubenswrapper[4813]: I1201 08:44:09.182760 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Dec 01 08:44:09 crc kubenswrapper[4813]: I1201 08:44:09.212552 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Dec 01 08:44:09 crc kubenswrapper[4813]: I1201 08:44:09.219157 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Dec 01 08:44:09 crc kubenswrapper[4813]: I1201 08:44:09.342868 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Dec 01 08:44:09 crc kubenswrapper[4813]: I1201 08:44:09.371332 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Dec 01 08:44:09 crc kubenswrapper[4813]: I1201 08:44:09.405683 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Dec 01 08:44:09 crc kubenswrapper[4813]: I1201 08:44:09.431832 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Dec 01 08:44:09 crc kubenswrapper[4813]: I1201 08:44:09.467065 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Dec 01 08:44:09 crc kubenswrapper[4813]: I1201 08:44:09.508571 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Dec 01 08:44:09 crc kubenswrapper[4813]: I1201 08:44:09.516099 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Dec 01 08:44:09 crc kubenswrapper[4813]: I1201 08:44:09.615166 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Dec 01 08:44:09 crc kubenswrapper[4813]: I1201 08:44:09.764654 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Dec 01 08:44:09 crc kubenswrapper[4813]: I1201 08:44:09.829472 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Dec 01 08:44:09 crc kubenswrapper[4813]: I1201 08:44:09.881433 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Dec 01 08:44:09 crc kubenswrapper[4813]: I1201 08:44:09.914556 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Dec 01 08:44:09 crc kubenswrapper[4813]: I1201 08:44:09.915927 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Dec 01 08:44:09 crc kubenswrapper[4813]: I1201 08:44:09.973185 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Dec 01 08:44:10 crc kubenswrapper[4813]: I1201 08:44:10.033055 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Dec 01 08:44:10 crc kubenswrapper[4813]: I1201 08:44:10.149287 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Dec 01 08:44:10 crc kubenswrapper[4813]: I1201 08:44:10.168518 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Dec 01 08:44:10 crc kubenswrapper[4813]: I1201 08:44:10.231112 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Dec 01 08:44:10 crc kubenswrapper[4813]: I1201 08:44:10.298473 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Dec 01 08:44:10 crc kubenswrapper[4813]: I1201 08:44:10.342117 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Dec 01 08:44:10 crc kubenswrapper[4813]: I1201 08:44:10.369598 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Dec 01 08:44:10 crc kubenswrapper[4813]: I1201 08:44:10.460940 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Dec 01 08:44:10 crc kubenswrapper[4813]: I1201 08:44:10.482715 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Dec 01 08:44:10 crc kubenswrapper[4813]: I1201 08:44:10.529408 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Dec 01 08:44:10 crc kubenswrapper[4813]: I1201 08:44:10.571755 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Dec 01 08:44:10 crc kubenswrapper[4813]: I1201 08:44:10.577613 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Dec 01 08:44:10 crc kubenswrapper[4813]: I1201 08:44:10.598411 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Dec 01 08:44:10 crc kubenswrapper[4813]: I1201 08:44:10.607364 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Dec 01 08:44:10 crc kubenswrapper[4813]: I1201 08:44:10.610783 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Dec 01 08:44:10 crc kubenswrapper[4813]: I1201 08:44:10.658111 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Dec 01 08:44:10 crc kubenswrapper[4813]: I1201 08:44:10.683824 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Dec 01 08:44:10 crc kubenswrapper[4813]: I1201 08:44:10.754998 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Dec 01 08:44:10 crc kubenswrapper[4813]: I1201 08:44:10.827142 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Dec 01 08:44:10 crc kubenswrapper[4813]: I1201 08:44:10.840349 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Dec 01 08:44:10 crc kubenswrapper[4813]: I1201 08:44:10.862599 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Dec 01 08:44:10 crc kubenswrapper[4813]: I1201 08:44:10.874329 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Dec 01 08:44:10 crc kubenswrapper[4813]: I1201 08:44:10.885288 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 01 08:44:10 crc kubenswrapper[4813]: I1201 08:44:10.918424 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Dec 01 08:44:11 crc kubenswrapper[4813]: I1201 08:44:11.021844 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Dec 01 08:44:11 crc kubenswrapper[4813]: I1201 08:44:11.058321 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Dec 01 08:44:11 crc kubenswrapper[4813]: I1201 08:44:11.075286 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Dec 01 08:44:11 crc kubenswrapper[4813]: I1201 08:44:11.122016 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Dec 01 08:44:11 crc kubenswrapper[4813]: I1201 08:44:11.162658 4813 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Dec 01 08:44:11 crc kubenswrapper[4813]: I1201 08:44:11.191949 4813 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Dec 01 08:44:11 crc kubenswrapper[4813]: I1201 08:44:11.195220 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Dec 01 08:44:11 crc kubenswrapper[4813]: I1201 08:44:11.196338 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 01 08:44:11 crc kubenswrapper[4813]: I1201 08:44:11.196402 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 01 08:44:11 crc kubenswrapper[4813]: I1201 08:44:11.196422 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-65c4c84884-2vrp6"] Dec 01 08:44:11 crc kubenswrapper[4813]: I1201 08:44:11.197013 4813 scope.go:117] "RemoveContainer" containerID="cb26e0972ffc5a5896767a7b439651cebc260ceb18183d3a16b4cf0d87dc81ff" Dec 01 08:44:11 crc kubenswrapper[4813]: I1201 08:44:11.209914 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 08:44:11 crc kubenswrapper[4813]: I1201 08:44:11.223275 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=20.223236925 podStartE2EDuration="20.223236925s" podCreationTimestamp="2025-12-01 08:43:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:44:11.2230708 +0000 UTC m=+235.845992426" watchObservedRunningTime="2025-12-01 08:44:11.223236925 +0000 UTC m=+235.846158511" Dec 01 08:44:11 crc kubenswrapper[4813]: I1201 08:44:11.324260 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Dec 01 08:44:11 crc kubenswrapper[4813]: I1201 08:44:11.355850 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Dec 01 08:44:11 crc kubenswrapper[4813]: I1201 08:44:11.401842 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Dec 01 08:44:11 crc kubenswrapper[4813]: I1201 08:44:11.457348 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Dec 01 08:44:11 crc kubenswrapper[4813]: I1201 08:44:11.470322 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Dec 01 08:44:11 crc kubenswrapper[4813]: I1201 08:44:11.485661 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Dec 01 08:44:11 crc kubenswrapper[4813]: I1201 08:44:11.497783 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-authentication_oauth-openshift-65c4c84884-2vrp6_9781d8cf-9f56-4d33-b05d-0877b490f389/oauth-openshift/1.log" Dec 01 08:44:11 crc kubenswrapper[4813]: I1201 08:44:11.497886 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-65c4c84884-2vrp6" event={"ID":"9781d8cf-9f56-4d33-b05d-0877b490f389","Type":"ContainerStarted","Data":"fdffc17417005273918369fe2cb746bf45b969faa1bc1972899dc28bef15fe76"} Dec 01 08:44:11 crc kubenswrapper[4813]: I1201 08:44:11.498438 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-65c4c84884-2vrp6" Dec 01 08:44:11 crc kubenswrapper[4813]: I1201 08:44:11.500006 4813 patch_prober.go:28] interesting pod/oauth-openshift-65c4c84884-2vrp6 container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.56:6443/healthz\": dial tcp 10.217.0.56:6443: connect: connection refused" start-of-body= Dec 01 08:44:11 crc kubenswrapper[4813]: I1201 08:44:11.500056 4813 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-65c4c84884-2vrp6" podUID="9781d8cf-9f56-4d33-b05d-0877b490f389" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.56:6443/healthz\": dial tcp 10.217.0.56:6443: connect: connection refused" Dec 01 08:44:11 crc kubenswrapper[4813]: I1201 08:44:11.506178 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Dec 01 08:44:11 crc kubenswrapper[4813]: I1201 08:44:11.518756 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-65c4c84884-2vrp6" podStartSLOduration=78.518728725 podStartE2EDuration="1m18.518728725s" podCreationTimestamp="2025-12-01 08:42:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:43:52.384055643 +0000 UTC m=+217.006977229" watchObservedRunningTime="2025-12-01 08:44:11.518728725 +0000 UTC m=+236.141650311" Dec 01 08:44:11 crc kubenswrapper[4813]: I1201 08:44:11.846386 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Dec 01 08:44:11 crc kubenswrapper[4813]: I1201 08:44:11.867389 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Dec 01 08:44:11 crc kubenswrapper[4813]: I1201 08:44:11.868805 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Dec 01 08:44:11 crc kubenswrapper[4813]: I1201 08:44:11.883894 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Dec 01 08:44:11 crc kubenswrapper[4813]: I1201 08:44:11.954321 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Dec 01 08:44:11 crc kubenswrapper[4813]: I1201 08:44:11.962335 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Dec 01 08:44:12 crc kubenswrapper[4813]: I1201 08:44:12.107277 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Dec 01 08:44:12 crc kubenswrapper[4813]: I1201 08:44:12.221174 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Dec 01 08:44:12 crc kubenswrapper[4813]: I1201 08:44:12.243045 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Dec 01 08:44:12 crc kubenswrapper[4813]: I1201 08:44:12.281132 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Dec 01 08:44:12 crc kubenswrapper[4813]: I1201 08:44:12.296813 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Dec 01 08:44:12 crc kubenswrapper[4813]: I1201 08:44:12.308512 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Dec 01 08:44:12 crc kubenswrapper[4813]: I1201 08:44:12.331671 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Dec 01 08:44:12 crc kubenswrapper[4813]: I1201 08:44:12.354483 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Dec 01 08:44:12 crc kubenswrapper[4813]: I1201 08:44:12.426877 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Dec 01 08:44:12 crc kubenswrapper[4813]: I1201 08:44:12.428095 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Dec 01 08:44:12 crc kubenswrapper[4813]: I1201 08:44:12.435693 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Dec 01 08:44:12 crc kubenswrapper[4813]: I1201 08:44:12.454611 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Dec 01 08:44:12 crc kubenswrapper[4813]: I1201 08:44:12.510439 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-65c4c84884-2vrp6" Dec 01 08:44:12 crc kubenswrapper[4813]: I1201 08:44:12.546291 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Dec 01 08:44:12 crc kubenswrapper[4813]: I1201 08:44:12.570869 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Dec 01 08:44:12 crc kubenswrapper[4813]: I1201 08:44:12.641944 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Dec 01 08:44:12 crc kubenswrapper[4813]: I1201 08:44:12.688716 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Dec 01 08:44:12 crc kubenswrapper[4813]: I1201 08:44:12.713589 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Dec 01 08:44:12 crc kubenswrapper[4813]: I1201 08:44:12.770306 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Dec 01 08:44:12 crc kubenswrapper[4813]: I1201 08:44:12.845195 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Dec 01 08:44:12 crc kubenswrapper[4813]: I1201 08:44:12.867953 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Dec 01 08:44:12 crc kubenswrapper[4813]: I1201 08:44:12.927477 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Dec 01 08:44:13 crc kubenswrapper[4813]: I1201 08:44:13.007897 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Dec 01 08:44:13 crc kubenswrapper[4813]: I1201 08:44:13.130703 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Dec 01 08:44:13 crc kubenswrapper[4813]: I1201 08:44:13.130814 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Dec 01 08:44:13 crc kubenswrapper[4813]: I1201 08:44:13.142950 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Dec 01 08:44:13 crc kubenswrapper[4813]: I1201 08:44:13.190403 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Dec 01 08:44:13 crc kubenswrapper[4813]: I1201 08:44:13.210373 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 01 08:44:13 crc kubenswrapper[4813]: I1201 08:44:13.310288 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Dec 01 08:44:13 crc kubenswrapper[4813]: I1201 08:44:13.396127 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Dec 01 08:44:13 crc kubenswrapper[4813]: I1201 08:44:13.424379 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Dec 01 08:44:13 crc kubenswrapper[4813]: I1201 08:44:13.489603 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Dec 01 08:44:13 crc kubenswrapper[4813]: I1201 08:44:13.505987 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Dec 01 08:44:13 crc kubenswrapper[4813]: I1201 08:44:13.520738 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Dec 01 08:44:13 crc kubenswrapper[4813]: I1201 08:44:13.576167 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Dec 01 08:44:13 crc kubenswrapper[4813]: I1201 08:44:13.588003 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Dec 01 08:44:13 crc kubenswrapper[4813]: I1201 08:44:13.669574 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Dec 01 08:44:13 crc kubenswrapper[4813]: I1201 08:44:13.690784 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Dec 01 08:44:13 crc kubenswrapper[4813]: I1201 08:44:13.768269 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Dec 01 08:44:13 crc kubenswrapper[4813]: I1201 08:44:13.788997 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Dec 01 08:44:13 crc kubenswrapper[4813]: I1201 08:44:13.933828 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Dec 01 08:44:13 crc kubenswrapper[4813]: I1201 08:44:13.935988 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Dec 01 08:44:14 crc kubenswrapper[4813]: I1201 08:44:14.024923 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Dec 01 08:44:14 crc kubenswrapper[4813]: I1201 08:44:14.031719 4813 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Dec 01 08:44:14 crc kubenswrapper[4813]: I1201 08:44:14.031996 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" containerID="cri-o://c8e6209fa56fb03f0f1bf4859a3d4994d5a1b490494d427c97727ff1cf8a6fdd" gracePeriod=5 Dec 01 08:44:14 crc kubenswrapper[4813]: I1201 08:44:14.076688 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Dec 01 08:44:14 crc kubenswrapper[4813]: I1201 08:44:14.140458 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Dec 01 08:44:14 crc kubenswrapper[4813]: I1201 08:44:14.152243 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Dec 01 08:44:14 crc kubenswrapper[4813]: I1201 08:44:14.220152 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 01 08:44:14 crc kubenswrapper[4813]: I1201 08:44:14.294810 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Dec 01 08:44:14 crc kubenswrapper[4813]: I1201 08:44:14.353791 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Dec 01 08:44:14 crc kubenswrapper[4813]: I1201 08:44:14.549068 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Dec 01 08:44:14 crc kubenswrapper[4813]: I1201 08:44:14.554607 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Dec 01 08:44:14 crc kubenswrapper[4813]: I1201 08:44:14.677849 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 01 08:44:14 crc kubenswrapper[4813]: I1201 08:44:14.719385 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Dec 01 08:44:14 crc kubenswrapper[4813]: I1201 08:44:14.846410 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Dec 01 08:44:14 crc kubenswrapper[4813]: I1201 08:44:14.900869 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Dec 01 08:44:14 crc kubenswrapper[4813]: I1201 08:44:14.929752 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Dec 01 08:44:14 crc kubenswrapper[4813]: I1201 08:44:14.931697 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Dec 01 08:44:14 crc kubenswrapper[4813]: I1201 08:44:14.934230 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Dec 01 08:44:15 crc kubenswrapper[4813]: I1201 08:44:15.081284 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Dec 01 08:44:15 crc kubenswrapper[4813]: I1201 08:44:15.085999 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Dec 01 08:44:15 crc kubenswrapper[4813]: I1201 08:44:15.278219 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Dec 01 08:44:15 crc kubenswrapper[4813]: I1201 08:44:15.317313 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Dec 01 08:44:15 crc kubenswrapper[4813]: I1201 08:44:15.365095 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Dec 01 08:44:15 crc kubenswrapper[4813]: I1201 08:44:15.405584 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Dec 01 08:44:15 crc kubenswrapper[4813]: I1201 08:44:15.498722 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Dec 01 08:44:15 crc kubenswrapper[4813]: I1201 08:44:15.532750 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Dec 01 08:44:15 crc kubenswrapper[4813]: I1201 08:44:15.574019 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Dec 01 08:44:15 crc kubenswrapper[4813]: I1201 08:44:15.897920 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Dec 01 08:44:16 crc kubenswrapper[4813]: I1201 08:44:16.033723 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Dec 01 08:44:16 crc kubenswrapper[4813]: I1201 08:44:16.437365 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Dec 01 08:44:16 crc kubenswrapper[4813]: I1201 08:44:16.526680 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Dec 01 08:44:16 crc kubenswrapper[4813]: I1201 08:44:16.826328 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Dec 01 08:44:16 crc kubenswrapper[4813]: I1201 08:44:16.990885 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Dec 01 08:44:16 crc kubenswrapper[4813]: I1201 08:44:16.994915 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Dec 01 08:44:17 crc kubenswrapper[4813]: I1201 08:44:17.324732 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Dec 01 08:44:17 crc kubenswrapper[4813]: I1201 08:44:17.559045 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Dec 01 08:44:17 crc kubenswrapper[4813]: I1201 08:44:17.823360 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Dec 01 08:44:19 crc kubenswrapper[4813]: I1201 08:44:19.581620 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Dec 01 08:44:19 crc kubenswrapper[4813]: I1201 08:44:19.582270 4813 generic.go:334] "Generic (PLEG): container finished" podID="f85e55b1a89d02b0cb034b1ea31ed45a" containerID="c8e6209fa56fb03f0f1bf4859a3d4994d5a1b490494d427c97727ff1cf8a6fdd" exitCode=137 Dec 01 08:44:19 crc kubenswrapper[4813]: I1201 08:44:19.632845 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Dec 01 08:44:19 crc kubenswrapper[4813]: I1201 08:44:19.633076 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 01 08:44:19 crc kubenswrapper[4813]: I1201 08:44:19.781831 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 01 08:44:19 crc kubenswrapper[4813]: I1201 08:44:19.781892 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 01 08:44:19 crc kubenswrapper[4813]: I1201 08:44:19.781938 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 01 08:44:19 crc kubenswrapper[4813]: I1201 08:44:19.782017 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 01 08:44:19 crc kubenswrapper[4813]: I1201 08:44:19.782086 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 01 08:44:19 crc kubenswrapper[4813]: I1201 08:44:19.782351 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock" (OuterVolumeSpecName: "var-lock") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 08:44:19 crc kubenswrapper[4813]: I1201 08:44:19.782353 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests" (OuterVolumeSpecName: "manifests") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "manifests". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 08:44:19 crc kubenswrapper[4813]: I1201 08:44:19.782429 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 08:44:19 crc kubenswrapper[4813]: I1201 08:44:19.782525 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log" (OuterVolumeSpecName: "var-log") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 08:44:19 crc kubenswrapper[4813]: I1201 08:44:19.824575 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir" (OuterVolumeSpecName: "pod-resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "pod-resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 08:44:19 crc kubenswrapper[4813]: I1201 08:44:19.883912 4813 reconciler_common.go:293] "Volume detached for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") on node \"crc\" DevicePath \"\"" Dec 01 08:44:19 crc kubenswrapper[4813]: I1201 08:44:19.883990 4813 reconciler_common.go:293] "Volume detached for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") on node \"crc\" DevicePath \"\"" Dec 01 08:44:19 crc kubenswrapper[4813]: I1201 08:44:19.884004 4813 reconciler_common.go:293] "Volume detached for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") on node \"crc\" DevicePath \"\"" Dec 01 08:44:19 crc kubenswrapper[4813]: I1201 08:44:19.884012 4813 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") on node \"crc\" DevicePath \"\"" Dec 01 08:44:19 crc kubenswrapper[4813]: I1201 08:44:19.884021 4813 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") on node \"crc\" DevicePath \"\"" Dec 01 08:44:20 crc kubenswrapper[4813]: I1201 08:44:20.406599 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" path="/var/lib/kubelet/pods/f85e55b1a89d02b0cb034b1ea31ed45a/volumes" Dec 01 08:44:20 crc kubenswrapper[4813]: I1201 08:44:20.593796 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Dec 01 08:44:20 crc kubenswrapper[4813]: I1201 08:44:20.594028 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 01 08:44:20 crc kubenswrapper[4813]: I1201 08:44:20.594047 4813 scope.go:117] "RemoveContainer" containerID="c8e6209fa56fb03f0f1bf4859a3d4994d5a1b490494d427c97727ff1cf8a6fdd" Dec 01 08:44:32 crc kubenswrapper[4813]: I1201 08:44:32.173056 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Dec 01 08:44:36 crc kubenswrapper[4813]: I1201 08:44:36.997318 4813 generic.go:334] "Generic (PLEG): container finished" podID="c342090f-f331-484d-b287-929d8303f5d7" containerID="675727fde89d8e51f7ed3ff2b83d39e2917fa51a8b9eb117e74b96316e396b4f" exitCode=0 Dec 01 08:44:36 crc kubenswrapper[4813]: I1201 08:44:36.997457 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-qsskc" event={"ID":"c342090f-f331-484d-b287-929d8303f5d7","Type":"ContainerDied","Data":"675727fde89d8e51f7ed3ff2b83d39e2917fa51a8b9eb117e74b96316e396b4f"} Dec 01 08:44:36 crc kubenswrapper[4813]: I1201 08:44:36.998376 4813 scope.go:117] "RemoveContainer" containerID="675727fde89d8e51f7ed3ff2b83d39e2917fa51a8b9eb117e74b96316e396b4f" Dec 01 08:44:38 crc kubenswrapper[4813]: I1201 08:44:38.007597 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-qsskc" event={"ID":"c342090f-f331-484d-b287-929d8303f5d7","Type":"ContainerStarted","Data":"9c9df537248bcf5a73e1bcdb2c8e30af1d91f39aac66106534a6c2aa1f5fab7a"} Dec 01 08:44:38 crc kubenswrapper[4813]: I1201 08:44:38.009594 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-qsskc" Dec 01 08:44:38 crc kubenswrapper[4813]: I1201 08:44:38.013560 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-qsskc" Dec 01 08:44:48 crc kubenswrapper[4813]: I1201 08:44:48.170858 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-l4zh5"] Dec 01 08:44:48 crc kubenswrapper[4813]: I1201 08:44:48.171908 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-l4zh5" podUID="705b2456-4ba7-4775-84ca-4dcea64b6755" containerName="controller-manager" containerID="cri-o://2e0e06d86b703d0484153ad01443faf32d8bb1a0bbbf5e6ecbecc5c8847cce3f" gracePeriod=30 Dec 01 08:44:48 crc kubenswrapper[4813]: I1201 08:44:48.269936 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-2mb7d"] Dec 01 08:44:48 crc kubenswrapper[4813]: I1201 08:44:48.270198 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-2mb7d" podUID="0f0340e8-faa5-447b-a5cd-a878e982d98f" containerName="route-controller-manager" containerID="cri-o://a886c4c7bb5d2c717fea11a48e826ddbcdd7e2724cea3114371eee53b1ed91d1" gracePeriod=30 Dec 01 08:44:48 crc kubenswrapper[4813]: I1201 08:44:48.578041 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-l4zh5" Dec 01 08:44:48 crc kubenswrapper[4813]: I1201 08:44:48.625827 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-2mb7d" Dec 01 08:44:48 crc kubenswrapper[4813]: I1201 08:44:48.726806 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjxkr\" (UniqueName: \"kubernetes.io/projected/705b2456-4ba7-4775-84ca-4dcea64b6755-kube-api-access-pjxkr\") pod \"705b2456-4ba7-4775-84ca-4dcea64b6755\" (UID: \"705b2456-4ba7-4775-84ca-4dcea64b6755\") " Dec 01 08:44:48 crc kubenswrapper[4813]: I1201 08:44:48.726854 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/705b2456-4ba7-4775-84ca-4dcea64b6755-proxy-ca-bundles\") pod \"705b2456-4ba7-4775-84ca-4dcea64b6755\" (UID: \"705b2456-4ba7-4775-84ca-4dcea64b6755\") " Dec 01 08:44:48 crc kubenswrapper[4813]: I1201 08:44:48.726946 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/705b2456-4ba7-4775-84ca-4dcea64b6755-config\") pod \"705b2456-4ba7-4775-84ca-4dcea64b6755\" (UID: \"705b2456-4ba7-4775-84ca-4dcea64b6755\") " Dec 01 08:44:48 crc kubenswrapper[4813]: I1201 08:44:48.726980 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/705b2456-4ba7-4775-84ca-4dcea64b6755-client-ca\") pod \"705b2456-4ba7-4775-84ca-4dcea64b6755\" (UID: \"705b2456-4ba7-4775-84ca-4dcea64b6755\") " Dec 01 08:44:48 crc kubenswrapper[4813]: I1201 08:44:48.727163 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0f0340e8-faa5-447b-a5cd-a878e982d98f-serving-cert\") pod \"0f0340e8-faa5-447b-a5cd-a878e982d98f\" (UID: \"0f0340e8-faa5-447b-a5cd-a878e982d98f\") " Dec 01 08:44:48 crc kubenswrapper[4813]: I1201 08:44:48.727219 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/705b2456-4ba7-4775-84ca-4dcea64b6755-serving-cert\") pod \"705b2456-4ba7-4775-84ca-4dcea64b6755\" (UID: \"705b2456-4ba7-4775-84ca-4dcea64b6755\") " Dec 01 08:44:48 crc kubenswrapper[4813]: I1201 08:44:48.727257 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0f0340e8-faa5-447b-a5cd-a878e982d98f-config\") pod \"0f0340e8-faa5-447b-a5cd-a878e982d98f\" (UID: \"0f0340e8-faa5-447b-a5cd-a878e982d98f\") " Dec 01 08:44:48 crc kubenswrapper[4813]: I1201 08:44:48.727821 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/705b2456-4ba7-4775-84ca-4dcea64b6755-client-ca" (OuterVolumeSpecName: "client-ca") pod "705b2456-4ba7-4775-84ca-4dcea64b6755" (UID: "705b2456-4ba7-4775-84ca-4dcea64b6755"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:44:48 crc kubenswrapper[4813]: I1201 08:44:48.727815 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/705b2456-4ba7-4775-84ca-4dcea64b6755-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "705b2456-4ba7-4775-84ca-4dcea64b6755" (UID: "705b2456-4ba7-4775-84ca-4dcea64b6755"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:44:48 crc kubenswrapper[4813]: I1201 08:44:48.728063 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/705b2456-4ba7-4775-84ca-4dcea64b6755-config" (OuterVolumeSpecName: "config") pod "705b2456-4ba7-4775-84ca-4dcea64b6755" (UID: "705b2456-4ba7-4775-84ca-4dcea64b6755"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:44:48 crc kubenswrapper[4813]: I1201 08:44:48.728366 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0f0340e8-faa5-447b-a5cd-a878e982d98f-config" (OuterVolumeSpecName: "config") pod "0f0340e8-faa5-447b-a5cd-a878e982d98f" (UID: "0f0340e8-faa5-447b-a5cd-a878e982d98f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:44:48 crc kubenswrapper[4813]: I1201 08:44:48.733829 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/705b2456-4ba7-4775-84ca-4dcea64b6755-kube-api-access-pjxkr" (OuterVolumeSpecName: "kube-api-access-pjxkr") pod "705b2456-4ba7-4775-84ca-4dcea64b6755" (UID: "705b2456-4ba7-4775-84ca-4dcea64b6755"). InnerVolumeSpecName "kube-api-access-pjxkr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:44:48 crc kubenswrapper[4813]: I1201 08:44:48.733865 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0f0340e8-faa5-447b-a5cd-a878e982d98f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0f0340e8-faa5-447b-a5cd-a878e982d98f" (UID: "0f0340e8-faa5-447b-a5cd-a878e982d98f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:44:48 crc kubenswrapper[4813]: I1201 08:44:48.734525 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/705b2456-4ba7-4775-84ca-4dcea64b6755-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "705b2456-4ba7-4775-84ca-4dcea64b6755" (UID: "705b2456-4ba7-4775-84ca-4dcea64b6755"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:44:48 crc kubenswrapper[4813]: I1201 08:44:48.828476 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/0f0340e8-faa5-447b-a5cd-a878e982d98f-client-ca\") pod \"0f0340e8-faa5-447b-a5cd-a878e982d98f\" (UID: \"0f0340e8-faa5-447b-a5cd-a878e982d98f\") " Dec 01 08:44:48 crc kubenswrapper[4813]: I1201 08:44:48.828731 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-72d4v\" (UniqueName: \"kubernetes.io/projected/0f0340e8-faa5-447b-a5cd-a878e982d98f-kube-api-access-72d4v\") pod \"0f0340e8-faa5-447b-a5cd-a878e982d98f\" (UID: \"0f0340e8-faa5-447b-a5cd-a878e982d98f\") " Dec 01 08:44:48 crc kubenswrapper[4813]: I1201 08:44:48.829125 4813 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/705b2456-4ba7-4775-84ca-4dcea64b6755-config\") on node \"crc\" DevicePath \"\"" Dec 01 08:44:48 crc kubenswrapper[4813]: I1201 08:44:48.829150 4813 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/705b2456-4ba7-4775-84ca-4dcea64b6755-client-ca\") on node \"crc\" DevicePath \"\"" Dec 01 08:44:48 crc kubenswrapper[4813]: I1201 08:44:48.829163 4813 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0f0340e8-faa5-447b-a5cd-a878e982d98f-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 01 08:44:48 crc kubenswrapper[4813]: I1201 08:44:48.829176 4813 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/705b2456-4ba7-4775-84ca-4dcea64b6755-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 01 08:44:48 crc kubenswrapper[4813]: I1201 08:44:48.829188 4813 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0f0340e8-faa5-447b-a5cd-a878e982d98f-config\") on node \"crc\" DevicePath \"\"" Dec 01 08:44:48 crc kubenswrapper[4813]: I1201 08:44:48.829200 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjxkr\" (UniqueName: \"kubernetes.io/projected/705b2456-4ba7-4775-84ca-4dcea64b6755-kube-api-access-pjxkr\") on node \"crc\" DevicePath \"\"" Dec 01 08:44:48 crc kubenswrapper[4813]: I1201 08:44:48.829213 4813 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/705b2456-4ba7-4775-84ca-4dcea64b6755-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 01 08:44:48 crc kubenswrapper[4813]: I1201 08:44:48.829317 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0f0340e8-faa5-447b-a5cd-a878e982d98f-client-ca" (OuterVolumeSpecName: "client-ca") pod "0f0340e8-faa5-447b-a5cd-a878e982d98f" (UID: "0f0340e8-faa5-447b-a5cd-a878e982d98f"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:44:48 crc kubenswrapper[4813]: I1201 08:44:48.832328 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0f0340e8-faa5-447b-a5cd-a878e982d98f-kube-api-access-72d4v" (OuterVolumeSpecName: "kube-api-access-72d4v") pod "0f0340e8-faa5-447b-a5cd-a878e982d98f" (UID: "0f0340e8-faa5-447b-a5cd-a878e982d98f"). InnerVolumeSpecName "kube-api-access-72d4v". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:44:48 crc kubenswrapper[4813]: I1201 08:44:48.930505 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-72d4v\" (UniqueName: \"kubernetes.io/projected/0f0340e8-faa5-447b-a5cd-a878e982d98f-kube-api-access-72d4v\") on node \"crc\" DevicePath \"\"" Dec 01 08:44:48 crc kubenswrapper[4813]: I1201 08:44:48.930565 4813 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/0f0340e8-faa5-447b-a5cd-a878e982d98f-client-ca\") on node \"crc\" DevicePath \"\"" Dec 01 08:44:49 crc kubenswrapper[4813]: I1201 08:44:49.083537 4813 generic.go:334] "Generic (PLEG): container finished" podID="0f0340e8-faa5-447b-a5cd-a878e982d98f" containerID="a886c4c7bb5d2c717fea11a48e826ddbcdd7e2724cea3114371eee53b1ed91d1" exitCode=0 Dec 01 08:44:49 crc kubenswrapper[4813]: I1201 08:44:49.083661 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-2mb7d" Dec 01 08:44:49 crc kubenswrapper[4813]: I1201 08:44:49.084151 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-2mb7d" event={"ID":"0f0340e8-faa5-447b-a5cd-a878e982d98f","Type":"ContainerDied","Data":"a886c4c7bb5d2c717fea11a48e826ddbcdd7e2724cea3114371eee53b1ed91d1"} Dec 01 08:44:49 crc kubenswrapper[4813]: I1201 08:44:49.084410 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-2mb7d" event={"ID":"0f0340e8-faa5-447b-a5cd-a878e982d98f","Type":"ContainerDied","Data":"81ace5696b7b1e97479c500c0dae3e9c2f5ad6341c22e14161695ed61e7ab950"} Dec 01 08:44:49 crc kubenswrapper[4813]: I1201 08:44:49.084510 4813 scope.go:117] "RemoveContainer" containerID="a886c4c7bb5d2c717fea11a48e826ddbcdd7e2724cea3114371eee53b1ed91d1" Dec 01 08:44:49 crc kubenswrapper[4813]: I1201 08:44:49.087279 4813 generic.go:334] "Generic (PLEG): container finished" podID="705b2456-4ba7-4775-84ca-4dcea64b6755" containerID="2e0e06d86b703d0484153ad01443faf32d8bb1a0bbbf5e6ecbecc5c8847cce3f" exitCode=0 Dec 01 08:44:49 crc kubenswrapper[4813]: I1201 08:44:49.087340 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-l4zh5" event={"ID":"705b2456-4ba7-4775-84ca-4dcea64b6755","Type":"ContainerDied","Data":"2e0e06d86b703d0484153ad01443faf32d8bb1a0bbbf5e6ecbecc5c8847cce3f"} Dec 01 08:44:49 crc kubenswrapper[4813]: I1201 08:44:49.087384 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-l4zh5" event={"ID":"705b2456-4ba7-4775-84ca-4dcea64b6755","Type":"ContainerDied","Data":"7301dd69d74eda5bd74707420800383217ecf5e7c3028a1e77774f282ad69285"} Dec 01 08:44:49 crc kubenswrapper[4813]: I1201 08:44:49.087464 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-l4zh5" Dec 01 08:44:49 crc kubenswrapper[4813]: I1201 08:44:49.116467 4813 scope.go:117] "RemoveContainer" containerID="a886c4c7bb5d2c717fea11a48e826ddbcdd7e2724cea3114371eee53b1ed91d1" Dec 01 08:44:49 crc kubenswrapper[4813]: E1201 08:44:49.117461 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a886c4c7bb5d2c717fea11a48e826ddbcdd7e2724cea3114371eee53b1ed91d1\": container with ID starting with a886c4c7bb5d2c717fea11a48e826ddbcdd7e2724cea3114371eee53b1ed91d1 not found: ID does not exist" containerID="a886c4c7bb5d2c717fea11a48e826ddbcdd7e2724cea3114371eee53b1ed91d1" Dec 01 08:44:49 crc kubenswrapper[4813]: I1201 08:44:49.117527 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a886c4c7bb5d2c717fea11a48e826ddbcdd7e2724cea3114371eee53b1ed91d1"} err="failed to get container status \"a886c4c7bb5d2c717fea11a48e826ddbcdd7e2724cea3114371eee53b1ed91d1\": rpc error: code = NotFound desc = could not find container \"a886c4c7bb5d2c717fea11a48e826ddbcdd7e2724cea3114371eee53b1ed91d1\": container with ID starting with a886c4c7bb5d2c717fea11a48e826ddbcdd7e2724cea3114371eee53b1ed91d1 not found: ID does not exist" Dec 01 08:44:49 crc kubenswrapper[4813]: I1201 08:44:49.117569 4813 scope.go:117] "RemoveContainer" containerID="2e0e06d86b703d0484153ad01443faf32d8bb1a0bbbf5e6ecbecc5c8847cce3f" Dec 01 08:44:49 crc kubenswrapper[4813]: I1201 08:44:49.180585 4813 scope.go:117] "RemoveContainer" containerID="2e0e06d86b703d0484153ad01443faf32d8bb1a0bbbf5e6ecbecc5c8847cce3f" Dec 01 08:44:49 crc kubenswrapper[4813]: I1201 08:44:49.180720 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-2mb7d"] Dec 01 08:44:49 crc kubenswrapper[4813]: E1201 08:44:49.181144 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2e0e06d86b703d0484153ad01443faf32d8bb1a0bbbf5e6ecbecc5c8847cce3f\": container with ID starting with 2e0e06d86b703d0484153ad01443faf32d8bb1a0bbbf5e6ecbecc5c8847cce3f not found: ID does not exist" containerID="2e0e06d86b703d0484153ad01443faf32d8bb1a0bbbf5e6ecbecc5c8847cce3f" Dec 01 08:44:49 crc kubenswrapper[4813]: I1201 08:44:49.181183 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2e0e06d86b703d0484153ad01443faf32d8bb1a0bbbf5e6ecbecc5c8847cce3f"} err="failed to get container status \"2e0e06d86b703d0484153ad01443faf32d8bb1a0bbbf5e6ecbecc5c8847cce3f\": rpc error: code = NotFound desc = could not find container \"2e0e06d86b703d0484153ad01443faf32d8bb1a0bbbf5e6ecbecc5c8847cce3f\": container with ID starting with 2e0e06d86b703d0484153ad01443faf32d8bb1a0bbbf5e6ecbecc5c8847cce3f not found: ID does not exist" Dec 01 08:44:49 crc kubenswrapper[4813]: I1201 08:44:49.184373 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-2mb7d"] Dec 01 08:44:49 crc kubenswrapper[4813]: I1201 08:44:49.190724 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-l4zh5"] Dec 01 08:44:49 crc kubenswrapper[4813]: I1201 08:44:49.194730 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-l4zh5"] Dec 01 08:44:49 crc kubenswrapper[4813]: I1201 08:44:49.337422 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-f86fd7585-cfv88"] Dec 01 08:44:49 crc kubenswrapper[4813]: E1201 08:44:49.338065 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0f0340e8-faa5-447b-a5cd-a878e982d98f" containerName="route-controller-manager" Dec 01 08:44:49 crc kubenswrapper[4813]: I1201 08:44:49.338107 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="0f0340e8-faa5-447b-a5cd-a878e982d98f" containerName="route-controller-manager" Dec 01 08:44:49 crc kubenswrapper[4813]: E1201 08:44:49.338127 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="269d86e5-dc67-48b3-a955-201b986f6c97" containerName="installer" Dec 01 08:44:49 crc kubenswrapper[4813]: I1201 08:44:49.338135 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="269d86e5-dc67-48b3-a955-201b986f6c97" containerName="installer" Dec 01 08:44:49 crc kubenswrapper[4813]: E1201 08:44:49.338146 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Dec 01 08:44:49 crc kubenswrapper[4813]: I1201 08:44:49.338154 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Dec 01 08:44:49 crc kubenswrapper[4813]: E1201 08:44:49.338168 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="705b2456-4ba7-4775-84ca-4dcea64b6755" containerName="controller-manager" Dec 01 08:44:49 crc kubenswrapper[4813]: I1201 08:44:49.338177 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="705b2456-4ba7-4775-84ca-4dcea64b6755" containerName="controller-manager" Dec 01 08:44:49 crc kubenswrapper[4813]: I1201 08:44:49.338336 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="269d86e5-dc67-48b3-a955-201b986f6c97" containerName="installer" Dec 01 08:44:49 crc kubenswrapper[4813]: I1201 08:44:49.338361 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="0f0340e8-faa5-447b-a5cd-a878e982d98f" containerName="route-controller-manager" Dec 01 08:44:49 crc kubenswrapper[4813]: I1201 08:44:49.338374 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="705b2456-4ba7-4775-84ca-4dcea64b6755" containerName="controller-manager" Dec 01 08:44:49 crc kubenswrapper[4813]: I1201 08:44:49.338385 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Dec 01 08:44:49 crc kubenswrapper[4813]: I1201 08:44:49.338937 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-f86fd7585-cfv88" Dec 01 08:44:49 crc kubenswrapper[4813]: I1201 08:44:49.345935 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-84f5659d8d-6mdwm"] Dec 01 08:44:49 crc kubenswrapper[4813]: I1201 08:44:49.346767 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-84f5659d8d-6mdwm" Dec 01 08:44:49 crc kubenswrapper[4813]: I1201 08:44:49.350306 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 01 08:44:49 crc kubenswrapper[4813]: I1201 08:44:49.350376 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 01 08:44:49 crc kubenswrapper[4813]: I1201 08:44:49.350648 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Dec 01 08:44:49 crc kubenswrapper[4813]: I1201 08:44:49.351131 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Dec 01 08:44:49 crc kubenswrapper[4813]: I1201 08:44:49.351172 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 01 08:44:49 crc kubenswrapper[4813]: I1201 08:44:49.351326 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Dec 01 08:44:49 crc kubenswrapper[4813]: I1201 08:44:49.351432 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Dec 01 08:44:49 crc kubenswrapper[4813]: I1201 08:44:49.351482 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Dec 01 08:44:49 crc kubenswrapper[4813]: I1201 08:44:49.351843 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Dec 01 08:44:49 crc kubenswrapper[4813]: I1201 08:44:49.352073 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 01 08:44:49 crc kubenswrapper[4813]: I1201 08:44:49.354167 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 01 08:44:49 crc kubenswrapper[4813]: I1201 08:44:49.354954 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-f86fd7585-cfv88"] Dec 01 08:44:49 crc kubenswrapper[4813]: I1201 08:44:49.355227 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 01 08:44:49 crc kubenswrapper[4813]: I1201 08:44:49.359796 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 01 08:44:49 crc kubenswrapper[4813]: I1201 08:44:49.360778 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-84f5659d8d-6mdwm"] Dec 01 08:44:49 crc kubenswrapper[4813]: I1201 08:44:49.541790 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/93834113-aa88-44f3-bf6b-4a07e78cfef4-client-ca\") pod \"route-controller-manager-84f5659d8d-6mdwm\" (UID: \"93834113-aa88-44f3-bf6b-4a07e78cfef4\") " pod="openshift-route-controller-manager/route-controller-manager-84f5659d8d-6mdwm" Dec 01 08:44:49 crc kubenswrapper[4813]: I1201 08:44:49.542361 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-69xlg\" (UniqueName: \"kubernetes.io/projected/e0324384-89bb-42c0-8597-55fee0b4bb17-kube-api-access-69xlg\") pod \"controller-manager-f86fd7585-cfv88\" (UID: \"e0324384-89bb-42c0-8597-55fee0b4bb17\") " pod="openshift-controller-manager/controller-manager-f86fd7585-cfv88" Dec 01 08:44:49 crc kubenswrapper[4813]: I1201 08:44:49.542632 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e0324384-89bb-42c0-8597-55fee0b4bb17-serving-cert\") pod \"controller-manager-f86fd7585-cfv88\" (UID: \"e0324384-89bb-42c0-8597-55fee0b4bb17\") " pod="openshift-controller-manager/controller-manager-f86fd7585-cfv88" Dec 01 08:44:49 crc kubenswrapper[4813]: I1201 08:44:49.542803 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e0324384-89bb-42c0-8597-55fee0b4bb17-client-ca\") pod \"controller-manager-f86fd7585-cfv88\" (UID: \"e0324384-89bb-42c0-8597-55fee0b4bb17\") " pod="openshift-controller-manager/controller-manager-f86fd7585-cfv88" Dec 01 08:44:49 crc kubenswrapper[4813]: I1201 08:44:49.542948 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e0324384-89bb-42c0-8597-55fee0b4bb17-config\") pod \"controller-manager-f86fd7585-cfv88\" (UID: \"e0324384-89bb-42c0-8597-55fee0b4bb17\") " pod="openshift-controller-manager/controller-manager-f86fd7585-cfv88" Dec 01 08:44:49 crc kubenswrapper[4813]: I1201 08:44:49.543148 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/93834113-aa88-44f3-bf6b-4a07e78cfef4-serving-cert\") pod \"route-controller-manager-84f5659d8d-6mdwm\" (UID: \"93834113-aa88-44f3-bf6b-4a07e78cfef4\") " pod="openshift-route-controller-manager/route-controller-manager-84f5659d8d-6mdwm" Dec 01 08:44:49 crc kubenswrapper[4813]: I1201 08:44:49.543348 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p5hmz\" (UniqueName: \"kubernetes.io/projected/93834113-aa88-44f3-bf6b-4a07e78cfef4-kube-api-access-p5hmz\") pod \"route-controller-manager-84f5659d8d-6mdwm\" (UID: \"93834113-aa88-44f3-bf6b-4a07e78cfef4\") " pod="openshift-route-controller-manager/route-controller-manager-84f5659d8d-6mdwm" Dec 01 08:44:49 crc kubenswrapper[4813]: I1201 08:44:49.543511 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e0324384-89bb-42c0-8597-55fee0b4bb17-proxy-ca-bundles\") pod \"controller-manager-f86fd7585-cfv88\" (UID: \"e0324384-89bb-42c0-8597-55fee0b4bb17\") " pod="openshift-controller-manager/controller-manager-f86fd7585-cfv88" Dec 01 08:44:49 crc kubenswrapper[4813]: I1201 08:44:49.543714 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/93834113-aa88-44f3-bf6b-4a07e78cfef4-config\") pod \"route-controller-manager-84f5659d8d-6mdwm\" (UID: \"93834113-aa88-44f3-bf6b-4a07e78cfef4\") " pod="openshift-route-controller-manager/route-controller-manager-84f5659d8d-6mdwm" Dec 01 08:44:49 crc kubenswrapper[4813]: I1201 08:44:49.645241 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e0324384-89bb-42c0-8597-55fee0b4bb17-serving-cert\") pod \"controller-manager-f86fd7585-cfv88\" (UID: \"e0324384-89bb-42c0-8597-55fee0b4bb17\") " pod="openshift-controller-manager/controller-manager-f86fd7585-cfv88" Dec 01 08:44:49 crc kubenswrapper[4813]: I1201 08:44:49.645313 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e0324384-89bb-42c0-8597-55fee0b4bb17-client-ca\") pod \"controller-manager-f86fd7585-cfv88\" (UID: \"e0324384-89bb-42c0-8597-55fee0b4bb17\") " pod="openshift-controller-manager/controller-manager-f86fd7585-cfv88" Dec 01 08:44:49 crc kubenswrapper[4813]: I1201 08:44:49.645337 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e0324384-89bb-42c0-8597-55fee0b4bb17-config\") pod \"controller-manager-f86fd7585-cfv88\" (UID: \"e0324384-89bb-42c0-8597-55fee0b4bb17\") " pod="openshift-controller-manager/controller-manager-f86fd7585-cfv88" Dec 01 08:44:49 crc kubenswrapper[4813]: I1201 08:44:49.645356 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/93834113-aa88-44f3-bf6b-4a07e78cfef4-serving-cert\") pod \"route-controller-manager-84f5659d8d-6mdwm\" (UID: \"93834113-aa88-44f3-bf6b-4a07e78cfef4\") " pod="openshift-route-controller-manager/route-controller-manager-84f5659d8d-6mdwm" Dec 01 08:44:49 crc kubenswrapper[4813]: I1201 08:44:49.645374 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p5hmz\" (UniqueName: \"kubernetes.io/projected/93834113-aa88-44f3-bf6b-4a07e78cfef4-kube-api-access-p5hmz\") pod \"route-controller-manager-84f5659d8d-6mdwm\" (UID: \"93834113-aa88-44f3-bf6b-4a07e78cfef4\") " pod="openshift-route-controller-manager/route-controller-manager-84f5659d8d-6mdwm" Dec 01 08:44:49 crc kubenswrapper[4813]: I1201 08:44:49.645396 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e0324384-89bb-42c0-8597-55fee0b4bb17-proxy-ca-bundles\") pod \"controller-manager-f86fd7585-cfv88\" (UID: \"e0324384-89bb-42c0-8597-55fee0b4bb17\") " pod="openshift-controller-manager/controller-manager-f86fd7585-cfv88" Dec 01 08:44:49 crc kubenswrapper[4813]: I1201 08:44:49.645446 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/93834113-aa88-44f3-bf6b-4a07e78cfef4-config\") pod \"route-controller-manager-84f5659d8d-6mdwm\" (UID: \"93834113-aa88-44f3-bf6b-4a07e78cfef4\") " pod="openshift-route-controller-manager/route-controller-manager-84f5659d8d-6mdwm" Dec 01 08:44:49 crc kubenswrapper[4813]: I1201 08:44:49.645467 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/93834113-aa88-44f3-bf6b-4a07e78cfef4-client-ca\") pod \"route-controller-manager-84f5659d8d-6mdwm\" (UID: \"93834113-aa88-44f3-bf6b-4a07e78cfef4\") " pod="openshift-route-controller-manager/route-controller-manager-84f5659d8d-6mdwm" Dec 01 08:44:49 crc kubenswrapper[4813]: I1201 08:44:49.646201 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-69xlg\" (UniqueName: \"kubernetes.io/projected/e0324384-89bb-42c0-8597-55fee0b4bb17-kube-api-access-69xlg\") pod \"controller-manager-f86fd7585-cfv88\" (UID: \"e0324384-89bb-42c0-8597-55fee0b4bb17\") " pod="openshift-controller-manager/controller-manager-f86fd7585-cfv88" Dec 01 08:44:49 crc kubenswrapper[4813]: I1201 08:44:49.647788 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/93834113-aa88-44f3-bf6b-4a07e78cfef4-client-ca\") pod \"route-controller-manager-84f5659d8d-6mdwm\" (UID: \"93834113-aa88-44f3-bf6b-4a07e78cfef4\") " pod="openshift-route-controller-manager/route-controller-manager-84f5659d8d-6mdwm" Dec 01 08:44:49 crc kubenswrapper[4813]: I1201 08:44:49.648442 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/93834113-aa88-44f3-bf6b-4a07e78cfef4-config\") pod \"route-controller-manager-84f5659d8d-6mdwm\" (UID: \"93834113-aa88-44f3-bf6b-4a07e78cfef4\") " pod="openshift-route-controller-manager/route-controller-manager-84f5659d8d-6mdwm" Dec 01 08:44:49 crc kubenswrapper[4813]: I1201 08:44:49.648571 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e0324384-89bb-42c0-8597-55fee0b4bb17-client-ca\") pod \"controller-manager-f86fd7585-cfv88\" (UID: \"e0324384-89bb-42c0-8597-55fee0b4bb17\") " pod="openshift-controller-manager/controller-manager-f86fd7585-cfv88" Dec 01 08:44:49 crc kubenswrapper[4813]: I1201 08:44:49.652300 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/93834113-aa88-44f3-bf6b-4a07e78cfef4-serving-cert\") pod \"route-controller-manager-84f5659d8d-6mdwm\" (UID: \"93834113-aa88-44f3-bf6b-4a07e78cfef4\") " pod="openshift-route-controller-manager/route-controller-manager-84f5659d8d-6mdwm" Dec 01 08:44:49 crc kubenswrapper[4813]: I1201 08:44:49.659671 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e0324384-89bb-42c0-8597-55fee0b4bb17-serving-cert\") pod \"controller-manager-f86fd7585-cfv88\" (UID: \"e0324384-89bb-42c0-8597-55fee0b4bb17\") " pod="openshift-controller-manager/controller-manager-f86fd7585-cfv88" Dec 01 08:44:49 crc kubenswrapper[4813]: I1201 08:44:49.660101 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e0324384-89bb-42c0-8597-55fee0b4bb17-config\") pod \"controller-manager-f86fd7585-cfv88\" (UID: \"e0324384-89bb-42c0-8597-55fee0b4bb17\") " pod="openshift-controller-manager/controller-manager-f86fd7585-cfv88" Dec 01 08:44:49 crc kubenswrapper[4813]: I1201 08:44:49.660557 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e0324384-89bb-42c0-8597-55fee0b4bb17-proxy-ca-bundles\") pod \"controller-manager-f86fd7585-cfv88\" (UID: \"e0324384-89bb-42c0-8597-55fee0b4bb17\") " pod="openshift-controller-manager/controller-manager-f86fd7585-cfv88" Dec 01 08:44:49 crc kubenswrapper[4813]: I1201 08:44:49.667895 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-69xlg\" (UniqueName: \"kubernetes.io/projected/e0324384-89bb-42c0-8597-55fee0b4bb17-kube-api-access-69xlg\") pod \"controller-manager-f86fd7585-cfv88\" (UID: \"e0324384-89bb-42c0-8597-55fee0b4bb17\") " pod="openshift-controller-manager/controller-manager-f86fd7585-cfv88" Dec 01 08:44:49 crc kubenswrapper[4813]: I1201 08:44:49.671451 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p5hmz\" (UniqueName: \"kubernetes.io/projected/93834113-aa88-44f3-bf6b-4a07e78cfef4-kube-api-access-p5hmz\") pod \"route-controller-manager-84f5659d8d-6mdwm\" (UID: \"93834113-aa88-44f3-bf6b-4a07e78cfef4\") " pod="openshift-route-controller-manager/route-controller-manager-84f5659d8d-6mdwm" Dec 01 08:44:49 crc kubenswrapper[4813]: I1201 08:44:49.750329 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-84f5659d8d-6mdwm" Dec 01 08:44:49 crc kubenswrapper[4813]: I1201 08:44:49.963273 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-f86fd7585-cfv88" Dec 01 08:44:50 crc kubenswrapper[4813]: I1201 08:44:50.155407 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-84f5659d8d-6mdwm"] Dec 01 08:44:50 crc kubenswrapper[4813]: I1201 08:44:50.163528 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-f86fd7585-cfv88"] Dec 01 08:44:50 crc kubenswrapper[4813]: W1201 08:44:50.169394 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod93834113_aa88_44f3_bf6b_4a07e78cfef4.slice/crio-279f06cffb3c3fe9475f7b9b41be3562c86149acad1286811f2049a2c0d124b1 WatchSource:0}: Error finding container 279f06cffb3c3fe9475f7b9b41be3562c86149acad1286811f2049a2c0d124b1: Status 404 returned error can't find the container with id 279f06cffb3c3fe9475f7b9b41be3562c86149acad1286811f2049a2c0d124b1 Dec 01 08:44:50 crc kubenswrapper[4813]: I1201 08:44:50.405613 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0f0340e8-faa5-447b-a5cd-a878e982d98f" path="/var/lib/kubelet/pods/0f0340e8-faa5-447b-a5cd-a878e982d98f/volumes" Dec 01 08:44:50 crc kubenswrapper[4813]: I1201 08:44:50.406950 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="705b2456-4ba7-4775-84ca-4dcea64b6755" path="/var/lib/kubelet/pods/705b2456-4ba7-4775-84ca-4dcea64b6755/volumes" Dec 01 08:44:51 crc kubenswrapper[4813]: I1201 08:44:51.110439 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-84f5659d8d-6mdwm" event={"ID":"93834113-aa88-44f3-bf6b-4a07e78cfef4","Type":"ContainerStarted","Data":"869bb1c5e9c619ddce571cbb2f8cd1bac46873d00cbe3e34fc6d7035c1d6a127"} Dec 01 08:44:51 crc kubenswrapper[4813]: I1201 08:44:51.110503 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-84f5659d8d-6mdwm" event={"ID":"93834113-aa88-44f3-bf6b-4a07e78cfef4","Type":"ContainerStarted","Data":"279f06cffb3c3fe9475f7b9b41be3562c86149acad1286811f2049a2c0d124b1"} Dec 01 08:44:51 crc kubenswrapper[4813]: I1201 08:44:51.110901 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-84f5659d8d-6mdwm" Dec 01 08:44:51 crc kubenswrapper[4813]: I1201 08:44:51.113516 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-f86fd7585-cfv88" event={"ID":"e0324384-89bb-42c0-8597-55fee0b4bb17","Type":"ContainerStarted","Data":"5ee93d1d8abaaf8342f8e83f8c2bff101787fbf77c30d25963d6aba49c033174"} Dec 01 08:44:51 crc kubenswrapper[4813]: I1201 08:44:51.113566 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-f86fd7585-cfv88" event={"ID":"e0324384-89bb-42c0-8597-55fee0b4bb17","Type":"ContainerStarted","Data":"a1d32a47c79704b9835946ce477b8adab192b21cf6edd19eed2e214e4e8812db"} Dec 01 08:44:51 crc kubenswrapper[4813]: I1201 08:44:51.113816 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-f86fd7585-cfv88" Dec 01 08:44:51 crc kubenswrapper[4813]: I1201 08:44:51.116906 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-84f5659d8d-6mdwm" Dec 01 08:44:51 crc kubenswrapper[4813]: I1201 08:44:51.123689 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-f86fd7585-cfv88" Dec 01 08:44:51 crc kubenswrapper[4813]: I1201 08:44:51.160457 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-84f5659d8d-6mdwm" podStartSLOduration=2.160376894 podStartE2EDuration="2.160376894s" podCreationTimestamp="2025-12-01 08:44:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:44:51.153311907 +0000 UTC m=+275.776233503" watchObservedRunningTime="2025-12-01 08:44:51.160376894 +0000 UTC m=+275.783298530" Dec 01 08:44:51 crc kubenswrapper[4813]: I1201 08:44:51.180859 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-f86fd7585-cfv88" podStartSLOduration=2.180833842 podStartE2EDuration="2.180833842s" podCreationTimestamp="2025-12-01 08:44:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:44:51.180787581 +0000 UTC m=+275.803709187" watchObservedRunningTime="2025-12-01 08:44:51.180833842 +0000 UTC m=+275.803755418" Dec 01 08:45:00 crc kubenswrapper[4813]: I1201 08:45:00.207959 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29409645-lwkbv"] Dec 01 08:45:00 crc kubenswrapper[4813]: I1201 08:45:00.209381 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29409645-lwkbv" Dec 01 08:45:00 crc kubenswrapper[4813]: I1201 08:45:00.212211 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 01 08:45:00 crc kubenswrapper[4813]: I1201 08:45:00.213753 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29409645-lwkbv"] Dec 01 08:45:00 crc kubenswrapper[4813]: I1201 08:45:00.215392 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 01 08:45:00 crc kubenswrapper[4813]: I1201 08:45:00.331040 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/9a00fc8f-1b64-4198-ba46-e93661b8e9db-config-volume\") pod \"collect-profiles-29409645-lwkbv\" (UID: \"9a00fc8f-1b64-4198-ba46-e93661b8e9db\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29409645-lwkbv" Dec 01 08:45:00 crc kubenswrapper[4813]: I1201 08:45:00.331351 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/9a00fc8f-1b64-4198-ba46-e93661b8e9db-secret-volume\") pod \"collect-profiles-29409645-lwkbv\" (UID: \"9a00fc8f-1b64-4198-ba46-e93661b8e9db\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29409645-lwkbv" Dec 01 08:45:00 crc kubenswrapper[4813]: I1201 08:45:00.331432 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gcx5m\" (UniqueName: \"kubernetes.io/projected/9a00fc8f-1b64-4198-ba46-e93661b8e9db-kube-api-access-gcx5m\") pod \"collect-profiles-29409645-lwkbv\" (UID: \"9a00fc8f-1b64-4198-ba46-e93661b8e9db\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29409645-lwkbv" Dec 01 08:45:00 crc kubenswrapper[4813]: I1201 08:45:00.433192 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gcx5m\" (UniqueName: \"kubernetes.io/projected/9a00fc8f-1b64-4198-ba46-e93661b8e9db-kube-api-access-gcx5m\") pod \"collect-profiles-29409645-lwkbv\" (UID: \"9a00fc8f-1b64-4198-ba46-e93661b8e9db\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29409645-lwkbv" Dec 01 08:45:00 crc kubenswrapper[4813]: I1201 08:45:00.433295 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/9a00fc8f-1b64-4198-ba46-e93661b8e9db-config-volume\") pod \"collect-profiles-29409645-lwkbv\" (UID: \"9a00fc8f-1b64-4198-ba46-e93661b8e9db\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29409645-lwkbv" Dec 01 08:45:00 crc kubenswrapper[4813]: I1201 08:45:00.433320 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/9a00fc8f-1b64-4198-ba46-e93661b8e9db-secret-volume\") pod \"collect-profiles-29409645-lwkbv\" (UID: \"9a00fc8f-1b64-4198-ba46-e93661b8e9db\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29409645-lwkbv" Dec 01 08:45:00 crc kubenswrapper[4813]: I1201 08:45:00.435098 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/9a00fc8f-1b64-4198-ba46-e93661b8e9db-config-volume\") pod \"collect-profiles-29409645-lwkbv\" (UID: \"9a00fc8f-1b64-4198-ba46-e93661b8e9db\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29409645-lwkbv" Dec 01 08:45:00 crc kubenswrapper[4813]: I1201 08:45:00.443057 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/9a00fc8f-1b64-4198-ba46-e93661b8e9db-secret-volume\") pod \"collect-profiles-29409645-lwkbv\" (UID: \"9a00fc8f-1b64-4198-ba46-e93661b8e9db\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29409645-lwkbv" Dec 01 08:45:00 crc kubenswrapper[4813]: I1201 08:45:00.458357 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gcx5m\" (UniqueName: \"kubernetes.io/projected/9a00fc8f-1b64-4198-ba46-e93661b8e9db-kube-api-access-gcx5m\") pod \"collect-profiles-29409645-lwkbv\" (UID: \"9a00fc8f-1b64-4198-ba46-e93661b8e9db\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29409645-lwkbv" Dec 01 08:45:00 crc kubenswrapper[4813]: I1201 08:45:00.533477 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29409645-lwkbv" Dec 01 08:45:01 crc kubenswrapper[4813]: I1201 08:45:01.021203 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29409645-lwkbv"] Dec 01 08:45:01 crc kubenswrapper[4813]: I1201 08:45:01.228646 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29409645-lwkbv" event={"ID":"9a00fc8f-1b64-4198-ba46-e93661b8e9db","Type":"ContainerStarted","Data":"91bb5c2dc07830eb7b3182bb172ffa1afa12fb76ce788e88013b926ea11d242b"} Dec 01 08:45:01 crc kubenswrapper[4813]: I1201 08:45:01.228699 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29409645-lwkbv" event={"ID":"9a00fc8f-1b64-4198-ba46-e93661b8e9db","Type":"ContainerStarted","Data":"dcf3ea3e18fc217232bea101e75348f805c45e19f9ece480c1514f8c56dfd625"} Dec 01 08:45:01 crc kubenswrapper[4813]: I1201 08:45:01.254068 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29409645-lwkbv" podStartSLOduration=1.2540411009999999 podStartE2EDuration="1.254041101s" podCreationTimestamp="2025-12-01 08:45:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:45:01.241242117 +0000 UTC m=+285.864163723" watchObservedRunningTime="2025-12-01 08:45:01.254041101 +0000 UTC m=+285.876962687" Dec 01 08:45:02 crc kubenswrapper[4813]: I1201 08:45:02.237400 4813 generic.go:334] "Generic (PLEG): container finished" podID="9a00fc8f-1b64-4198-ba46-e93661b8e9db" containerID="91bb5c2dc07830eb7b3182bb172ffa1afa12fb76ce788e88013b926ea11d242b" exitCode=0 Dec 01 08:45:02 crc kubenswrapper[4813]: I1201 08:45:02.237506 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29409645-lwkbv" event={"ID":"9a00fc8f-1b64-4198-ba46-e93661b8e9db","Type":"ContainerDied","Data":"91bb5c2dc07830eb7b3182bb172ffa1afa12fb76ce788e88013b926ea11d242b"} Dec 01 08:45:03 crc kubenswrapper[4813]: I1201 08:45:03.655232 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29409645-lwkbv" Dec 01 08:45:03 crc kubenswrapper[4813]: I1201 08:45:03.781060 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/9a00fc8f-1b64-4198-ba46-e93661b8e9db-secret-volume\") pod \"9a00fc8f-1b64-4198-ba46-e93661b8e9db\" (UID: \"9a00fc8f-1b64-4198-ba46-e93661b8e9db\") " Dec 01 08:45:03 crc kubenswrapper[4813]: I1201 08:45:03.781206 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/9a00fc8f-1b64-4198-ba46-e93661b8e9db-config-volume\") pod \"9a00fc8f-1b64-4198-ba46-e93661b8e9db\" (UID: \"9a00fc8f-1b64-4198-ba46-e93661b8e9db\") " Dec 01 08:45:03 crc kubenswrapper[4813]: I1201 08:45:03.781251 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gcx5m\" (UniqueName: \"kubernetes.io/projected/9a00fc8f-1b64-4198-ba46-e93661b8e9db-kube-api-access-gcx5m\") pod \"9a00fc8f-1b64-4198-ba46-e93661b8e9db\" (UID: \"9a00fc8f-1b64-4198-ba46-e93661b8e9db\") " Dec 01 08:45:03 crc kubenswrapper[4813]: I1201 08:45:03.782786 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9a00fc8f-1b64-4198-ba46-e93661b8e9db-config-volume" (OuterVolumeSpecName: "config-volume") pod "9a00fc8f-1b64-4198-ba46-e93661b8e9db" (UID: "9a00fc8f-1b64-4198-ba46-e93661b8e9db"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:45:03 crc kubenswrapper[4813]: I1201 08:45:03.788783 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9a00fc8f-1b64-4198-ba46-e93661b8e9db-kube-api-access-gcx5m" (OuterVolumeSpecName: "kube-api-access-gcx5m") pod "9a00fc8f-1b64-4198-ba46-e93661b8e9db" (UID: "9a00fc8f-1b64-4198-ba46-e93661b8e9db"). InnerVolumeSpecName "kube-api-access-gcx5m". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:45:03 crc kubenswrapper[4813]: I1201 08:45:03.789581 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9a00fc8f-1b64-4198-ba46-e93661b8e9db-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "9a00fc8f-1b64-4198-ba46-e93661b8e9db" (UID: "9a00fc8f-1b64-4198-ba46-e93661b8e9db"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:45:03 crc kubenswrapper[4813]: I1201 08:45:03.882870 4813 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/9a00fc8f-1b64-4198-ba46-e93661b8e9db-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 01 08:45:03 crc kubenswrapper[4813]: I1201 08:45:03.882921 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gcx5m\" (UniqueName: \"kubernetes.io/projected/9a00fc8f-1b64-4198-ba46-e93661b8e9db-kube-api-access-gcx5m\") on node \"crc\" DevicePath \"\"" Dec 01 08:45:03 crc kubenswrapper[4813]: I1201 08:45:03.882943 4813 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/9a00fc8f-1b64-4198-ba46-e93661b8e9db-config-volume\") on node \"crc\" DevicePath \"\"" Dec 01 08:45:04 crc kubenswrapper[4813]: I1201 08:45:04.252173 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29409645-lwkbv" event={"ID":"9a00fc8f-1b64-4198-ba46-e93661b8e9db","Type":"ContainerDied","Data":"dcf3ea3e18fc217232bea101e75348f805c45e19f9ece480c1514f8c56dfd625"} Dec 01 08:45:04 crc kubenswrapper[4813]: I1201 08:45:04.252216 4813 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="dcf3ea3e18fc217232bea101e75348f805c45e19f9ece480c1514f8c56dfd625" Dec 01 08:45:04 crc kubenswrapper[4813]: I1201 08:45:04.252298 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29409645-lwkbv" Dec 01 08:45:28 crc kubenswrapper[4813]: I1201 08:45:28.182909 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-f86fd7585-cfv88"] Dec 01 08:45:28 crc kubenswrapper[4813]: I1201 08:45:28.187063 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-f86fd7585-cfv88" podUID="e0324384-89bb-42c0-8597-55fee0b4bb17" containerName="controller-manager" containerID="cri-o://5ee93d1d8abaaf8342f8e83f8c2bff101787fbf77c30d25963d6aba49c033174" gracePeriod=30 Dec 01 08:45:28 crc kubenswrapper[4813]: I1201 08:45:28.467798 4813 generic.go:334] "Generic (PLEG): container finished" podID="e0324384-89bb-42c0-8597-55fee0b4bb17" containerID="5ee93d1d8abaaf8342f8e83f8c2bff101787fbf77c30d25963d6aba49c033174" exitCode=0 Dec 01 08:45:28 crc kubenswrapper[4813]: I1201 08:45:28.467937 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-f86fd7585-cfv88" event={"ID":"e0324384-89bb-42c0-8597-55fee0b4bb17","Type":"ContainerDied","Data":"5ee93d1d8abaaf8342f8e83f8c2bff101787fbf77c30d25963d6aba49c033174"} Dec 01 08:45:28 crc kubenswrapper[4813]: I1201 08:45:28.659835 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-f86fd7585-cfv88" Dec 01 08:45:28 crc kubenswrapper[4813]: I1201 08:45:28.783539 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-69xlg\" (UniqueName: \"kubernetes.io/projected/e0324384-89bb-42c0-8597-55fee0b4bb17-kube-api-access-69xlg\") pod \"e0324384-89bb-42c0-8597-55fee0b4bb17\" (UID: \"e0324384-89bb-42c0-8597-55fee0b4bb17\") " Dec 01 08:45:28 crc kubenswrapper[4813]: I1201 08:45:28.783615 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e0324384-89bb-42c0-8597-55fee0b4bb17-serving-cert\") pod \"e0324384-89bb-42c0-8597-55fee0b4bb17\" (UID: \"e0324384-89bb-42c0-8597-55fee0b4bb17\") " Dec 01 08:45:28 crc kubenswrapper[4813]: I1201 08:45:28.783658 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e0324384-89bb-42c0-8597-55fee0b4bb17-client-ca\") pod \"e0324384-89bb-42c0-8597-55fee0b4bb17\" (UID: \"e0324384-89bb-42c0-8597-55fee0b4bb17\") " Dec 01 08:45:28 crc kubenswrapper[4813]: I1201 08:45:28.783702 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e0324384-89bb-42c0-8597-55fee0b4bb17-config\") pod \"e0324384-89bb-42c0-8597-55fee0b4bb17\" (UID: \"e0324384-89bb-42c0-8597-55fee0b4bb17\") " Dec 01 08:45:28 crc kubenswrapper[4813]: I1201 08:45:28.783768 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e0324384-89bb-42c0-8597-55fee0b4bb17-proxy-ca-bundles\") pod \"e0324384-89bb-42c0-8597-55fee0b4bb17\" (UID: \"e0324384-89bb-42c0-8597-55fee0b4bb17\") " Dec 01 08:45:28 crc kubenswrapper[4813]: I1201 08:45:28.784807 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e0324384-89bb-42c0-8597-55fee0b4bb17-client-ca" (OuterVolumeSpecName: "client-ca") pod "e0324384-89bb-42c0-8597-55fee0b4bb17" (UID: "e0324384-89bb-42c0-8597-55fee0b4bb17"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:45:28 crc kubenswrapper[4813]: I1201 08:45:28.784818 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e0324384-89bb-42c0-8597-55fee0b4bb17-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "e0324384-89bb-42c0-8597-55fee0b4bb17" (UID: "e0324384-89bb-42c0-8597-55fee0b4bb17"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:45:28 crc kubenswrapper[4813]: I1201 08:45:28.785429 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e0324384-89bb-42c0-8597-55fee0b4bb17-config" (OuterVolumeSpecName: "config") pod "e0324384-89bb-42c0-8597-55fee0b4bb17" (UID: "e0324384-89bb-42c0-8597-55fee0b4bb17"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:45:28 crc kubenswrapper[4813]: I1201 08:45:28.790887 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e0324384-89bb-42c0-8597-55fee0b4bb17-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e0324384-89bb-42c0-8597-55fee0b4bb17" (UID: "e0324384-89bb-42c0-8597-55fee0b4bb17"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:45:28 crc kubenswrapper[4813]: I1201 08:45:28.796126 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e0324384-89bb-42c0-8597-55fee0b4bb17-kube-api-access-69xlg" (OuterVolumeSpecName: "kube-api-access-69xlg") pod "e0324384-89bb-42c0-8597-55fee0b4bb17" (UID: "e0324384-89bb-42c0-8597-55fee0b4bb17"). InnerVolumeSpecName "kube-api-access-69xlg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:45:28 crc kubenswrapper[4813]: I1201 08:45:28.885092 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-69xlg\" (UniqueName: \"kubernetes.io/projected/e0324384-89bb-42c0-8597-55fee0b4bb17-kube-api-access-69xlg\") on node \"crc\" DevicePath \"\"" Dec 01 08:45:28 crc kubenswrapper[4813]: I1201 08:45:28.885143 4813 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e0324384-89bb-42c0-8597-55fee0b4bb17-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 01 08:45:28 crc kubenswrapper[4813]: I1201 08:45:28.885158 4813 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e0324384-89bb-42c0-8597-55fee0b4bb17-client-ca\") on node \"crc\" DevicePath \"\"" Dec 01 08:45:28 crc kubenswrapper[4813]: I1201 08:45:28.885188 4813 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e0324384-89bb-42c0-8597-55fee0b4bb17-config\") on node \"crc\" DevicePath \"\"" Dec 01 08:45:28 crc kubenswrapper[4813]: I1201 08:45:28.885203 4813 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e0324384-89bb-42c0-8597-55fee0b4bb17-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 01 08:45:29 crc kubenswrapper[4813]: I1201 08:45:29.374898 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-789645f746-8k644"] Dec 01 08:45:29 crc kubenswrapper[4813]: E1201 08:45:29.375568 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a00fc8f-1b64-4198-ba46-e93661b8e9db" containerName="collect-profiles" Dec 01 08:45:29 crc kubenswrapper[4813]: I1201 08:45:29.375602 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a00fc8f-1b64-4198-ba46-e93661b8e9db" containerName="collect-profiles" Dec 01 08:45:29 crc kubenswrapper[4813]: E1201 08:45:29.375629 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e0324384-89bb-42c0-8597-55fee0b4bb17" containerName="controller-manager" Dec 01 08:45:29 crc kubenswrapper[4813]: I1201 08:45:29.375636 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="e0324384-89bb-42c0-8597-55fee0b4bb17" containerName="controller-manager" Dec 01 08:45:29 crc kubenswrapper[4813]: I1201 08:45:29.375787 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="e0324384-89bb-42c0-8597-55fee0b4bb17" containerName="controller-manager" Dec 01 08:45:29 crc kubenswrapper[4813]: I1201 08:45:29.375803 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="9a00fc8f-1b64-4198-ba46-e93661b8e9db" containerName="collect-profiles" Dec 01 08:45:29 crc kubenswrapper[4813]: I1201 08:45:29.380269 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-789645f746-8k644" Dec 01 08:45:29 crc kubenswrapper[4813]: I1201 08:45:29.383799 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-789645f746-8k644"] Dec 01 08:45:29 crc kubenswrapper[4813]: I1201 08:45:29.456848 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/1c2327af-db2c-4c64-8cec-8316684c5155-proxy-ca-bundles\") pod \"controller-manager-789645f746-8k644\" (UID: \"1c2327af-db2c-4c64-8cec-8316684c5155\") " pod="openshift-controller-manager/controller-manager-789645f746-8k644" Dec 01 08:45:29 crc kubenswrapper[4813]: I1201 08:45:29.456938 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1c2327af-db2c-4c64-8cec-8316684c5155-config\") pod \"controller-manager-789645f746-8k644\" (UID: \"1c2327af-db2c-4c64-8cec-8316684c5155\") " pod="openshift-controller-manager/controller-manager-789645f746-8k644" Dec 01 08:45:29 crc kubenswrapper[4813]: I1201 08:45:29.456986 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1c2327af-db2c-4c64-8cec-8316684c5155-client-ca\") pod \"controller-manager-789645f746-8k644\" (UID: \"1c2327af-db2c-4c64-8cec-8316684c5155\") " pod="openshift-controller-manager/controller-manager-789645f746-8k644" Dec 01 08:45:29 crc kubenswrapper[4813]: I1201 08:45:29.457011 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c8zbj\" (UniqueName: \"kubernetes.io/projected/1c2327af-db2c-4c64-8cec-8316684c5155-kube-api-access-c8zbj\") pod \"controller-manager-789645f746-8k644\" (UID: \"1c2327af-db2c-4c64-8cec-8316684c5155\") " pod="openshift-controller-manager/controller-manager-789645f746-8k644" Dec 01 08:45:29 crc kubenswrapper[4813]: I1201 08:45:29.457291 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1c2327af-db2c-4c64-8cec-8316684c5155-serving-cert\") pod \"controller-manager-789645f746-8k644\" (UID: \"1c2327af-db2c-4c64-8cec-8316684c5155\") " pod="openshift-controller-manager/controller-manager-789645f746-8k644" Dec 01 08:45:29 crc kubenswrapper[4813]: I1201 08:45:29.487302 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-f86fd7585-cfv88" event={"ID":"e0324384-89bb-42c0-8597-55fee0b4bb17","Type":"ContainerDied","Data":"a1d32a47c79704b9835946ce477b8adab192b21cf6edd19eed2e214e4e8812db"} Dec 01 08:45:29 crc kubenswrapper[4813]: I1201 08:45:29.487384 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-f86fd7585-cfv88" Dec 01 08:45:29 crc kubenswrapper[4813]: I1201 08:45:29.487431 4813 scope.go:117] "RemoveContainer" containerID="5ee93d1d8abaaf8342f8e83f8c2bff101787fbf77c30d25963d6aba49c033174" Dec 01 08:45:29 crc kubenswrapper[4813]: I1201 08:45:29.520863 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-f86fd7585-cfv88"] Dec 01 08:45:29 crc kubenswrapper[4813]: I1201 08:45:29.524043 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-f86fd7585-cfv88"] Dec 01 08:45:29 crc kubenswrapper[4813]: I1201 08:45:29.558758 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1c2327af-db2c-4c64-8cec-8316684c5155-config\") pod \"controller-manager-789645f746-8k644\" (UID: \"1c2327af-db2c-4c64-8cec-8316684c5155\") " pod="openshift-controller-manager/controller-manager-789645f746-8k644" Dec 01 08:45:29 crc kubenswrapper[4813]: I1201 08:45:29.558837 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1c2327af-db2c-4c64-8cec-8316684c5155-client-ca\") pod \"controller-manager-789645f746-8k644\" (UID: \"1c2327af-db2c-4c64-8cec-8316684c5155\") " pod="openshift-controller-manager/controller-manager-789645f746-8k644" Dec 01 08:45:29 crc kubenswrapper[4813]: I1201 08:45:29.558868 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c8zbj\" (UniqueName: \"kubernetes.io/projected/1c2327af-db2c-4c64-8cec-8316684c5155-kube-api-access-c8zbj\") pod \"controller-manager-789645f746-8k644\" (UID: \"1c2327af-db2c-4c64-8cec-8316684c5155\") " pod="openshift-controller-manager/controller-manager-789645f746-8k644" Dec 01 08:45:29 crc kubenswrapper[4813]: I1201 08:45:29.558996 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1c2327af-db2c-4c64-8cec-8316684c5155-serving-cert\") pod \"controller-manager-789645f746-8k644\" (UID: \"1c2327af-db2c-4c64-8cec-8316684c5155\") " pod="openshift-controller-manager/controller-manager-789645f746-8k644" Dec 01 08:45:29 crc kubenswrapper[4813]: I1201 08:45:29.559057 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/1c2327af-db2c-4c64-8cec-8316684c5155-proxy-ca-bundles\") pod \"controller-manager-789645f746-8k644\" (UID: \"1c2327af-db2c-4c64-8cec-8316684c5155\") " pod="openshift-controller-manager/controller-manager-789645f746-8k644" Dec 01 08:45:29 crc kubenswrapper[4813]: I1201 08:45:29.560354 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1c2327af-db2c-4c64-8cec-8316684c5155-client-ca\") pod \"controller-manager-789645f746-8k644\" (UID: \"1c2327af-db2c-4c64-8cec-8316684c5155\") " pod="openshift-controller-manager/controller-manager-789645f746-8k644" Dec 01 08:45:29 crc kubenswrapper[4813]: I1201 08:45:29.560846 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/1c2327af-db2c-4c64-8cec-8316684c5155-proxy-ca-bundles\") pod \"controller-manager-789645f746-8k644\" (UID: \"1c2327af-db2c-4c64-8cec-8316684c5155\") " pod="openshift-controller-manager/controller-manager-789645f746-8k644" Dec 01 08:45:29 crc kubenswrapper[4813]: I1201 08:45:29.561593 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1c2327af-db2c-4c64-8cec-8316684c5155-config\") pod \"controller-manager-789645f746-8k644\" (UID: \"1c2327af-db2c-4c64-8cec-8316684c5155\") " pod="openshift-controller-manager/controller-manager-789645f746-8k644" Dec 01 08:45:29 crc kubenswrapper[4813]: I1201 08:45:29.566665 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1c2327af-db2c-4c64-8cec-8316684c5155-serving-cert\") pod \"controller-manager-789645f746-8k644\" (UID: \"1c2327af-db2c-4c64-8cec-8316684c5155\") " pod="openshift-controller-manager/controller-manager-789645f746-8k644" Dec 01 08:45:29 crc kubenswrapper[4813]: I1201 08:45:29.576196 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c8zbj\" (UniqueName: \"kubernetes.io/projected/1c2327af-db2c-4c64-8cec-8316684c5155-kube-api-access-c8zbj\") pod \"controller-manager-789645f746-8k644\" (UID: \"1c2327af-db2c-4c64-8cec-8316684c5155\") " pod="openshift-controller-manager/controller-manager-789645f746-8k644" Dec 01 08:45:29 crc kubenswrapper[4813]: I1201 08:45:29.702365 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-789645f746-8k644" Dec 01 08:45:30 crc kubenswrapper[4813]: I1201 08:45:30.014163 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-789645f746-8k644"] Dec 01 08:45:30 crc kubenswrapper[4813]: I1201 08:45:30.403109 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e0324384-89bb-42c0-8597-55fee0b4bb17" path="/var/lib/kubelet/pods/e0324384-89bb-42c0-8597-55fee0b4bb17/volumes" Dec 01 08:45:30 crc kubenswrapper[4813]: I1201 08:45:30.495101 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-789645f746-8k644" event={"ID":"1c2327af-db2c-4c64-8cec-8316684c5155","Type":"ContainerStarted","Data":"4cd4e4fc29aa29928e064252449ce8c99f8dfae08f54596abfa289bd24edf6f9"} Dec 01 08:45:30 crc kubenswrapper[4813]: I1201 08:45:30.495506 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-789645f746-8k644" Dec 01 08:45:30 crc kubenswrapper[4813]: I1201 08:45:30.495525 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-789645f746-8k644" event={"ID":"1c2327af-db2c-4c64-8cec-8316684c5155","Type":"ContainerStarted","Data":"768be75248c3f39de173ee730ae766a961d344f3ed93ae73349394faeb39b441"} Dec 01 08:45:30 crc kubenswrapper[4813]: I1201 08:45:30.500997 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-789645f746-8k644" Dec 01 08:45:30 crc kubenswrapper[4813]: I1201 08:45:30.516153 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-789645f746-8k644" podStartSLOduration=2.5161167669999998 podStartE2EDuration="2.516116767s" podCreationTimestamp="2025-12-01 08:45:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:45:30.51244815 +0000 UTC m=+315.135369736" watchObservedRunningTime="2025-12-01 08:45:30.516116767 +0000 UTC m=+315.139038353" Dec 01 08:45:34 crc kubenswrapper[4813]: I1201 08:45:34.040482 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-qh5hd"] Dec 01 08:45:34 crc kubenswrapper[4813]: I1201 08:45:34.042976 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-qh5hd" podUID="26f4804b-998f-488c-a501-d1cc4ef7f0b4" containerName="registry-server" containerID="cri-o://1acc7744b9f14eab07ba59bc0c898cc8e8c944a526dd3599948ffbbaf19aabfd" gracePeriod=30 Dec 01 08:45:34 crc kubenswrapper[4813]: I1201 08:45:34.050924 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-7fqxs"] Dec 01 08:45:34 crc kubenswrapper[4813]: I1201 08:45:34.051415 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-7fqxs" podUID="0959ee73-31e7-4d58-b124-3e971d777225" containerName="registry-server" containerID="cri-o://e33294dfa539d5c4653c0b4bf83d01a72b7e2848567f6ec9f836d922d084cbd4" gracePeriod=30 Dec 01 08:45:34 crc kubenswrapper[4813]: I1201 08:45:34.061610 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-qsskc"] Dec 01 08:45:34 crc kubenswrapper[4813]: I1201 08:45:34.061854 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-qsskc" podUID="c342090f-f331-484d-b287-929d8303f5d7" containerName="marketplace-operator" containerID="cri-o://9c9df537248bcf5a73e1bcdb2c8e30af1d91f39aac66106534a6c2aa1f5fab7a" gracePeriod=30 Dec 01 08:45:34 crc kubenswrapper[4813]: I1201 08:45:34.073923 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-lk5db"] Dec 01 08:45:34 crc kubenswrapper[4813]: I1201 08:45:34.074286 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-lk5db" podUID="4f1619ff-e7ed-42ee-b7a9-484c18d6677d" containerName="registry-server" containerID="cri-o://9ffc53906faf61c55a06e12e23404613fa25399b22cf92a6835b44ae83f7eb5a" gracePeriod=30 Dec 01 08:45:34 crc kubenswrapper[4813]: I1201 08:45:34.083343 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-wbvcp"] Dec 01 08:45:34 crc kubenswrapper[4813]: I1201 08:45:34.083664 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-wbvcp" podUID="979ea876-4dd3-4d31-9144-7a9a6e914c67" containerName="registry-server" containerID="cri-o://b449598529a2bf89bc7d1b288f3aef72364ec65d95f4efa1630f2e0298648fcb" gracePeriod=30 Dec 01 08:45:34 crc kubenswrapper[4813]: I1201 08:45:34.165929 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-nmwnn"] Dec 01 08:45:34 crc kubenswrapper[4813]: I1201 08:45:34.174689 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-nmwnn" Dec 01 08:45:34 crc kubenswrapper[4813]: I1201 08:45:34.197413 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-nmwnn"] Dec 01 08:45:34 crc kubenswrapper[4813]: I1201 08:45:34.327630 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4qdr7\" (UniqueName: \"kubernetes.io/projected/8d57deaa-d1bd-421f-84c3-8682472f1491-kube-api-access-4qdr7\") pod \"marketplace-operator-79b997595-nmwnn\" (UID: \"8d57deaa-d1bd-421f-84c3-8682472f1491\") " pod="openshift-marketplace/marketplace-operator-79b997595-nmwnn" Dec 01 08:45:34 crc kubenswrapper[4813]: I1201 08:45:34.327684 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8d57deaa-d1bd-421f-84c3-8682472f1491-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-nmwnn\" (UID: \"8d57deaa-d1bd-421f-84c3-8682472f1491\") " pod="openshift-marketplace/marketplace-operator-79b997595-nmwnn" Dec 01 08:45:34 crc kubenswrapper[4813]: I1201 08:45:34.327763 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/8d57deaa-d1bd-421f-84c3-8682472f1491-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-nmwnn\" (UID: \"8d57deaa-d1bd-421f-84c3-8682472f1491\") " pod="openshift-marketplace/marketplace-operator-79b997595-nmwnn" Dec 01 08:45:34 crc kubenswrapper[4813]: I1201 08:45:34.428878 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4qdr7\" (UniqueName: \"kubernetes.io/projected/8d57deaa-d1bd-421f-84c3-8682472f1491-kube-api-access-4qdr7\") pod \"marketplace-operator-79b997595-nmwnn\" (UID: \"8d57deaa-d1bd-421f-84c3-8682472f1491\") " pod="openshift-marketplace/marketplace-operator-79b997595-nmwnn" Dec 01 08:45:34 crc kubenswrapper[4813]: I1201 08:45:34.428935 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8d57deaa-d1bd-421f-84c3-8682472f1491-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-nmwnn\" (UID: \"8d57deaa-d1bd-421f-84c3-8682472f1491\") " pod="openshift-marketplace/marketplace-operator-79b997595-nmwnn" Dec 01 08:45:34 crc kubenswrapper[4813]: I1201 08:45:34.429024 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/8d57deaa-d1bd-421f-84c3-8682472f1491-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-nmwnn\" (UID: \"8d57deaa-d1bd-421f-84c3-8682472f1491\") " pod="openshift-marketplace/marketplace-operator-79b997595-nmwnn" Dec 01 08:45:34 crc kubenswrapper[4813]: I1201 08:45:34.430842 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8d57deaa-d1bd-421f-84c3-8682472f1491-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-nmwnn\" (UID: \"8d57deaa-d1bd-421f-84c3-8682472f1491\") " pod="openshift-marketplace/marketplace-operator-79b997595-nmwnn" Dec 01 08:45:34 crc kubenswrapper[4813]: I1201 08:45:34.456238 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/8d57deaa-d1bd-421f-84c3-8682472f1491-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-nmwnn\" (UID: \"8d57deaa-d1bd-421f-84c3-8682472f1491\") " pod="openshift-marketplace/marketplace-operator-79b997595-nmwnn" Dec 01 08:45:34 crc kubenswrapper[4813]: I1201 08:45:34.469570 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4qdr7\" (UniqueName: \"kubernetes.io/projected/8d57deaa-d1bd-421f-84c3-8682472f1491-kube-api-access-4qdr7\") pod \"marketplace-operator-79b997595-nmwnn\" (UID: \"8d57deaa-d1bd-421f-84c3-8682472f1491\") " pod="openshift-marketplace/marketplace-operator-79b997595-nmwnn" Dec 01 08:45:34 crc kubenswrapper[4813]: I1201 08:45:34.515557 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-nmwnn" Dec 01 08:45:34 crc kubenswrapper[4813]: I1201 08:45:34.525393 4813 generic.go:334] "Generic (PLEG): container finished" podID="979ea876-4dd3-4d31-9144-7a9a6e914c67" containerID="b449598529a2bf89bc7d1b288f3aef72364ec65d95f4efa1630f2e0298648fcb" exitCode=0 Dec 01 08:45:34 crc kubenswrapper[4813]: I1201 08:45:34.525454 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wbvcp" event={"ID":"979ea876-4dd3-4d31-9144-7a9a6e914c67","Type":"ContainerDied","Data":"b449598529a2bf89bc7d1b288f3aef72364ec65d95f4efa1630f2e0298648fcb"} Dec 01 08:45:34 crc kubenswrapper[4813]: I1201 08:45:34.527389 4813 generic.go:334] "Generic (PLEG): container finished" podID="26f4804b-998f-488c-a501-d1cc4ef7f0b4" containerID="1acc7744b9f14eab07ba59bc0c898cc8e8c944a526dd3599948ffbbaf19aabfd" exitCode=0 Dec 01 08:45:34 crc kubenswrapper[4813]: I1201 08:45:34.527451 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qh5hd" event={"ID":"26f4804b-998f-488c-a501-d1cc4ef7f0b4","Type":"ContainerDied","Data":"1acc7744b9f14eab07ba59bc0c898cc8e8c944a526dd3599948ffbbaf19aabfd"} Dec 01 08:45:34 crc kubenswrapper[4813]: I1201 08:45:34.533352 4813 generic.go:334] "Generic (PLEG): container finished" podID="4f1619ff-e7ed-42ee-b7a9-484c18d6677d" containerID="9ffc53906faf61c55a06e12e23404613fa25399b22cf92a6835b44ae83f7eb5a" exitCode=0 Dec 01 08:45:34 crc kubenswrapper[4813]: I1201 08:45:34.533429 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lk5db" event={"ID":"4f1619ff-e7ed-42ee-b7a9-484c18d6677d","Type":"ContainerDied","Data":"9ffc53906faf61c55a06e12e23404613fa25399b22cf92a6835b44ae83f7eb5a"} Dec 01 08:45:34 crc kubenswrapper[4813]: I1201 08:45:34.542534 4813 generic.go:334] "Generic (PLEG): container finished" podID="c342090f-f331-484d-b287-929d8303f5d7" containerID="9c9df537248bcf5a73e1bcdb2c8e30af1d91f39aac66106534a6c2aa1f5fab7a" exitCode=0 Dec 01 08:45:34 crc kubenswrapper[4813]: I1201 08:45:34.542598 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-qsskc" event={"ID":"c342090f-f331-484d-b287-929d8303f5d7","Type":"ContainerDied","Data":"9c9df537248bcf5a73e1bcdb2c8e30af1d91f39aac66106534a6c2aa1f5fab7a"} Dec 01 08:45:34 crc kubenswrapper[4813]: I1201 08:45:34.542635 4813 scope.go:117] "RemoveContainer" containerID="675727fde89d8e51f7ed3ff2b83d39e2917fa51a8b9eb117e74b96316e396b4f" Dec 01 08:45:34 crc kubenswrapper[4813]: I1201 08:45:34.545397 4813 generic.go:334] "Generic (PLEG): container finished" podID="0959ee73-31e7-4d58-b124-3e971d777225" containerID="e33294dfa539d5c4653c0b4bf83d01a72b7e2848567f6ec9f836d922d084cbd4" exitCode=0 Dec 01 08:45:34 crc kubenswrapper[4813]: I1201 08:45:34.545425 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7fqxs" event={"ID":"0959ee73-31e7-4d58-b124-3e971d777225","Type":"ContainerDied","Data":"e33294dfa539d5c4653c0b4bf83d01a72b7e2848567f6ec9f836d922d084cbd4"} Dec 01 08:45:34 crc kubenswrapper[4813]: I1201 08:45:34.590274 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-qh5hd" Dec 01 08:45:34 crc kubenswrapper[4813]: I1201 08:45:34.734519 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/26f4804b-998f-488c-a501-d1cc4ef7f0b4-utilities\") pod \"26f4804b-998f-488c-a501-d1cc4ef7f0b4\" (UID: \"26f4804b-998f-488c-a501-d1cc4ef7f0b4\") " Dec 01 08:45:34 crc kubenswrapper[4813]: I1201 08:45:34.734617 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/26f4804b-998f-488c-a501-d1cc4ef7f0b4-catalog-content\") pod \"26f4804b-998f-488c-a501-d1cc4ef7f0b4\" (UID: \"26f4804b-998f-488c-a501-d1cc4ef7f0b4\") " Dec 01 08:45:34 crc kubenswrapper[4813]: I1201 08:45:34.734652 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ns6n2\" (UniqueName: \"kubernetes.io/projected/26f4804b-998f-488c-a501-d1cc4ef7f0b4-kube-api-access-ns6n2\") pod \"26f4804b-998f-488c-a501-d1cc4ef7f0b4\" (UID: \"26f4804b-998f-488c-a501-d1cc4ef7f0b4\") " Dec 01 08:45:34 crc kubenswrapper[4813]: I1201 08:45:34.737498 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/26f4804b-998f-488c-a501-d1cc4ef7f0b4-utilities" (OuterVolumeSpecName: "utilities") pod "26f4804b-998f-488c-a501-d1cc4ef7f0b4" (UID: "26f4804b-998f-488c-a501-d1cc4ef7f0b4"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 08:45:34 crc kubenswrapper[4813]: I1201 08:45:34.739789 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/26f4804b-998f-488c-a501-d1cc4ef7f0b4-kube-api-access-ns6n2" (OuterVolumeSpecName: "kube-api-access-ns6n2") pod "26f4804b-998f-488c-a501-d1cc4ef7f0b4" (UID: "26f4804b-998f-488c-a501-d1cc4ef7f0b4"). InnerVolumeSpecName "kube-api-access-ns6n2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:45:34 crc kubenswrapper[4813]: I1201 08:45:34.784348 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7fqxs" Dec 01 08:45:34 crc kubenswrapper[4813]: I1201 08:45:34.811525 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wbvcp" Dec 01 08:45:34 crc kubenswrapper[4813]: I1201 08:45:34.813247 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/26f4804b-998f-488c-a501-d1cc4ef7f0b4-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "26f4804b-998f-488c-a501-d1cc4ef7f0b4" (UID: "26f4804b-998f-488c-a501-d1cc4ef7f0b4"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 08:45:34 crc kubenswrapper[4813]: I1201 08:45:34.831559 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-lk5db" Dec 01 08:45:34 crc kubenswrapper[4813]: I1201 08:45:34.835830 4813 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/26f4804b-998f-488c-a501-d1cc4ef7f0b4-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 08:45:34 crc kubenswrapper[4813]: I1201 08:45:34.835857 4813 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/26f4804b-998f-488c-a501-d1cc4ef7f0b4-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 08:45:34 crc kubenswrapper[4813]: I1201 08:45:34.835868 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ns6n2\" (UniqueName: \"kubernetes.io/projected/26f4804b-998f-488c-a501-d1cc4ef7f0b4-kube-api-access-ns6n2\") on node \"crc\" DevicePath \"\"" Dec 01 08:45:34 crc kubenswrapper[4813]: I1201 08:45:34.841110 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-qsskc" Dec 01 08:45:34 crc kubenswrapper[4813]: I1201 08:45:34.937220 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4f1619ff-e7ed-42ee-b7a9-484c18d6677d-utilities\") pod \"4f1619ff-e7ed-42ee-b7a9-484c18d6677d\" (UID: \"4f1619ff-e7ed-42ee-b7a9-484c18d6677d\") " Dec 01 08:45:34 crc kubenswrapper[4813]: I1201 08:45:34.937279 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vqnmj\" (UniqueName: \"kubernetes.io/projected/4f1619ff-e7ed-42ee-b7a9-484c18d6677d-kube-api-access-vqnmj\") pod \"4f1619ff-e7ed-42ee-b7a9-484c18d6677d\" (UID: \"4f1619ff-e7ed-42ee-b7a9-484c18d6677d\") " Dec 01 08:45:34 crc kubenswrapper[4813]: I1201 08:45:34.937297 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-srz6m\" (UniqueName: \"kubernetes.io/projected/0959ee73-31e7-4d58-b124-3e971d777225-kube-api-access-srz6m\") pod \"0959ee73-31e7-4d58-b124-3e971d777225\" (UID: \"0959ee73-31e7-4d58-b124-3e971d777225\") " Dec 01 08:45:34 crc kubenswrapper[4813]: I1201 08:45:34.937351 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/979ea876-4dd3-4d31-9144-7a9a6e914c67-catalog-content\") pod \"979ea876-4dd3-4d31-9144-7a9a6e914c67\" (UID: \"979ea876-4dd3-4d31-9144-7a9a6e914c67\") " Dec 01 08:45:34 crc kubenswrapper[4813]: I1201 08:45:34.937984 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0959ee73-31e7-4d58-b124-3e971d777225-catalog-content\") pod \"0959ee73-31e7-4d58-b124-3e971d777225\" (UID: \"0959ee73-31e7-4d58-b124-3e971d777225\") " Dec 01 08:45:34 crc kubenswrapper[4813]: I1201 08:45:34.938120 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4rzrc\" (UniqueName: \"kubernetes.io/projected/979ea876-4dd3-4d31-9144-7a9a6e914c67-kube-api-access-4rzrc\") pod \"979ea876-4dd3-4d31-9144-7a9a6e914c67\" (UID: \"979ea876-4dd3-4d31-9144-7a9a6e914c67\") " Dec 01 08:45:34 crc kubenswrapper[4813]: I1201 08:45:34.938148 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0959ee73-31e7-4d58-b124-3e971d777225-utilities\") pod \"0959ee73-31e7-4d58-b124-3e971d777225\" (UID: \"0959ee73-31e7-4d58-b124-3e971d777225\") " Dec 01 08:45:34 crc kubenswrapper[4813]: I1201 08:45:34.938171 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c342090f-f331-484d-b287-929d8303f5d7-marketplace-trusted-ca\") pod \"c342090f-f331-484d-b287-929d8303f5d7\" (UID: \"c342090f-f331-484d-b287-929d8303f5d7\") " Dec 01 08:45:34 crc kubenswrapper[4813]: I1201 08:45:34.938258 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4f1619ff-e7ed-42ee-b7a9-484c18d6677d-catalog-content\") pod \"4f1619ff-e7ed-42ee-b7a9-484c18d6677d\" (UID: \"4f1619ff-e7ed-42ee-b7a9-484c18d6677d\") " Dec 01 08:45:34 crc kubenswrapper[4813]: I1201 08:45:34.938283 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/979ea876-4dd3-4d31-9144-7a9a6e914c67-utilities\") pod \"979ea876-4dd3-4d31-9144-7a9a6e914c67\" (UID: \"979ea876-4dd3-4d31-9144-7a9a6e914c67\") " Dec 01 08:45:34 crc kubenswrapper[4813]: I1201 08:45:34.938329 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jm2f5\" (UniqueName: \"kubernetes.io/projected/c342090f-f331-484d-b287-929d8303f5d7-kube-api-access-jm2f5\") pod \"c342090f-f331-484d-b287-929d8303f5d7\" (UID: \"c342090f-f331-484d-b287-929d8303f5d7\") " Dec 01 08:45:34 crc kubenswrapper[4813]: I1201 08:45:34.938787 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0959ee73-31e7-4d58-b124-3e971d777225-utilities" (OuterVolumeSpecName: "utilities") pod "0959ee73-31e7-4d58-b124-3e971d777225" (UID: "0959ee73-31e7-4d58-b124-3e971d777225"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 08:45:34 crc kubenswrapper[4813]: I1201 08:45:34.939220 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c342090f-f331-484d-b287-929d8303f5d7-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "c342090f-f331-484d-b287-929d8303f5d7" (UID: "c342090f-f331-484d-b287-929d8303f5d7"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:45:34 crc kubenswrapper[4813]: I1201 08:45:34.939363 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/979ea876-4dd3-4d31-9144-7a9a6e914c67-utilities" (OuterVolumeSpecName: "utilities") pod "979ea876-4dd3-4d31-9144-7a9a6e914c67" (UID: "979ea876-4dd3-4d31-9144-7a9a6e914c67"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 08:45:34 crc kubenswrapper[4813]: I1201 08:45:34.942567 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4f1619ff-e7ed-42ee-b7a9-484c18d6677d-utilities" (OuterVolumeSpecName: "utilities") pod "4f1619ff-e7ed-42ee-b7a9-484c18d6677d" (UID: "4f1619ff-e7ed-42ee-b7a9-484c18d6677d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 08:45:34 crc kubenswrapper[4813]: I1201 08:45:34.947130 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0959ee73-31e7-4d58-b124-3e971d777225-kube-api-access-srz6m" (OuterVolumeSpecName: "kube-api-access-srz6m") pod "0959ee73-31e7-4d58-b124-3e971d777225" (UID: "0959ee73-31e7-4d58-b124-3e971d777225"). InnerVolumeSpecName "kube-api-access-srz6m". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:45:34 crc kubenswrapper[4813]: I1201 08:45:34.947145 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/979ea876-4dd3-4d31-9144-7a9a6e914c67-kube-api-access-4rzrc" (OuterVolumeSpecName: "kube-api-access-4rzrc") pod "979ea876-4dd3-4d31-9144-7a9a6e914c67" (UID: "979ea876-4dd3-4d31-9144-7a9a6e914c67"). InnerVolumeSpecName "kube-api-access-4rzrc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:45:34 crc kubenswrapper[4813]: I1201 08:45:34.947669 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c342090f-f331-484d-b287-929d8303f5d7-kube-api-access-jm2f5" (OuterVolumeSpecName: "kube-api-access-jm2f5") pod "c342090f-f331-484d-b287-929d8303f5d7" (UID: "c342090f-f331-484d-b287-929d8303f5d7"). InnerVolumeSpecName "kube-api-access-jm2f5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:45:34 crc kubenswrapper[4813]: I1201 08:45:34.948328 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4f1619ff-e7ed-42ee-b7a9-484c18d6677d-kube-api-access-vqnmj" (OuterVolumeSpecName: "kube-api-access-vqnmj") pod "4f1619ff-e7ed-42ee-b7a9-484c18d6677d" (UID: "4f1619ff-e7ed-42ee-b7a9-484c18d6677d"). InnerVolumeSpecName "kube-api-access-vqnmj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:45:34 crc kubenswrapper[4813]: I1201 08:45:34.958367 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4f1619ff-e7ed-42ee-b7a9-484c18d6677d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4f1619ff-e7ed-42ee-b7a9-484c18d6677d" (UID: "4f1619ff-e7ed-42ee-b7a9-484c18d6677d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 08:45:34 crc kubenswrapper[4813]: I1201 08:45:34.989379 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0959ee73-31e7-4d58-b124-3e971d777225-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0959ee73-31e7-4d58-b124-3e971d777225" (UID: "0959ee73-31e7-4d58-b124-3e971d777225"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 08:45:35 crc kubenswrapper[4813]: I1201 08:45:35.039379 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/c342090f-f331-484d-b287-929d8303f5d7-marketplace-operator-metrics\") pod \"c342090f-f331-484d-b287-929d8303f5d7\" (UID: \"c342090f-f331-484d-b287-929d8303f5d7\") " Dec 01 08:45:35 crc kubenswrapper[4813]: I1201 08:45:35.039671 4813 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4f1619ff-e7ed-42ee-b7a9-484c18d6677d-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 08:45:35 crc kubenswrapper[4813]: I1201 08:45:35.039697 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vqnmj\" (UniqueName: \"kubernetes.io/projected/4f1619ff-e7ed-42ee-b7a9-484c18d6677d-kube-api-access-vqnmj\") on node \"crc\" DevicePath \"\"" Dec 01 08:45:35 crc kubenswrapper[4813]: I1201 08:45:35.039711 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-srz6m\" (UniqueName: \"kubernetes.io/projected/0959ee73-31e7-4d58-b124-3e971d777225-kube-api-access-srz6m\") on node \"crc\" DevicePath \"\"" Dec 01 08:45:35 crc kubenswrapper[4813]: I1201 08:45:35.039724 4813 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0959ee73-31e7-4d58-b124-3e971d777225-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 08:45:35 crc kubenswrapper[4813]: I1201 08:45:35.039734 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4rzrc\" (UniqueName: \"kubernetes.io/projected/979ea876-4dd3-4d31-9144-7a9a6e914c67-kube-api-access-4rzrc\") on node \"crc\" DevicePath \"\"" Dec 01 08:45:35 crc kubenswrapper[4813]: I1201 08:45:35.039747 4813 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0959ee73-31e7-4d58-b124-3e971d777225-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 08:45:35 crc kubenswrapper[4813]: I1201 08:45:35.039758 4813 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c342090f-f331-484d-b287-929d8303f5d7-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 01 08:45:35 crc kubenswrapper[4813]: I1201 08:45:35.039769 4813 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4f1619ff-e7ed-42ee-b7a9-484c18d6677d-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 08:45:35 crc kubenswrapper[4813]: I1201 08:45:35.039782 4813 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/979ea876-4dd3-4d31-9144-7a9a6e914c67-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 08:45:35 crc kubenswrapper[4813]: I1201 08:45:35.039792 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jm2f5\" (UniqueName: \"kubernetes.io/projected/c342090f-f331-484d-b287-929d8303f5d7-kube-api-access-jm2f5\") on node \"crc\" DevicePath \"\"" Dec 01 08:45:35 crc kubenswrapper[4813]: I1201 08:45:35.045260 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c342090f-f331-484d-b287-929d8303f5d7-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "c342090f-f331-484d-b287-929d8303f5d7" (UID: "c342090f-f331-484d-b287-929d8303f5d7"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:45:35 crc kubenswrapper[4813]: I1201 08:45:35.123115 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/979ea876-4dd3-4d31-9144-7a9a6e914c67-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "979ea876-4dd3-4d31-9144-7a9a6e914c67" (UID: "979ea876-4dd3-4d31-9144-7a9a6e914c67"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 08:45:35 crc kubenswrapper[4813]: I1201 08:45:35.140680 4813 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/979ea876-4dd3-4d31-9144-7a9a6e914c67-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 08:45:35 crc kubenswrapper[4813]: I1201 08:45:35.141018 4813 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/c342090f-f331-484d-b287-929d8303f5d7-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Dec 01 08:45:35 crc kubenswrapper[4813]: I1201 08:45:35.163065 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-nmwnn"] Dec 01 08:45:35 crc kubenswrapper[4813]: I1201 08:45:35.552174 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-qsskc" event={"ID":"c342090f-f331-484d-b287-929d8303f5d7","Type":"ContainerDied","Data":"9bf28095a6c6ccd8c16c480dd35263958c6ff846cbc75ff9300b59da85db1125"} Dec 01 08:45:35 crc kubenswrapper[4813]: I1201 08:45:35.552241 4813 scope.go:117] "RemoveContainer" containerID="9c9df537248bcf5a73e1bcdb2c8e30af1d91f39aac66106534a6c2aa1f5fab7a" Dec 01 08:45:35 crc kubenswrapper[4813]: I1201 08:45:35.552252 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-qsskc" Dec 01 08:45:35 crc kubenswrapper[4813]: I1201 08:45:35.555192 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7fqxs" event={"ID":"0959ee73-31e7-4d58-b124-3e971d777225","Type":"ContainerDied","Data":"c3f740dd049aa2768738295f2186f891e43c9627bfba804b249e2eb6758d4a95"} Dec 01 08:45:35 crc kubenswrapper[4813]: I1201 08:45:35.555295 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7fqxs" Dec 01 08:45:35 crc kubenswrapper[4813]: I1201 08:45:35.558811 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wbvcp" event={"ID":"979ea876-4dd3-4d31-9144-7a9a6e914c67","Type":"ContainerDied","Data":"5cc0c87f16f0f1eb7a9a1dcd16afcdb84d301f455a3ac440e2afa8b17264b9aa"} Dec 01 08:45:35 crc kubenswrapper[4813]: I1201 08:45:35.558924 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wbvcp" Dec 01 08:45:35 crc kubenswrapper[4813]: I1201 08:45:35.568559 4813 scope.go:117] "RemoveContainer" containerID="e33294dfa539d5c4653c0b4bf83d01a72b7e2848567f6ec9f836d922d084cbd4" Dec 01 08:45:35 crc kubenswrapper[4813]: I1201 08:45:35.574143 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qh5hd" event={"ID":"26f4804b-998f-488c-a501-d1cc4ef7f0b4","Type":"ContainerDied","Data":"3b0157e891ec5eb8ab47e9c1ce21457c24e98cfae8c5c64214cfc597f5f1b146"} Dec 01 08:45:35 crc kubenswrapper[4813]: I1201 08:45:35.574262 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-qh5hd" Dec 01 08:45:35 crc kubenswrapper[4813]: I1201 08:45:35.581067 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lk5db" event={"ID":"4f1619ff-e7ed-42ee-b7a9-484c18d6677d","Type":"ContainerDied","Data":"d1fd115940029c23e9d5fd263b2f54473fa1d8fa42e3f02fb128dc0def171258"} Dec 01 08:45:35 crc kubenswrapper[4813]: I1201 08:45:35.581194 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-lk5db" Dec 01 08:45:35 crc kubenswrapper[4813]: I1201 08:45:35.586858 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-nmwnn" event={"ID":"8d57deaa-d1bd-421f-84c3-8682472f1491","Type":"ContainerStarted","Data":"d48f57b22e10aa6c72985aaddd0025105423e82a8cf1b9603505ac1515798430"} Dec 01 08:45:35 crc kubenswrapper[4813]: I1201 08:45:35.586912 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-nmwnn" event={"ID":"8d57deaa-d1bd-421f-84c3-8682472f1491","Type":"ContainerStarted","Data":"f62d67cbbba6e1c4f3e5f6ccec1f35d8a5fcc6ca903d4a5669048b407f583e91"} Dec 01 08:45:35 crc kubenswrapper[4813]: I1201 08:45:35.587370 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-nmwnn" Dec 01 08:45:35 crc kubenswrapper[4813]: I1201 08:45:35.588547 4813 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-nmwnn container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.62:8080/healthz\": dial tcp 10.217.0.62:8080: connect: connection refused" start-of-body= Dec 01 08:45:35 crc kubenswrapper[4813]: I1201 08:45:35.588613 4813 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-nmwnn" podUID="8d57deaa-d1bd-421f-84c3-8682472f1491" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.62:8080/healthz\": dial tcp 10.217.0.62:8080: connect: connection refused" Dec 01 08:45:35 crc kubenswrapper[4813]: I1201 08:45:35.598879 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-qsskc"] Dec 01 08:45:35 crc kubenswrapper[4813]: I1201 08:45:35.603841 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-qsskc"] Dec 01 08:45:35 crc kubenswrapper[4813]: I1201 08:45:35.606589 4813 scope.go:117] "RemoveContainer" containerID="bd47a660e132f743b3808aa5e85b3bbd8b48162275f7f73041d0564054db8d60" Dec 01 08:45:35 crc kubenswrapper[4813]: I1201 08:45:35.614244 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-7fqxs"] Dec 01 08:45:35 crc kubenswrapper[4813]: I1201 08:45:35.615671 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-7fqxs"] Dec 01 08:45:35 crc kubenswrapper[4813]: I1201 08:45:35.644737 4813 scope.go:117] "RemoveContainer" containerID="9940458ef29891b048dfad33fda694cea93fa1876d4ca455069e938ffa765d3a" Dec 01 08:45:35 crc kubenswrapper[4813]: I1201 08:45:35.673593 4813 scope.go:117] "RemoveContainer" containerID="b449598529a2bf89bc7d1b288f3aef72364ec65d95f4efa1630f2e0298648fcb" Dec 01 08:45:35 crc kubenswrapper[4813]: I1201 08:45:35.673642 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-qh5hd"] Dec 01 08:45:35 crc kubenswrapper[4813]: I1201 08:45:35.673705 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-qh5hd"] Dec 01 08:45:35 crc kubenswrapper[4813]: I1201 08:45:35.680060 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-wbvcp"] Dec 01 08:45:35 crc kubenswrapper[4813]: I1201 08:45:35.685171 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-wbvcp"] Dec 01 08:45:35 crc kubenswrapper[4813]: I1201 08:45:35.685386 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-nmwnn" podStartSLOduration=1.68536917 podStartE2EDuration="1.68536917s" podCreationTimestamp="2025-12-01 08:45:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:45:35.678006165 +0000 UTC m=+320.300927761" watchObservedRunningTime="2025-12-01 08:45:35.68536917 +0000 UTC m=+320.308290756" Dec 01 08:45:35 crc kubenswrapper[4813]: I1201 08:45:35.690056 4813 scope.go:117] "RemoveContainer" containerID="45a64db69ae0dfec8da00fa2f2586f5c5abb6678585119474fe56a698206b546" Dec 01 08:45:35 crc kubenswrapper[4813]: I1201 08:45:35.694748 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-lk5db"] Dec 01 08:45:35 crc kubenswrapper[4813]: I1201 08:45:35.694801 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-lk5db"] Dec 01 08:45:35 crc kubenswrapper[4813]: I1201 08:45:35.712294 4813 scope.go:117] "RemoveContainer" containerID="b59544951bce53f8c68dee91538682a664b7d5825a97dc0b5c017c5156d731cf" Dec 01 08:45:35 crc kubenswrapper[4813]: I1201 08:45:35.724621 4813 scope.go:117] "RemoveContainer" containerID="1acc7744b9f14eab07ba59bc0c898cc8e8c944a526dd3599948ffbbaf19aabfd" Dec 01 08:45:35 crc kubenswrapper[4813]: I1201 08:45:35.735168 4813 scope.go:117] "RemoveContainer" containerID="7650b8743cec1bafe177f09abd09ceb115bff7266e4465d450482e7dab874dd4" Dec 01 08:45:35 crc kubenswrapper[4813]: I1201 08:45:35.748516 4813 scope.go:117] "RemoveContainer" containerID="a85ebdc73ea8bec3e97712277dd9d3aa025ade7c5c066f36896adca48b049aa8" Dec 01 08:45:35 crc kubenswrapper[4813]: I1201 08:45:35.764603 4813 scope.go:117] "RemoveContainer" containerID="9ffc53906faf61c55a06e12e23404613fa25399b22cf92a6835b44ae83f7eb5a" Dec 01 08:45:35 crc kubenswrapper[4813]: I1201 08:45:35.785321 4813 scope.go:117] "RemoveContainer" containerID="e35636586ffe20bdca2438e4633c8f87e36115bddb54a527ea25d024c2d8b53c" Dec 01 08:45:35 crc kubenswrapper[4813]: I1201 08:45:35.806302 4813 scope.go:117] "RemoveContainer" containerID="be2b96d2a5411e24ef73a4a32274a3a7cf486c30a3b857f32b237216a1a4b83f" Dec 01 08:45:36 crc kubenswrapper[4813]: I1201 08:45:36.286492 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-dclft"] Dec 01 08:45:36 crc kubenswrapper[4813]: E1201 08:45:36.287034 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0959ee73-31e7-4d58-b124-3e971d777225" containerName="registry-server" Dec 01 08:45:36 crc kubenswrapper[4813]: I1201 08:45:36.287051 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="0959ee73-31e7-4d58-b124-3e971d777225" containerName="registry-server" Dec 01 08:45:36 crc kubenswrapper[4813]: E1201 08:45:36.287074 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0959ee73-31e7-4d58-b124-3e971d777225" containerName="extract-content" Dec 01 08:45:36 crc kubenswrapper[4813]: I1201 08:45:36.287081 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="0959ee73-31e7-4d58-b124-3e971d777225" containerName="extract-content" Dec 01 08:45:36 crc kubenswrapper[4813]: E1201 08:45:36.287088 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="26f4804b-998f-488c-a501-d1cc4ef7f0b4" containerName="registry-server" Dec 01 08:45:36 crc kubenswrapper[4813]: I1201 08:45:36.287094 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="26f4804b-998f-488c-a501-d1cc4ef7f0b4" containerName="registry-server" Dec 01 08:45:36 crc kubenswrapper[4813]: E1201 08:45:36.287105 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4f1619ff-e7ed-42ee-b7a9-484c18d6677d" containerName="extract-utilities" Dec 01 08:45:36 crc kubenswrapper[4813]: I1201 08:45:36.287110 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="4f1619ff-e7ed-42ee-b7a9-484c18d6677d" containerName="extract-utilities" Dec 01 08:45:36 crc kubenswrapper[4813]: E1201 08:45:36.287118 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="26f4804b-998f-488c-a501-d1cc4ef7f0b4" containerName="extract-content" Dec 01 08:45:36 crc kubenswrapper[4813]: I1201 08:45:36.287124 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="26f4804b-998f-488c-a501-d1cc4ef7f0b4" containerName="extract-content" Dec 01 08:45:36 crc kubenswrapper[4813]: E1201 08:45:36.287132 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="979ea876-4dd3-4d31-9144-7a9a6e914c67" containerName="extract-content" Dec 01 08:45:36 crc kubenswrapper[4813]: I1201 08:45:36.287138 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="979ea876-4dd3-4d31-9144-7a9a6e914c67" containerName="extract-content" Dec 01 08:45:36 crc kubenswrapper[4813]: E1201 08:45:36.287145 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c342090f-f331-484d-b287-929d8303f5d7" containerName="marketplace-operator" Dec 01 08:45:36 crc kubenswrapper[4813]: I1201 08:45:36.287152 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="c342090f-f331-484d-b287-929d8303f5d7" containerName="marketplace-operator" Dec 01 08:45:36 crc kubenswrapper[4813]: E1201 08:45:36.287160 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="979ea876-4dd3-4d31-9144-7a9a6e914c67" containerName="extract-utilities" Dec 01 08:45:36 crc kubenswrapper[4813]: I1201 08:45:36.287166 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="979ea876-4dd3-4d31-9144-7a9a6e914c67" containerName="extract-utilities" Dec 01 08:45:36 crc kubenswrapper[4813]: E1201 08:45:36.287175 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="979ea876-4dd3-4d31-9144-7a9a6e914c67" containerName="registry-server" Dec 01 08:45:36 crc kubenswrapper[4813]: I1201 08:45:36.287181 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="979ea876-4dd3-4d31-9144-7a9a6e914c67" containerName="registry-server" Dec 01 08:45:36 crc kubenswrapper[4813]: E1201 08:45:36.287189 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0959ee73-31e7-4d58-b124-3e971d777225" containerName="extract-utilities" Dec 01 08:45:36 crc kubenswrapper[4813]: I1201 08:45:36.287195 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="0959ee73-31e7-4d58-b124-3e971d777225" containerName="extract-utilities" Dec 01 08:45:36 crc kubenswrapper[4813]: E1201 08:45:36.287203 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="26f4804b-998f-488c-a501-d1cc4ef7f0b4" containerName="extract-utilities" Dec 01 08:45:36 crc kubenswrapper[4813]: I1201 08:45:36.287210 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="26f4804b-998f-488c-a501-d1cc4ef7f0b4" containerName="extract-utilities" Dec 01 08:45:36 crc kubenswrapper[4813]: E1201 08:45:36.287221 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4f1619ff-e7ed-42ee-b7a9-484c18d6677d" containerName="extract-content" Dec 01 08:45:36 crc kubenswrapper[4813]: I1201 08:45:36.287227 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="4f1619ff-e7ed-42ee-b7a9-484c18d6677d" containerName="extract-content" Dec 01 08:45:36 crc kubenswrapper[4813]: E1201 08:45:36.287235 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c342090f-f331-484d-b287-929d8303f5d7" containerName="marketplace-operator" Dec 01 08:45:36 crc kubenswrapper[4813]: I1201 08:45:36.287240 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="c342090f-f331-484d-b287-929d8303f5d7" containerName="marketplace-operator" Dec 01 08:45:36 crc kubenswrapper[4813]: E1201 08:45:36.287257 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4f1619ff-e7ed-42ee-b7a9-484c18d6677d" containerName="registry-server" Dec 01 08:45:36 crc kubenswrapper[4813]: I1201 08:45:36.287266 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="4f1619ff-e7ed-42ee-b7a9-484c18d6677d" containerName="registry-server" Dec 01 08:45:36 crc kubenswrapper[4813]: I1201 08:45:36.287357 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="0959ee73-31e7-4d58-b124-3e971d777225" containerName="registry-server" Dec 01 08:45:36 crc kubenswrapper[4813]: I1201 08:45:36.287368 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="26f4804b-998f-488c-a501-d1cc4ef7f0b4" containerName="registry-server" Dec 01 08:45:36 crc kubenswrapper[4813]: I1201 08:45:36.287379 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="4f1619ff-e7ed-42ee-b7a9-484c18d6677d" containerName="registry-server" Dec 01 08:45:36 crc kubenswrapper[4813]: I1201 08:45:36.287390 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="c342090f-f331-484d-b287-929d8303f5d7" containerName="marketplace-operator" Dec 01 08:45:36 crc kubenswrapper[4813]: I1201 08:45:36.287398 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="c342090f-f331-484d-b287-929d8303f5d7" containerName="marketplace-operator" Dec 01 08:45:36 crc kubenswrapper[4813]: I1201 08:45:36.287405 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="979ea876-4dd3-4d31-9144-7a9a6e914c67" containerName="registry-server" Dec 01 08:45:36 crc kubenswrapper[4813]: I1201 08:45:36.288167 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dclft" Dec 01 08:45:36 crc kubenswrapper[4813]: I1201 08:45:36.291425 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Dec 01 08:45:36 crc kubenswrapper[4813]: I1201 08:45:36.299207 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-dclft"] Dec 01 08:45:36 crc kubenswrapper[4813]: I1201 08:45:36.384142 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qhm2l\" (UniqueName: \"kubernetes.io/projected/5b4cbc77-2764-4d42-aa64-e040598935b7-kube-api-access-qhm2l\") pod \"redhat-marketplace-dclft\" (UID: \"5b4cbc77-2764-4d42-aa64-e040598935b7\") " pod="openshift-marketplace/redhat-marketplace-dclft" Dec 01 08:45:36 crc kubenswrapper[4813]: I1201 08:45:36.384528 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5b4cbc77-2764-4d42-aa64-e040598935b7-catalog-content\") pod \"redhat-marketplace-dclft\" (UID: \"5b4cbc77-2764-4d42-aa64-e040598935b7\") " pod="openshift-marketplace/redhat-marketplace-dclft" Dec 01 08:45:36 crc kubenswrapper[4813]: I1201 08:45:36.384715 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5b4cbc77-2764-4d42-aa64-e040598935b7-utilities\") pod \"redhat-marketplace-dclft\" (UID: \"5b4cbc77-2764-4d42-aa64-e040598935b7\") " pod="openshift-marketplace/redhat-marketplace-dclft" Dec 01 08:45:36 crc kubenswrapper[4813]: I1201 08:45:36.404339 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0959ee73-31e7-4d58-b124-3e971d777225" path="/var/lib/kubelet/pods/0959ee73-31e7-4d58-b124-3e971d777225/volumes" Dec 01 08:45:36 crc kubenswrapper[4813]: I1201 08:45:36.404946 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="26f4804b-998f-488c-a501-d1cc4ef7f0b4" path="/var/lib/kubelet/pods/26f4804b-998f-488c-a501-d1cc4ef7f0b4/volumes" Dec 01 08:45:36 crc kubenswrapper[4813]: I1201 08:45:36.405503 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4f1619ff-e7ed-42ee-b7a9-484c18d6677d" path="/var/lib/kubelet/pods/4f1619ff-e7ed-42ee-b7a9-484c18d6677d/volumes" Dec 01 08:45:36 crc kubenswrapper[4813]: I1201 08:45:36.406629 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="979ea876-4dd3-4d31-9144-7a9a6e914c67" path="/var/lib/kubelet/pods/979ea876-4dd3-4d31-9144-7a9a6e914c67/volumes" Dec 01 08:45:36 crc kubenswrapper[4813]: I1201 08:45:36.407241 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c342090f-f331-484d-b287-929d8303f5d7" path="/var/lib/kubelet/pods/c342090f-f331-484d-b287-929d8303f5d7/volumes" Dec 01 08:45:36 crc kubenswrapper[4813]: I1201 08:45:36.485955 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5b4cbc77-2764-4d42-aa64-e040598935b7-catalog-content\") pod \"redhat-marketplace-dclft\" (UID: \"5b4cbc77-2764-4d42-aa64-e040598935b7\") " pod="openshift-marketplace/redhat-marketplace-dclft" Dec 01 08:45:36 crc kubenswrapper[4813]: I1201 08:45:36.486086 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5b4cbc77-2764-4d42-aa64-e040598935b7-utilities\") pod \"redhat-marketplace-dclft\" (UID: \"5b4cbc77-2764-4d42-aa64-e040598935b7\") " pod="openshift-marketplace/redhat-marketplace-dclft" Dec 01 08:45:36 crc kubenswrapper[4813]: I1201 08:45:36.486123 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qhm2l\" (UniqueName: \"kubernetes.io/projected/5b4cbc77-2764-4d42-aa64-e040598935b7-kube-api-access-qhm2l\") pod \"redhat-marketplace-dclft\" (UID: \"5b4cbc77-2764-4d42-aa64-e040598935b7\") " pod="openshift-marketplace/redhat-marketplace-dclft" Dec 01 08:45:36 crc kubenswrapper[4813]: I1201 08:45:36.487532 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5b4cbc77-2764-4d42-aa64-e040598935b7-utilities\") pod \"redhat-marketplace-dclft\" (UID: \"5b4cbc77-2764-4d42-aa64-e040598935b7\") " pod="openshift-marketplace/redhat-marketplace-dclft" Dec 01 08:45:36 crc kubenswrapper[4813]: I1201 08:45:36.487669 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5b4cbc77-2764-4d42-aa64-e040598935b7-catalog-content\") pod \"redhat-marketplace-dclft\" (UID: \"5b4cbc77-2764-4d42-aa64-e040598935b7\") " pod="openshift-marketplace/redhat-marketplace-dclft" Dec 01 08:45:36 crc kubenswrapper[4813]: I1201 08:45:36.514613 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qhm2l\" (UniqueName: \"kubernetes.io/projected/5b4cbc77-2764-4d42-aa64-e040598935b7-kube-api-access-qhm2l\") pod \"redhat-marketplace-dclft\" (UID: \"5b4cbc77-2764-4d42-aa64-e040598935b7\") " pod="openshift-marketplace/redhat-marketplace-dclft" Dec 01 08:45:36 crc kubenswrapper[4813]: I1201 08:45:36.602424 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-nmwnn" Dec 01 08:45:36 crc kubenswrapper[4813]: I1201 08:45:36.608397 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dclft" Dec 01 08:45:37 crc kubenswrapper[4813]: I1201 08:45:37.011012 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-dclft"] Dec 01 08:45:37 crc kubenswrapper[4813]: W1201 08:45:37.020383 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5b4cbc77_2764_4d42_aa64_e040598935b7.slice/crio-d3c1d77930c9406b51a2f61b0095c2ac0154f63cf53d156caba2b9dfc726d820 WatchSource:0}: Error finding container d3c1d77930c9406b51a2f61b0095c2ac0154f63cf53d156caba2b9dfc726d820: Status 404 returned error can't find the container with id d3c1d77930c9406b51a2f61b0095c2ac0154f63cf53d156caba2b9dfc726d820 Dec 01 08:45:37 crc kubenswrapper[4813]: I1201 08:45:37.607434 4813 generic.go:334] "Generic (PLEG): container finished" podID="5b4cbc77-2764-4d42-aa64-e040598935b7" containerID="959056fa00da304d5df5fee59bcded691e0d41b63642e67d510f41fa96fb5376" exitCode=0 Dec 01 08:45:37 crc kubenswrapper[4813]: I1201 08:45:37.607558 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dclft" event={"ID":"5b4cbc77-2764-4d42-aa64-e040598935b7","Type":"ContainerDied","Data":"959056fa00da304d5df5fee59bcded691e0d41b63642e67d510f41fa96fb5376"} Dec 01 08:45:37 crc kubenswrapper[4813]: I1201 08:45:37.607790 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dclft" event={"ID":"5b4cbc77-2764-4d42-aa64-e040598935b7","Type":"ContainerStarted","Data":"d3c1d77930c9406b51a2f61b0095c2ac0154f63cf53d156caba2b9dfc726d820"} Dec 01 08:45:38 crc kubenswrapper[4813]: I1201 08:45:38.059955 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-h5ffj"] Dec 01 08:45:38 crc kubenswrapper[4813]: I1201 08:45:38.063110 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-h5ffj" Dec 01 08:45:38 crc kubenswrapper[4813]: I1201 08:45:38.065834 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Dec 01 08:45:38 crc kubenswrapper[4813]: I1201 08:45:38.066904 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-h5ffj"] Dec 01 08:45:38 crc kubenswrapper[4813]: I1201 08:45:38.205608 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9f704b3a-26b9-43b8-8598-02e769dfbb87-catalog-content\") pod \"redhat-operators-h5ffj\" (UID: \"9f704b3a-26b9-43b8-8598-02e769dfbb87\") " pod="openshift-marketplace/redhat-operators-h5ffj" Dec 01 08:45:38 crc kubenswrapper[4813]: I1201 08:45:38.205733 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9f704b3a-26b9-43b8-8598-02e769dfbb87-utilities\") pod \"redhat-operators-h5ffj\" (UID: \"9f704b3a-26b9-43b8-8598-02e769dfbb87\") " pod="openshift-marketplace/redhat-operators-h5ffj" Dec 01 08:45:38 crc kubenswrapper[4813]: I1201 08:45:38.205757 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lsngh\" (UniqueName: \"kubernetes.io/projected/9f704b3a-26b9-43b8-8598-02e769dfbb87-kube-api-access-lsngh\") pod \"redhat-operators-h5ffj\" (UID: \"9f704b3a-26b9-43b8-8598-02e769dfbb87\") " pod="openshift-marketplace/redhat-operators-h5ffj" Dec 01 08:45:38 crc kubenswrapper[4813]: I1201 08:45:38.306724 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9f704b3a-26b9-43b8-8598-02e769dfbb87-catalog-content\") pod \"redhat-operators-h5ffj\" (UID: \"9f704b3a-26b9-43b8-8598-02e769dfbb87\") " pod="openshift-marketplace/redhat-operators-h5ffj" Dec 01 08:45:38 crc kubenswrapper[4813]: I1201 08:45:38.306850 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9f704b3a-26b9-43b8-8598-02e769dfbb87-utilities\") pod \"redhat-operators-h5ffj\" (UID: \"9f704b3a-26b9-43b8-8598-02e769dfbb87\") " pod="openshift-marketplace/redhat-operators-h5ffj" Dec 01 08:45:38 crc kubenswrapper[4813]: I1201 08:45:38.306894 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lsngh\" (UniqueName: \"kubernetes.io/projected/9f704b3a-26b9-43b8-8598-02e769dfbb87-kube-api-access-lsngh\") pod \"redhat-operators-h5ffj\" (UID: \"9f704b3a-26b9-43b8-8598-02e769dfbb87\") " pod="openshift-marketplace/redhat-operators-h5ffj" Dec 01 08:45:38 crc kubenswrapper[4813]: I1201 08:45:38.307582 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9f704b3a-26b9-43b8-8598-02e769dfbb87-catalog-content\") pod \"redhat-operators-h5ffj\" (UID: \"9f704b3a-26b9-43b8-8598-02e769dfbb87\") " pod="openshift-marketplace/redhat-operators-h5ffj" Dec 01 08:45:38 crc kubenswrapper[4813]: I1201 08:45:38.307678 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9f704b3a-26b9-43b8-8598-02e769dfbb87-utilities\") pod \"redhat-operators-h5ffj\" (UID: \"9f704b3a-26b9-43b8-8598-02e769dfbb87\") " pod="openshift-marketplace/redhat-operators-h5ffj" Dec 01 08:45:38 crc kubenswrapper[4813]: I1201 08:45:38.330996 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lsngh\" (UniqueName: \"kubernetes.io/projected/9f704b3a-26b9-43b8-8598-02e769dfbb87-kube-api-access-lsngh\") pod \"redhat-operators-h5ffj\" (UID: \"9f704b3a-26b9-43b8-8598-02e769dfbb87\") " pod="openshift-marketplace/redhat-operators-h5ffj" Dec 01 08:45:38 crc kubenswrapper[4813]: I1201 08:45:38.384734 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-h5ffj" Dec 01 08:45:38 crc kubenswrapper[4813]: I1201 08:45:38.617189 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dclft" event={"ID":"5b4cbc77-2764-4d42-aa64-e040598935b7","Type":"ContainerStarted","Data":"014b557957795811f82b87d6462a27464ab9478de345ff502b7749aee246193f"} Dec 01 08:45:38 crc kubenswrapper[4813]: I1201 08:45:38.662717 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-f87fz"] Dec 01 08:45:38 crc kubenswrapper[4813]: I1201 08:45:38.664243 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-f87fz" Dec 01 08:45:38 crc kubenswrapper[4813]: I1201 08:45:38.666694 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Dec 01 08:45:38 crc kubenswrapper[4813]: I1201 08:45:38.671044 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-f87fz"] Dec 01 08:45:38 crc kubenswrapper[4813]: I1201 08:45:38.817996 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-h5ffj"] Dec 01 08:45:38 crc kubenswrapper[4813]: I1201 08:45:38.823055 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f6mts\" (UniqueName: \"kubernetes.io/projected/95c21bda-6191-4d26-a3d4-57056be08207-kube-api-access-f6mts\") pod \"community-operators-f87fz\" (UID: \"95c21bda-6191-4d26-a3d4-57056be08207\") " pod="openshift-marketplace/community-operators-f87fz" Dec 01 08:45:38 crc kubenswrapper[4813]: I1201 08:45:38.823121 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/95c21bda-6191-4d26-a3d4-57056be08207-utilities\") pod \"community-operators-f87fz\" (UID: \"95c21bda-6191-4d26-a3d4-57056be08207\") " pod="openshift-marketplace/community-operators-f87fz" Dec 01 08:45:38 crc kubenswrapper[4813]: I1201 08:45:38.823351 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/95c21bda-6191-4d26-a3d4-57056be08207-catalog-content\") pod \"community-operators-f87fz\" (UID: \"95c21bda-6191-4d26-a3d4-57056be08207\") " pod="openshift-marketplace/community-operators-f87fz" Dec 01 08:45:38 crc kubenswrapper[4813]: W1201 08:45:38.825756 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9f704b3a_26b9_43b8_8598_02e769dfbb87.slice/crio-a65f3368feb2d231fbf75ea411c61a9435c1c7f84b60a21606f18fa147556809 WatchSource:0}: Error finding container a65f3368feb2d231fbf75ea411c61a9435c1c7f84b60a21606f18fa147556809: Status 404 returned error can't find the container with id a65f3368feb2d231fbf75ea411c61a9435c1c7f84b60a21606f18fa147556809 Dec 01 08:45:38 crc kubenswrapper[4813]: I1201 08:45:38.924387 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/95c21bda-6191-4d26-a3d4-57056be08207-catalog-content\") pod \"community-operators-f87fz\" (UID: \"95c21bda-6191-4d26-a3d4-57056be08207\") " pod="openshift-marketplace/community-operators-f87fz" Dec 01 08:45:38 crc kubenswrapper[4813]: I1201 08:45:38.924730 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f6mts\" (UniqueName: \"kubernetes.io/projected/95c21bda-6191-4d26-a3d4-57056be08207-kube-api-access-f6mts\") pod \"community-operators-f87fz\" (UID: \"95c21bda-6191-4d26-a3d4-57056be08207\") " pod="openshift-marketplace/community-operators-f87fz" Dec 01 08:45:38 crc kubenswrapper[4813]: I1201 08:45:38.924769 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/95c21bda-6191-4d26-a3d4-57056be08207-utilities\") pod \"community-operators-f87fz\" (UID: \"95c21bda-6191-4d26-a3d4-57056be08207\") " pod="openshift-marketplace/community-operators-f87fz" Dec 01 08:45:38 crc kubenswrapper[4813]: I1201 08:45:38.924872 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/95c21bda-6191-4d26-a3d4-57056be08207-catalog-content\") pod \"community-operators-f87fz\" (UID: \"95c21bda-6191-4d26-a3d4-57056be08207\") " pod="openshift-marketplace/community-operators-f87fz" Dec 01 08:45:38 crc kubenswrapper[4813]: I1201 08:45:38.925221 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/95c21bda-6191-4d26-a3d4-57056be08207-utilities\") pod \"community-operators-f87fz\" (UID: \"95c21bda-6191-4d26-a3d4-57056be08207\") " pod="openshift-marketplace/community-operators-f87fz" Dec 01 08:45:38 crc kubenswrapper[4813]: I1201 08:45:38.952917 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-lk8b2"] Dec 01 08:45:38 crc kubenswrapper[4813]: I1201 08:45:38.953812 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-lk8b2" Dec 01 08:45:38 crc kubenswrapper[4813]: I1201 08:45:38.959518 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f6mts\" (UniqueName: \"kubernetes.io/projected/95c21bda-6191-4d26-a3d4-57056be08207-kube-api-access-f6mts\") pod \"community-operators-f87fz\" (UID: \"95c21bda-6191-4d26-a3d4-57056be08207\") " pod="openshift-marketplace/community-operators-f87fz" Dec 01 08:45:38 crc kubenswrapper[4813]: I1201 08:45:38.971469 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-lk8b2"] Dec 01 08:45:38 crc kubenswrapper[4813]: I1201 08:45:38.984033 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-f87fz" Dec 01 08:45:39 crc kubenswrapper[4813]: I1201 08:45:39.127933 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c765308d-b0bd-42fb-8695-83eeecc294b1-trusted-ca\") pod \"image-registry-66df7c8f76-lk8b2\" (UID: \"c765308d-b0bd-42fb-8695-83eeecc294b1\") " pod="openshift-image-registry/image-registry-66df7c8f76-lk8b2" Dec 01 08:45:39 crc kubenswrapper[4813]: I1201 08:45:39.128252 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/c765308d-b0bd-42fb-8695-83eeecc294b1-registry-tls\") pod \"image-registry-66df7c8f76-lk8b2\" (UID: \"c765308d-b0bd-42fb-8695-83eeecc294b1\") " pod="openshift-image-registry/image-registry-66df7c8f76-lk8b2" Dec 01 08:45:39 crc kubenswrapper[4813]: I1201 08:45:39.128285 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-lk8b2\" (UID: \"c765308d-b0bd-42fb-8695-83eeecc294b1\") " pod="openshift-image-registry/image-registry-66df7c8f76-lk8b2" Dec 01 08:45:39 crc kubenswrapper[4813]: I1201 08:45:39.128307 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/c765308d-b0bd-42fb-8695-83eeecc294b1-ca-trust-extracted\") pod \"image-registry-66df7c8f76-lk8b2\" (UID: \"c765308d-b0bd-42fb-8695-83eeecc294b1\") " pod="openshift-image-registry/image-registry-66df7c8f76-lk8b2" Dec 01 08:45:39 crc kubenswrapper[4813]: I1201 08:45:39.128460 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l5zvp\" (UniqueName: \"kubernetes.io/projected/c765308d-b0bd-42fb-8695-83eeecc294b1-kube-api-access-l5zvp\") pod \"image-registry-66df7c8f76-lk8b2\" (UID: \"c765308d-b0bd-42fb-8695-83eeecc294b1\") " pod="openshift-image-registry/image-registry-66df7c8f76-lk8b2" Dec 01 08:45:39 crc kubenswrapper[4813]: I1201 08:45:39.128515 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/c765308d-b0bd-42fb-8695-83eeecc294b1-registry-certificates\") pod \"image-registry-66df7c8f76-lk8b2\" (UID: \"c765308d-b0bd-42fb-8695-83eeecc294b1\") " pod="openshift-image-registry/image-registry-66df7c8f76-lk8b2" Dec 01 08:45:39 crc kubenswrapper[4813]: I1201 08:45:39.128643 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/c765308d-b0bd-42fb-8695-83eeecc294b1-bound-sa-token\") pod \"image-registry-66df7c8f76-lk8b2\" (UID: \"c765308d-b0bd-42fb-8695-83eeecc294b1\") " pod="openshift-image-registry/image-registry-66df7c8f76-lk8b2" Dec 01 08:45:39 crc kubenswrapper[4813]: I1201 08:45:39.128666 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/c765308d-b0bd-42fb-8695-83eeecc294b1-installation-pull-secrets\") pod \"image-registry-66df7c8f76-lk8b2\" (UID: \"c765308d-b0bd-42fb-8695-83eeecc294b1\") " pod="openshift-image-registry/image-registry-66df7c8f76-lk8b2" Dec 01 08:45:39 crc kubenswrapper[4813]: I1201 08:45:39.154698 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-lk8b2\" (UID: \"c765308d-b0bd-42fb-8695-83eeecc294b1\") " pod="openshift-image-registry/image-registry-66df7c8f76-lk8b2" Dec 01 08:45:39 crc kubenswrapper[4813]: I1201 08:45:39.229431 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/c765308d-b0bd-42fb-8695-83eeecc294b1-registry-certificates\") pod \"image-registry-66df7c8f76-lk8b2\" (UID: \"c765308d-b0bd-42fb-8695-83eeecc294b1\") " pod="openshift-image-registry/image-registry-66df7c8f76-lk8b2" Dec 01 08:45:39 crc kubenswrapper[4813]: I1201 08:45:39.229507 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/c765308d-b0bd-42fb-8695-83eeecc294b1-bound-sa-token\") pod \"image-registry-66df7c8f76-lk8b2\" (UID: \"c765308d-b0bd-42fb-8695-83eeecc294b1\") " pod="openshift-image-registry/image-registry-66df7c8f76-lk8b2" Dec 01 08:45:39 crc kubenswrapper[4813]: I1201 08:45:39.229528 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/c765308d-b0bd-42fb-8695-83eeecc294b1-installation-pull-secrets\") pod \"image-registry-66df7c8f76-lk8b2\" (UID: \"c765308d-b0bd-42fb-8695-83eeecc294b1\") " pod="openshift-image-registry/image-registry-66df7c8f76-lk8b2" Dec 01 08:45:39 crc kubenswrapper[4813]: I1201 08:45:39.229577 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c765308d-b0bd-42fb-8695-83eeecc294b1-trusted-ca\") pod \"image-registry-66df7c8f76-lk8b2\" (UID: \"c765308d-b0bd-42fb-8695-83eeecc294b1\") " pod="openshift-image-registry/image-registry-66df7c8f76-lk8b2" Dec 01 08:45:39 crc kubenswrapper[4813]: I1201 08:45:39.229600 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/c765308d-b0bd-42fb-8695-83eeecc294b1-registry-tls\") pod \"image-registry-66df7c8f76-lk8b2\" (UID: \"c765308d-b0bd-42fb-8695-83eeecc294b1\") " pod="openshift-image-registry/image-registry-66df7c8f76-lk8b2" Dec 01 08:45:39 crc kubenswrapper[4813]: I1201 08:45:39.229622 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/c765308d-b0bd-42fb-8695-83eeecc294b1-ca-trust-extracted\") pod \"image-registry-66df7c8f76-lk8b2\" (UID: \"c765308d-b0bd-42fb-8695-83eeecc294b1\") " pod="openshift-image-registry/image-registry-66df7c8f76-lk8b2" Dec 01 08:45:39 crc kubenswrapper[4813]: I1201 08:45:39.229666 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l5zvp\" (UniqueName: \"kubernetes.io/projected/c765308d-b0bd-42fb-8695-83eeecc294b1-kube-api-access-l5zvp\") pod \"image-registry-66df7c8f76-lk8b2\" (UID: \"c765308d-b0bd-42fb-8695-83eeecc294b1\") " pod="openshift-image-registry/image-registry-66df7c8f76-lk8b2" Dec 01 08:45:39 crc kubenswrapper[4813]: I1201 08:45:39.230625 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/c765308d-b0bd-42fb-8695-83eeecc294b1-ca-trust-extracted\") pod \"image-registry-66df7c8f76-lk8b2\" (UID: \"c765308d-b0bd-42fb-8695-83eeecc294b1\") " pod="openshift-image-registry/image-registry-66df7c8f76-lk8b2" Dec 01 08:45:39 crc kubenswrapper[4813]: I1201 08:45:39.230832 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/c765308d-b0bd-42fb-8695-83eeecc294b1-registry-certificates\") pod \"image-registry-66df7c8f76-lk8b2\" (UID: \"c765308d-b0bd-42fb-8695-83eeecc294b1\") " pod="openshift-image-registry/image-registry-66df7c8f76-lk8b2" Dec 01 08:45:39 crc kubenswrapper[4813]: I1201 08:45:39.230924 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c765308d-b0bd-42fb-8695-83eeecc294b1-trusted-ca\") pod \"image-registry-66df7c8f76-lk8b2\" (UID: \"c765308d-b0bd-42fb-8695-83eeecc294b1\") " pod="openshift-image-registry/image-registry-66df7c8f76-lk8b2" Dec 01 08:45:39 crc kubenswrapper[4813]: I1201 08:45:39.233694 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/c765308d-b0bd-42fb-8695-83eeecc294b1-installation-pull-secrets\") pod \"image-registry-66df7c8f76-lk8b2\" (UID: \"c765308d-b0bd-42fb-8695-83eeecc294b1\") " pod="openshift-image-registry/image-registry-66df7c8f76-lk8b2" Dec 01 08:45:39 crc kubenswrapper[4813]: I1201 08:45:39.235426 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/c765308d-b0bd-42fb-8695-83eeecc294b1-registry-tls\") pod \"image-registry-66df7c8f76-lk8b2\" (UID: \"c765308d-b0bd-42fb-8695-83eeecc294b1\") " pod="openshift-image-registry/image-registry-66df7c8f76-lk8b2" Dec 01 08:45:39 crc kubenswrapper[4813]: I1201 08:45:39.243636 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/c765308d-b0bd-42fb-8695-83eeecc294b1-bound-sa-token\") pod \"image-registry-66df7c8f76-lk8b2\" (UID: \"c765308d-b0bd-42fb-8695-83eeecc294b1\") " pod="openshift-image-registry/image-registry-66df7c8f76-lk8b2" Dec 01 08:45:39 crc kubenswrapper[4813]: I1201 08:45:39.250237 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l5zvp\" (UniqueName: \"kubernetes.io/projected/c765308d-b0bd-42fb-8695-83eeecc294b1-kube-api-access-l5zvp\") pod \"image-registry-66df7c8f76-lk8b2\" (UID: \"c765308d-b0bd-42fb-8695-83eeecc294b1\") " pod="openshift-image-registry/image-registry-66df7c8f76-lk8b2" Dec 01 08:45:39 crc kubenswrapper[4813]: I1201 08:45:39.269279 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-lk8b2" Dec 01 08:45:39 crc kubenswrapper[4813]: I1201 08:45:39.386988 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-f87fz"] Dec 01 08:45:39 crc kubenswrapper[4813]: W1201 08:45:39.396762 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod95c21bda_6191_4d26_a3d4_57056be08207.slice/crio-20d479d603ed9fb9fdc455947e824c53f27d241f036092d4a8d1d84ee424b08b WatchSource:0}: Error finding container 20d479d603ed9fb9fdc455947e824c53f27d241f036092d4a8d1d84ee424b08b: Status 404 returned error can't find the container with id 20d479d603ed9fb9fdc455947e824c53f27d241f036092d4a8d1d84ee424b08b Dec 01 08:45:39 crc kubenswrapper[4813]: I1201 08:45:39.626023 4813 generic.go:334] "Generic (PLEG): container finished" podID="5b4cbc77-2764-4d42-aa64-e040598935b7" containerID="014b557957795811f82b87d6462a27464ab9478de345ff502b7749aee246193f" exitCode=0 Dec 01 08:45:39 crc kubenswrapper[4813]: I1201 08:45:39.626099 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dclft" event={"ID":"5b4cbc77-2764-4d42-aa64-e040598935b7","Type":"ContainerDied","Data":"014b557957795811f82b87d6462a27464ab9478de345ff502b7749aee246193f"} Dec 01 08:45:39 crc kubenswrapper[4813]: I1201 08:45:39.627934 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-f87fz" event={"ID":"95c21bda-6191-4d26-a3d4-57056be08207","Type":"ContainerStarted","Data":"47b6be9ac5ff9b0460c4ab1405750211c203b0922ca5a5cfee383a703eea7523"} Dec 01 08:45:39 crc kubenswrapper[4813]: I1201 08:45:39.627977 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-f87fz" event={"ID":"95c21bda-6191-4d26-a3d4-57056be08207","Type":"ContainerStarted","Data":"20d479d603ed9fb9fdc455947e824c53f27d241f036092d4a8d1d84ee424b08b"} Dec 01 08:45:39 crc kubenswrapper[4813]: I1201 08:45:39.630146 4813 generic.go:334] "Generic (PLEG): container finished" podID="9f704b3a-26b9-43b8-8598-02e769dfbb87" containerID="ac98a8bcbe71ad71e6de09807da20370dc041a9c7bcb9b7f72ea9516196a52d2" exitCode=0 Dec 01 08:45:39 crc kubenswrapper[4813]: I1201 08:45:39.630182 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-h5ffj" event={"ID":"9f704b3a-26b9-43b8-8598-02e769dfbb87","Type":"ContainerDied","Data":"ac98a8bcbe71ad71e6de09807da20370dc041a9c7bcb9b7f72ea9516196a52d2"} Dec 01 08:45:39 crc kubenswrapper[4813]: I1201 08:45:39.630204 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-h5ffj" event={"ID":"9f704b3a-26b9-43b8-8598-02e769dfbb87","Type":"ContainerStarted","Data":"a65f3368feb2d231fbf75ea411c61a9435c1c7f84b60a21606f18fa147556809"} Dec 01 08:45:39 crc kubenswrapper[4813]: I1201 08:45:39.702950 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-lk8b2"] Dec 01 08:45:40 crc kubenswrapper[4813]: I1201 08:45:40.470371 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-gwz7r"] Dec 01 08:45:40 crc kubenswrapper[4813]: I1201 08:45:40.474235 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-gwz7r" Dec 01 08:45:40 crc kubenswrapper[4813]: I1201 08:45:40.476628 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Dec 01 08:45:40 crc kubenswrapper[4813]: I1201 08:45:40.483076 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-gwz7r"] Dec 01 08:45:40 crc kubenswrapper[4813]: I1201 08:45:40.636638 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-h5ffj" event={"ID":"9f704b3a-26b9-43b8-8598-02e769dfbb87","Type":"ContainerStarted","Data":"55ef8d1a929a45b48902af848bd03d58395942b2c23b28297f9c34cd9b21ce7b"} Dec 01 08:45:40 crc kubenswrapper[4813]: I1201 08:45:40.638576 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-lk8b2" event={"ID":"c765308d-b0bd-42fb-8695-83eeecc294b1","Type":"ContainerStarted","Data":"0c56bbf7fc3e8f0a286c6387cd2f3bc486900a475888fd35017fe1d4b603be3a"} Dec 01 08:45:40 crc kubenswrapper[4813]: I1201 08:45:40.638605 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-lk8b2" event={"ID":"c765308d-b0bd-42fb-8695-83eeecc294b1","Type":"ContainerStarted","Data":"a0038b5cd4c090128be2ab1cd143b1a964946211b7a8905f0df4ac03b16a7aa3"} Dec 01 08:45:40 crc kubenswrapper[4813]: I1201 08:45:40.638718 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-lk8b2" Dec 01 08:45:40 crc kubenswrapper[4813]: I1201 08:45:40.641056 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dclft" event={"ID":"5b4cbc77-2764-4d42-aa64-e040598935b7","Type":"ContainerStarted","Data":"fa0aaef39ef1013bc8262ce5712525181601812bb15c2382f41c305da7fa6e50"} Dec 01 08:45:40 crc kubenswrapper[4813]: I1201 08:45:40.642553 4813 generic.go:334] "Generic (PLEG): container finished" podID="95c21bda-6191-4d26-a3d4-57056be08207" containerID="47b6be9ac5ff9b0460c4ab1405750211c203b0922ca5a5cfee383a703eea7523" exitCode=0 Dec 01 08:45:40 crc kubenswrapper[4813]: I1201 08:45:40.642595 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-f87fz" event={"ID":"95c21bda-6191-4d26-a3d4-57056be08207","Type":"ContainerDied","Data":"47b6be9ac5ff9b0460c4ab1405750211c203b0922ca5a5cfee383a703eea7523"} Dec 01 08:45:40 crc kubenswrapper[4813]: I1201 08:45:40.642616 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-f87fz" event={"ID":"95c21bda-6191-4d26-a3d4-57056be08207","Type":"ContainerStarted","Data":"f378fc2307de94025e65d2434ed34141e308e192cc1eeba1d4f2b75609300d13"} Dec 01 08:45:40 crc kubenswrapper[4813]: I1201 08:45:40.652849 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/03720b75-74be-44af-85fd-6b5a55578ca6-utilities\") pod \"certified-operators-gwz7r\" (UID: \"03720b75-74be-44af-85fd-6b5a55578ca6\") " pod="openshift-marketplace/certified-operators-gwz7r" Dec 01 08:45:40 crc kubenswrapper[4813]: I1201 08:45:40.652893 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zpwgn\" (UniqueName: \"kubernetes.io/projected/03720b75-74be-44af-85fd-6b5a55578ca6-kube-api-access-zpwgn\") pod \"certified-operators-gwz7r\" (UID: \"03720b75-74be-44af-85fd-6b5a55578ca6\") " pod="openshift-marketplace/certified-operators-gwz7r" Dec 01 08:45:40 crc kubenswrapper[4813]: I1201 08:45:40.652950 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/03720b75-74be-44af-85fd-6b5a55578ca6-catalog-content\") pod \"certified-operators-gwz7r\" (UID: \"03720b75-74be-44af-85fd-6b5a55578ca6\") " pod="openshift-marketplace/certified-operators-gwz7r" Dec 01 08:45:40 crc kubenswrapper[4813]: I1201 08:45:40.682113 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-dclft" podStartSLOduration=2.20922005 podStartE2EDuration="4.68209354s" podCreationTimestamp="2025-12-01 08:45:36 +0000 UTC" firstStartedPulling="2025-12-01 08:45:37.610272458 +0000 UTC m=+322.233194044" lastFinishedPulling="2025-12-01 08:45:40.083145948 +0000 UTC m=+324.706067534" observedRunningTime="2025-12-01 08:45:40.679732901 +0000 UTC m=+325.302654487" watchObservedRunningTime="2025-12-01 08:45:40.68209354 +0000 UTC m=+325.305015126" Dec 01 08:45:40 crc kubenswrapper[4813]: I1201 08:45:40.704653 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-lk8b2" podStartSLOduration=2.704631009 podStartE2EDuration="2.704631009s" podCreationTimestamp="2025-12-01 08:45:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:45:40.699523619 +0000 UTC m=+325.322445225" watchObservedRunningTime="2025-12-01 08:45:40.704631009 +0000 UTC m=+325.327552615" Dec 01 08:45:40 crc kubenswrapper[4813]: I1201 08:45:40.754195 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/03720b75-74be-44af-85fd-6b5a55578ca6-utilities\") pod \"certified-operators-gwz7r\" (UID: \"03720b75-74be-44af-85fd-6b5a55578ca6\") " pod="openshift-marketplace/certified-operators-gwz7r" Dec 01 08:45:40 crc kubenswrapper[4813]: I1201 08:45:40.754245 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zpwgn\" (UniqueName: \"kubernetes.io/projected/03720b75-74be-44af-85fd-6b5a55578ca6-kube-api-access-zpwgn\") pod \"certified-operators-gwz7r\" (UID: \"03720b75-74be-44af-85fd-6b5a55578ca6\") " pod="openshift-marketplace/certified-operators-gwz7r" Dec 01 08:45:40 crc kubenswrapper[4813]: I1201 08:45:40.754341 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/03720b75-74be-44af-85fd-6b5a55578ca6-catalog-content\") pod \"certified-operators-gwz7r\" (UID: \"03720b75-74be-44af-85fd-6b5a55578ca6\") " pod="openshift-marketplace/certified-operators-gwz7r" Dec 01 08:45:40 crc kubenswrapper[4813]: I1201 08:45:40.754725 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/03720b75-74be-44af-85fd-6b5a55578ca6-utilities\") pod \"certified-operators-gwz7r\" (UID: \"03720b75-74be-44af-85fd-6b5a55578ca6\") " pod="openshift-marketplace/certified-operators-gwz7r" Dec 01 08:45:40 crc kubenswrapper[4813]: I1201 08:45:40.755024 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/03720b75-74be-44af-85fd-6b5a55578ca6-catalog-content\") pod \"certified-operators-gwz7r\" (UID: \"03720b75-74be-44af-85fd-6b5a55578ca6\") " pod="openshift-marketplace/certified-operators-gwz7r" Dec 01 08:45:40 crc kubenswrapper[4813]: I1201 08:45:40.792885 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zpwgn\" (UniqueName: \"kubernetes.io/projected/03720b75-74be-44af-85fd-6b5a55578ca6-kube-api-access-zpwgn\") pod \"certified-operators-gwz7r\" (UID: \"03720b75-74be-44af-85fd-6b5a55578ca6\") " pod="openshift-marketplace/certified-operators-gwz7r" Dec 01 08:45:40 crc kubenswrapper[4813]: I1201 08:45:40.806752 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-gwz7r" Dec 01 08:45:41 crc kubenswrapper[4813]: I1201 08:45:41.264891 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-gwz7r"] Dec 01 08:45:41 crc kubenswrapper[4813]: I1201 08:45:41.658157 4813 generic.go:334] "Generic (PLEG): container finished" podID="9f704b3a-26b9-43b8-8598-02e769dfbb87" containerID="55ef8d1a929a45b48902af848bd03d58395942b2c23b28297f9c34cd9b21ce7b" exitCode=0 Dec 01 08:45:41 crc kubenswrapper[4813]: I1201 08:45:41.658212 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-h5ffj" event={"ID":"9f704b3a-26b9-43b8-8598-02e769dfbb87","Type":"ContainerDied","Data":"55ef8d1a929a45b48902af848bd03d58395942b2c23b28297f9c34cd9b21ce7b"} Dec 01 08:45:41 crc kubenswrapper[4813]: I1201 08:45:41.660376 4813 generic.go:334] "Generic (PLEG): container finished" podID="03720b75-74be-44af-85fd-6b5a55578ca6" containerID="64ee7a4f4867f782282191fd42431ef32389eed9179c9d18d5975d499d8d5d09" exitCode=0 Dec 01 08:45:41 crc kubenswrapper[4813]: I1201 08:45:41.660474 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gwz7r" event={"ID":"03720b75-74be-44af-85fd-6b5a55578ca6","Type":"ContainerDied","Data":"64ee7a4f4867f782282191fd42431ef32389eed9179c9d18d5975d499d8d5d09"} Dec 01 08:45:41 crc kubenswrapper[4813]: I1201 08:45:41.660674 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gwz7r" event={"ID":"03720b75-74be-44af-85fd-6b5a55578ca6","Type":"ContainerStarted","Data":"21a7a2984d08c31a6ad18dd1f5eec909c75436ef2715243ecde99de5709ff21d"} Dec 01 08:45:41 crc kubenswrapper[4813]: I1201 08:45:41.671246 4813 generic.go:334] "Generic (PLEG): container finished" podID="95c21bda-6191-4d26-a3d4-57056be08207" containerID="f378fc2307de94025e65d2434ed34141e308e192cc1eeba1d4f2b75609300d13" exitCode=0 Dec 01 08:45:41 crc kubenswrapper[4813]: I1201 08:45:41.671795 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-f87fz" event={"ID":"95c21bda-6191-4d26-a3d4-57056be08207","Type":"ContainerDied","Data":"f378fc2307de94025e65d2434ed34141e308e192cc1eeba1d4f2b75609300d13"} Dec 01 08:45:42 crc kubenswrapper[4813]: I1201 08:45:42.679014 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-f87fz" event={"ID":"95c21bda-6191-4d26-a3d4-57056be08207","Type":"ContainerStarted","Data":"34584b5be69b1a718787a943d1f3d98ed6c09574e5da11dd0f2acb99aa018729"} Dec 01 08:45:42 crc kubenswrapper[4813]: I1201 08:45:42.681657 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-h5ffj" event={"ID":"9f704b3a-26b9-43b8-8598-02e769dfbb87","Type":"ContainerStarted","Data":"120663d1fcb5d242f2e9a6d02d93df0d1a8239dac5e9e82da85f55d17db23447"} Dec 01 08:45:42 crc kubenswrapper[4813]: I1201 08:45:42.686336 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gwz7r" event={"ID":"03720b75-74be-44af-85fd-6b5a55578ca6","Type":"ContainerStarted","Data":"91a482b866eca4db7a4963dc9b8da69c28aa82f993a61380ff15950ffc547feb"} Dec 01 08:45:42 crc kubenswrapper[4813]: I1201 08:45:42.711260 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-f87fz" podStartSLOduration=2.21108713 podStartE2EDuration="4.711236296s" podCreationTimestamp="2025-12-01 08:45:38 +0000 UTC" firstStartedPulling="2025-12-01 08:45:39.629122884 +0000 UTC m=+324.252044460" lastFinishedPulling="2025-12-01 08:45:42.12927204 +0000 UTC m=+326.752193626" observedRunningTime="2025-12-01 08:45:42.707254984 +0000 UTC m=+327.330176580" watchObservedRunningTime="2025-12-01 08:45:42.711236296 +0000 UTC m=+327.334157892" Dec 01 08:45:42 crc kubenswrapper[4813]: I1201 08:45:42.728362 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-h5ffj" podStartSLOduration=2.259893313 podStartE2EDuration="4.728342581s" podCreationTimestamp="2025-12-01 08:45:38 +0000 UTC" firstStartedPulling="2025-12-01 08:45:39.630931346 +0000 UTC m=+324.253852932" lastFinishedPulling="2025-12-01 08:45:42.099380584 +0000 UTC m=+326.722302200" observedRunningTime="2025-12-01 08:45:42.727075625 +0000 UTC m=+327.349997221" watchObservedRunningTime="2025-12-01 08:45:42.728342581 +0000 UTC m=+327.351264167" Dec 01 08:45:43 crc kubenswrapper[4813]: I1201 08:45:43.693662 4813 generic.go:334] "Generic (PLEG): container finished" podID="03720b75-74be-44af-85fd-6b5a55578ca6" containerID="91a482b866eca4db7a4963dc9b8da69c28aa82f993a61380ff15950ffc547feb" exitCode=0 Dec 01 08:45:43 crc kubenswrapper[4813]: I1201 08:45:43.693786 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gwz7r" event={"ID":"03720b75-74be-44af-85fd-6b5a55578ca6","Type":"ContainerDied","Data":"91a482b866eca4db7a4963dc9b8da69c28aa82f993a61380ff15950ffc547feb"} Dec 01 08:45:44 crc kubenswrapper[4813]: I1201 08:45:44.705086 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gwz7r" event={"ID":"03720b75-74be-44af-85fd-6b5a55578ca6","Type":"ContainerStarted","Data":"b3801e3a9b3be33eb13b7f21b1d51b65844ebe03683ac2439a63093d69ff79e9"} Dec 01 08:45:44 crc kubenswrapper[4813]: I1201 08:45:44.723696 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-gwz7r" podStartSLOduration=2.010897639 podStartE2EDuration="4.723677199s" podCreationTimestamp="2025-12-01 08:45:40 +0000 UTC" firstStartedPulling="2025-12-01 08:45:41.663042531 +0000 UTC m=+326.285964117" lastFinishedPulling="2025-12-01 08:45:44.375822091 +0000 UTC m=+328.998743677" observedRunningTime="2025-12-01 08:45:44.722553657 +0000 UTC m=+329.345475253" watchObservedRunningTime="2025-12-01 08:45:44.723677199 +0000 UTC m=+329.346598775" Dec 01 08:45:46 crc kubenswrapper[4813]: I1201 08:45:46.609151 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-dclft" Dec 01 08:45:46 crc kubenswrapper[4813]: I1201 08:45:46.609617 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-dclft" Dec 01 08:45:46 crc kubenswrapper[4813]: I1201 08:45:46.681046 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-dclft" Dec 01 08:45:46 crc kubenswrapper[4813]: I1201 08:45:46.766898 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-dclft" Dec 01 08:45:47 crc kubenswrapper[4813]: I1201 08:45:47.221415 4813 patch_prober.go:28] interesting pod/machine-config-daemon-9wn4q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 08:45:47 crc kubenswrapper[4813]: I1201 08:45:47.221834 4813 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9wn4q" podUID="3d423402-e08f-42e9-b60c-6eb9bf2f6d48" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 08:45:48 crc kubenswrapper[4813]: I1201 08:45:48.178124 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-84f5659d8d-6mdwm"] Dec 01 08:45:48 crc kubenswrapper[4813]: I1201 08:45:48.178379 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-84f5659d8d-6mdwm" podUID="93834113-aa88-44f3-bf6b-4a07e78cfef4" containerName="route-controller-manager" containerID="cri-o://869bb1c5e9c619ddce571cbb2f8cd1bac46873d00cbe3e34fc6d7035c1d6a127" gracePeriod=30 Dec 01 08:45:48 crc kubenswrapper[4813]: I1201 08:45:48.385873 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-h5ffj" Dec 01 08:45:48 crc kubenswrapper[4813]: I1201 08:45:48.385953 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-h5ffj" Dec 01 08:45:48 crc kubenswrapper[4813]: I1201 08:45:48.732206 4813 generic.go:334] "Generic (PLEG): container finished" podID="93834113-aa88-44f3-bf6b-4a07e78cfef4" containerID="869bb1c5e9c619ddce571cbb2f8cd1bac46873d00cbe3e34fc6d7035c1d6a127" exitCode=0 Dec 01 08:45:48 crc kubenswrapper[4813]: I1201 08:45:48.732368 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-84f5659d8d-6mdwm" event={"ID":"93834113-aa88-44f3-bf6b-4a07e78cfef4","Type":"ContainerDied","Data":"869bb1c5e9c619ddce571cbb2f8cd1bac46873d00cbe3e34fc6d7035c1d6a127"} Dec 01 08:45:48 crc kubenswrapper[4813]: I1201 08:45:48.985374 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-f87fz" Dec 01 08:45:48 crc kubenswrapper[4813]: I1201 08:45:48.985438 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-f87fz" Dec 01 08:45:49 crc kubenswrapper[4813]: I1201 08:45:49.058370 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-f87fz" Dec 01 08:45:49 crc kubenswrapper[4813]: I1201 08:45:49.186617 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-84f5659d8d-6mdwm" Dec 01 08:45:49 crc kubenswrapper[4813]: I1201 08:45:49.218331 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6c59b6cf4c-lg7vm"] Dec 01 08:45:49 crc kubenswrapper[4813]: E1201 08:45:49.218733 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="93834113-aa88-44f3-bf6b-4a07e78cfef4" containerName="route-controller-manager" Dec 01 08:45:49 crc kubenswrapper[4813]: I1201 08:45:49.218773 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="93834113-aa88-44f3-bf6b-4a07e78cfef4" containerName="route-controller-manager" Dec 01 08:45:49 crc kubenswrapper[4813]: I1201 08:45:49.220364 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="93834113-aa88-44f3-bf6b-4a07e78cfef4" containerName="route-controller-manager" Dec 01 08:45:49 crc kubenswrapper[4813]: I1201 08:45:49.221096 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6c59b6cf4c-lg7vm" Dec 01 08:45:49 crc kubenswrapper[4813]: I1201 08:45:49.247823 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6c59b6cf4c-lg7vm"] Dec 01 08:45:49 crc kubenswrapper[4813]: I1201 08:45:49.261518 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/93834113-aa88-44f3-bf6b-4a07e78cfef4-config\") pod \"93834113-aa88-44f3-bf6b-4a07e78cfef4\" (UID: \"93834113-aa88-44f3-bf6b-4a07e78cfef4\") " Dec 01 08:45:49 crc kubenswrapper[4813]: I1201 08:45:49.261683 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/93834113-aa88-44f3-bf6b-4a07e78cfef4-serving-cert\") pod \"93834113-aa88-44f3-bf6b-4a07e78cfef4\" (UID: \"93834113-aa88-44f3-bf6b-4a07e78cfef4\") " Dec 01 08:45:49 crc kubenswrapper[4813]: I1201 08:45:49.261751 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/93834113-aa88-44f3-bf6b-4a07e78cfef4-client-ca\") pod \"93834113-aa88-44f3-bf6b-4a07e78cfef4\" (UID: \"93834113-aa88-44f3-bf6b-4a07e78cfef4\") " Dec 01 08:45:49 crc kubenswrapper[4813]: I1201 08:45:49.261824 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p5hmz\" (UniqueName: \"kubernetes.io/projected/93834113-aa88-44f3-bf6b-4a07e78cfef4-kube-api-access-p5hmz\") pod \"93834113-aa88-44f3-bf6b-4a07e78cfef4\" (UID: \"93834113-aa88-44f3-bf6b-4a07e78cfef4\") " Dec 01 08:45:49 crc kubenswrapper[4813]: I1201 08:45:49.262487 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/66cd0011-17bd-416d-b7f7-aecd41bc6874-config\") pod \"route-controller-manager-6c59b6cf4c-lg7vm\" (UID: \"66cd0011-17bd-416d-b7f7-aecd41bc6874\") " pod="openshift-route-controller-manager/route-controller-manager-6c59b6cf4c-lg7vm" Dec 01 08:45:49 crc kubenswrapper[4813]: I1201 08:45:49.262607 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/66cd0011-17bd-416d-b7f7-aecd41bc6874-client-ca\") pod \"route-controller-manager-6c59b6cf4c-lg7vm\" (UID: \"66cd0011-17bd-416d-b7f7-aecd41bc6874\") " pod="openshift-route-controller-manager/route-controller-manager-6c59b6cf4c-lg7vm" Dec 01 08:45:49 crc kubenswrapper[4813]: I1201 08:45:49.262669 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/66cd0011-17bd-416d-b7f7-aecd41bc6874-serving-cert\") pod \"route-controller-manager-6c59b6cf4c-lg7vm\" (UID: \"66cd0011-17bd-416d-b7f7-aecd41bc6874\") " pod="openshift-route-controller-manager/route-controller-manager-6c59b6cf4c-lg7vm" Dec 01 08:45:49 crc kubenswrapper[4813]: I1201 08:45:49.262743 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9k9dv\" (UniqueName: \"kubernetes.io/projected/66cd0011-17bd-416d-b7f7-aecd41bc6874-kube-api-access-9k9dv\") pod \"route-controller-manager-6c59b6cf4c-lg7vm\" (UID: \"66cd0011-17bd-416d-b7f7-aecd41bc6874\") " pod="openshift-route-controller-manager/route-controller-manager-6c59b6cf4c-lg7vm" Dec 01 08:45:49 crc kubenswrapper[4813]: I1201 08:45:49.263648 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/93834113-aa88-44f3-bf6b-4a07e78cfef4-config" (OuterVolumeSpecName: "config") pod "93834113-aa88-44f3-bf6b-4a07e78cfef4" (UID: "93834113-aa88-44f3-bf6b-4a07e78cfef4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:45:49 crc kubenswrapper[4813]: I1201 08:45:49.265472 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/93834113-aa88-44f3-bf6b-4a07e78cfef4-client-ca" (OuterVolumeSpecName: "client-ca") pod "93834113-aa88-44f3-bf6b-4a07e78cfef4" (UID: "93834113-aa88-44f3-bf6b-4a07e78cfef4"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:45:49 crc kubenswrapper[4813]: I1201 08:45:49.291399 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/93834113-aa88-44f3-bf6b-4a07e78cfef4-kube-api-access-p5hmz" (OuterVolumeSpecName: "kube-api-access-p5hmz") pod "93834113-aa88-44f3-bf6b-4a07e78cfef4" (UID: "93834113-aa88-44f3-bf6b-4a07e78cfef4"). InnerVolumeSpecName "kube-api-access-p5hmz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:45:49 crc kubenswrapper[4813]: I1201 08:45:49.291450 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/93834113-aa88-44f3-bf6b-4a07e78cfef4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "93834113-aa88-44f3-bf6b-4a07e78cfef4" (UID: "93834113-aa88-44f3-bf6b-4a07e78cfef4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:45:49 crc kubenswrapper[4813]: I1201 08:45:49.364181 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/66cd0011-17bd-416d-b7f7-aecd41bc6874-serving-cert\") pod \"route-controller-manager-6c59b6cf4c-lg7vm\" (UID: \"66cd0011-17bd-416d-b7f7-aecd41bc6874\") " pod="openshift-route-controller-manager/route-controller-manager-6c59b6cf4c-lg7vm" Dec 01 08:45:49 crc kubenswrapper[4813]: I1201 08:45:49.364262 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9k9dv\" (UniqueName: \"kubernetes.io/projected/66cd0011-17bd-416d-b7f7-aecd41bc6874-kube-api-access-9k9dv\") pod \"route-controller-manager-6c59b6cf4c-lg7vm\" (UID: \"66cd0011-17bd-416d-b7f7-aecd41bc6874\") " pod="openshift-route-controller-manager/route-controller-manager-6c59b6cf4c-lg7vm" Dec 01 08:45:49 crc kubenswrapper[4813]: I1201 08:45:49.364313 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/66cd0011-17bd-416d-b7f7-aecd41bc6874-config\") pod \"route-controller-manager-6c59b6cf4c-lg7vm\" (UID: \"66cd0011-17bd-416d-b7f7-aecd41bc6874\") " pod="openshift-route-controller-manager/route-controller-manager-6c59b6cf4c-lg7vm" Dec 01 08:45:49 crc kubenswrapper[4813]: I1201 08:45:49.364375 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/66cd0011-17bd-416d-b7f7-aecd41bc6874-client-ca\") pod \"route-controller-manager-6c59b6cf4c-lg7vm\" (UID: \"66cd0011-17bd-416d-b7f7-aecd41bc6874\") " pod="openshift-route-controller-manager/route-controller-manager-6c59b6cf4c-lg7vm" Dec 01 08:45:49 crc kubenswrapper[4813]: I1201 08:45:49.364430 4813 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/93834113-aa88-44f3-bf6b-4a07e78cfef4-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 01 08:45:49 crc kubenswrapper[4813]: I1201 08:45:49.364449 4813 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/93834113-aa88-44f3-bf6b-4a07e78cfef4-client-ca\") on node \"crc\" DevicePath \"\"" Dec 01 08:45:49 crc kubenswrapper[4813]: I1201 08:45:49.364463 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p5hmz\" (UniqueName: \"kubernetes.io/projected/93834113-aa88-44f3-bf6b-4a07e78cfef4-kube-api-access-p5hmz\") on node \"crc\" DevicePath \"\"" Dec 01 08:45:49 crc kubenswrapper[4813]: I1201 08:45:49.364479 4813 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/93834113-aa88-44f3-bf6b-4a07e78cfef4-config\") on node \"crc\" DevicePath \"\"" Dec 01 08:45:49 crc kubenswrapper[4813]: I1201 08:45:49.365530 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/66cd0011-17bd-416d-b7f7-aecd41bc6874-client-ca\") pod \"route-controller-manager-6c59b6cf4c-lg7vm\" (UID: \"66cd0011-17bd-416d-b7f7-aecd41bc6874\") " pod="openshift-route-controller-manager/route-controller-manager-6c59b6cf4c-lg7vm" Dec 01 08:45:49 crc kubenswrapper[4813]: I1201 08:45:49.369161 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/66cd0011-17bd-416d-b7f7-aecd41bc6874-config\") pod \"route-controller-manager-6c59b6cf4c-lg7vm\" (UID: \"66cd0011-17bd-416d-b7f7-aecd41bc6874\") " pod="openshift-route-controller-manager/route-controller-manager-6c59b6cf4c-lg7vm" Dec 01 08:45:49 crc kubenswrapper[4813]: I1201 08:45:49.369691 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/66cd0011-17bd-416d-b7f7-aecd41bc6874-serving-cert\") pod \"route-controller-manager-6c59b6cf4c-lg7vm\" (UID: \"66cd0011-17bd-416d-b7f7-aecd41bc6874\") " pod="openshift-route-controller-manager/route-controller-manager-6c59b6cf4c-lg7vm" Dec 01 08:45:49 crc kubenswrapper[4813]: I1201 08:45:49.382935 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9k9dv\" (UniqueName: \"kubernetes.io/projected/66cd0011-17bd-416d-b7f7-aecd41bc6874-kube-api-access-9k9dv\") pod \"route-controller-manager-6c59b6cf4c-lg7vm\" (UID: \"66cd0011-17bd-416d-b7f7-aecd41bc6874\") " pod="openshift-route-controller-manager/route-controller-manager-6c59b6cf4c-lg7vm" Dec 01 08:45:49 crc kubenswrapper[4813]: I1201 08:45:49.450995 4813 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-h5ffj" podUID="9f704b3a-26b9-43b8-8598-02e769dfbb87" containerName="registry-server" probeResult="failure" output=< Dec 01 08:45:49 crc kubenswrapper[4813]: timeout: failed to connect service ":50051" within 1s Dec 01 08:45:49 crc kubenswrapper[4813]: > Dec 01 08:45:49 crc kubenswrapper[4813]: I1201 08:45:49.537336 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6c59b6cf4c-lg7vm" Dec 01 08:45:49 crc kubenswrapper[4813]: I1201 08:45:49.741994 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-84f5659d8d-6mdwm" Dec 01 08:45:49 crc kubenswrapper[4813]: I1201 08:45:49.742232 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-84f5659d8d-6mdwm" event={"ID":"93834113-aa88-44f3-bf6b-4a07e78cfef4","Type":"ContainerDied","Data":"279f06cffb3c3fe9475f7b9b41be3562c86149acad1286811f2049a2c0d124b1"} Dec 01 08:45:49 crc kubenswrapper[4813]: I1201 08:45:49.742290 4813 scope.go:117] "RemoveContainer" containerID="869bb1c5e9c619ddce571cbb2f8cd1bac46873d00cbe3e34fc6d7035c1d6a127" Dec 01 08:45:49 crc kubenswrapper[4813]: I1201 08:45:49.786699 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-84f5659d8d-6mdwm"] Dec 01 08:45:49 crc kubenswrapper[4813]: I1201 08:45:49.791872 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-f87fz" Dec 01 08:45:49 crc kubenswrapper[4813]: I1201 08:45:49.798620 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-84f5659d8d-6mdwm"] Dec 01 08:45:49 crc kubenswrapper[4813]: I1201 08:45:49.982701 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6c59b6cf4c-lg7vm"] Dec 01 08:45:49 crc kubenswrapper[4813]: W1201 08:45:49.993883 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod66cd0011_17bd_416d_b7f7_aecd41bc6874.slice/crio-a83f4b721965fcdec3c9922250718c2d3c8f882c3b735f8e6884674f420c7108 WatchSource:0}: Error finding container a83f4b721965fcdec3c9922250718c2d3c8f882c3b735f8e6884674f420c7108: Status 404 returned error can't find the container with id a83f4b721965fcdec3c9922250718c2d3c8f882c3b735f8e6884674f420c7108 Dec 01 08:45:50 crc kubenswrapper[4813]: I1201 08:45:50.403592 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="93834113-aa88-44f3-bf6b-4a07e78cfef4" path="/var/lib/kubelet/pods/93834113-aa88-44f3-bf6b-4a07e78cfef4/volumes" Dec 01 08:45:50 crc kubenswrapper[4813]: I1201 08:45:50.750875 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6c59b6cf4c-lg7vm" event={"ID":"66cd0011-17bd-416d-b7f7-aecd41bc6874","Type":"ContainerStarted","Data":"e40bb3d9c86324397b2794c4d279dc7accad186c2e9b9052738e726b83fe4785"} Dec 01 08:45:50 crc kubenswrapper[4813]: I1201 08:45:50.750915 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6c59b6cf4c-lg7vm" event={"ID":"66cd0011-17bd-416d-b7f7-aecd41bc6874","Type":"ContainerStarted","Data":"a83f4b721965fcdec3c9922250718c2d3c8f882c3b735f8e6884674f420c7108"} Dec 01 08:45:50 crc kubenswrapper[4813]: I1201 08:45:50.751843 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6c59b6cf4c-lg7vm" Dec 01 08:45:50 crc kubenswrapper[4813]: I1201 08:45:50.755859 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6c59b6cf4c-lg7vm" Dec 01 08:45:50 crc kubenswrapper[4813]: I1201 08:45:50.768084 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6c59b6cf4c-lg7vm" podStartSLOduration=2.7680635479999998 podStartE2EDuration="2.768063548s" podCreationTimestamp="2025-12-01 08:45:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:45:50.763953862 +0000 UTC m=+335.386875448" watchObservedRunningTime="2025-12-01 08:45:50.768063548 +0000 UTC m=+335.390985134" Dec 01 08:45:50 crc kubenswrapper[4813]: I1201 08:45:50.807982 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-gwz7r" Dec 01 08:45:50 crc kubenswrapper[4813]: I1201 08:45:50.808053 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-gwz7r" Dec 01 08:45:50 crc kubenswrapper[4813]: I1201 08:45:50.858873 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-gwz7r" Dec 01 08:45:51 crc kubenswrapper[4813]: I1201 08:45:51.796452 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-gwz7r" Dec 01 08:45:58 crc kubenswrapper[4813]: I1201 08:45:58.457809 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-h5ffj" Dec 01 08:45:58 crc kubenswrapper[4813]: I1201 08:45:58.506210 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-h5ffj" Dec 01 08:45:59 crc kubenswrapper[4813]: I1201 08:45:59.281098 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-lk8b2" Dec 01 08:45:59 crc kubenswrapper[4813]: I1201 08:45:59.354310 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-mj64s"] Dec 01 08:46:17 crc kubenswrapper[4813]: I1201 08:46:17.221555 4813 patch_prober.go:28] interesting pod/machine-config-daemon-9wn4q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 08:46:17 crc kubenswrapper[4813]: I1201 08:46:17.222440 4813 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9wn4q" podUID="3d423402-e08f-42e9-b60c-6eb9bf2f6d48" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 08:46:24 crc kubenswrapper[4813]: I1201 08:46:24.423196 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-mj64s" podUID="a2b231fa-0ecc-44da-9259-a7dc7261127c" containerName="registry" containerID="cri-o://ac60085c4717fdae779b48e0b9b874818666761bf46e67459bdf3c2cd59f37de" gracePeriod=30 Dec 01 08:46:25 crc kubenswrapper[4813]: I1201 08:46:25.107779 4813 generic.go:334] "Generic (PLEG): container finished" podID="a2b231fa-0ecc-44da-9259-a7dc7261127c" containerID="ac60085c4717fdae779b48e0b9b874818666761bf46e67459bdf3c2cd59f37de" exitCode=0 Dec 01 08:46:25 crc kubenswrapper[4813]: I1201 08:46:25.107933 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-mj64s" event={"ID":"a2b231fa-0ecc-44da-9259-a7dc7261127c","Type":"ContainerDied","Data":"ac60085c4717fdae779b48e0b9b874818666761bf46e67459bdf3c2cd59f37de"} Dec 01 08:46:25 crc kubenswrapper[4813]: I1201 08:46:25.335629 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-mj64s" Dec 01 08:46:25 crc kubenswrapper[4813]: I1201 08:46:25.484805 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/a2b231fa-0ecc-44da-9259-a7dc7261127c-registry-tls\") pod \"a2b231fa-0ecc-44da-9259-a7dc7261127c\" (UID: \"a2b231fa-0ecc-44da-9259-a7dc7261127c\") " Dec 01 08:46:25 crc kubenswrapper[4813]: I1201 08:46:25.485059 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/a2b231fa-0ecc-44da-9259-a7dc7261127c-registry-certificates\") pod \"a2b231fa-0ecc-44da-9259-a7dc7261127c\" (UID: \"a2b231fa-0ecc-44da-9259-a7dc7261127c\") " Dec 01 08:46:25 crc kubenswrapper[4813]: I1201 08:46:25.485176 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a2b231fa-0ecc-44da-9259-a7dc7261127c-bound-sa-token\") pod \"a2b231fa-0ecc-44da-9259-a7dc7261127c\" (UID: \"a2b231fa-0ecc-44da-9259-a7dc7261127c\") " Dec 01 08:46:25 crc kubenswrapper[4813]: I1201 08:46:25.485213 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/a2b231fa-0ecc-44da-9259-a7dc7261127c-installation-pull-secrets\") pod \"a2b231fa-0ecc-44da-9259-a7dc7261127c\" (UID: \"a2b231fa-0ecc-44da-9259-a7dc7261127c\") " Dec 01 08:46:25 crc kubenswrapper[4813]: I1201 08:46:25.485257 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/a2b231fa-0ecc-44da-9259-a7dc7261127c-ca-trust-extracted\") pod \"a2b231fa-0ecc-44da-9259-a7dc7261127c\" (UID: \"a2b231fa-0ecc-44da-9259-a7dc7261127c\") " Dec 01 08:46:25 crc kubenswrapper[4813]: I1201 08:46:25.485323 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9p9xx\" (UniqueName: \"kubernetes.io/projected/a2b231fa-0ecc-44da-9259-a7dc7261127c-kube-api-access-9p9xx\") pod \"a2b231fa-0ecc-44da-9259-a7dc7261127c\" (UID: \"a2b231fa-0ecc-44da-9259-a7dc7261127c\") " Dec 01 08:46:25 crc kubenswrapper[4813]: I1201 08:46:25.485370 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a2b231fa-0ecc-44da-9259-a7dc7261127c-trusted-ca\") pod \"a2b231fa-0ecc-44da-9259-a7dc7261127c\" (UID: \"a2b231fa-0ecc-44da-9259-a7dc7261127c\") " Dec 01 08:46:25 crc kubenswrapper[4813]: I1201 08:46:25.485616 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"a2b231fa-0ecc-44da-9259-a7dc7261127c\" (UID: \"a2b231fa-0ecc-44da-9259-a7dc7261127c\") " Dec 01 08:46:25 crc kubenswrapper[4813]: I1201 08:46:25.486510 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a2b231fa-0ecc-44da-9259-a7dc7261127c-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "a2b231fa-0ecc-44da-9259-a7dc7261127c" (UID: "a2b231fa-0ecc-44da-9259-a7dc7261127c"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:46:25 crc kubenswrapper[4813]: I1201 08:46:25.487892 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a2b231fa-0ecc-44da-9259-a7dc7261127c-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a2b231fa-0ecc-44da-9259-a7dc7261127c" (UID: "a2b231fa-0ecc-44da-9259-a7dc7261127c"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:46:25 crc kubenswrapper[4813]: I1201 08:46:25.491930 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a2b231fa-0ecc-44da-9259-a7dc7261127c-kube-api-access-9p9xx" (OuterVolumeSpecName: "kube-api-access-9p9xx") pod "a2b231fa-0ecc-44da-9259-a7dc7261127c" (UID: "a2b231fa-0ecc-44da-9259-a7dc7261127c"). InnerVolumeSpecName "kube-api-access-9p9xx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:46:25 crc kubenswrapper[4813]: I1201 08:46:25.493426 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a2b231fa-0ecc-44da-9259-a7dc7261127c-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a2b231fa-0ecc-44da-9259-a7dc7261127c" (UID: "a2b231fa-0ecc-44da-9259-a7dc7261127c"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:46:25 crc kubenswrapper[4813]: I1201 08:46:25.497604 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a2b231fa-0ecc-44da-9259-a7dc7261127c-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "a2b231fa-0ecc-44da-9259-a7dc7261127c" (UID: "a2b231fa-0ecc-44da-9259-a7dc7261127c"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:46:25 crc kubenswrapper[4813]: I1201 08:46:25.504358 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a2b231fa-0ecc-44da-9259-a7dc7261127c-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "a2b231fa-0ecc-44da-9259-a7dc7261127c" (UID: "a2b231fa-0ecc-44da-9259-a7dc7261127c"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:46:25 crc kubenswrapper[4813]: I1201 08:46:25.506497 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "a2b231fa-0ecc-44da-9259-a7dc7261127c" (UID: "a2b231fa-0ecc-44da-9259-a7dc7261127c"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Dec 01 08:46:25 crc kubenswrapper[4813]: I1201 08:46:25.508394 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a2b231fa-0ecc-44da-9259-a7dc7261127c-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "a2b231fa-0ecc-44da-9259-a7dc7261127c" (UID: "a2b231fa-0ecc-44da-9259-a7dc7261127c"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 08:46:25 crc kubenswrapper[4813]: I1201 08:46:25.587042 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9p9xx\" (UniqueName: \"kubernetes.io/projected/a2b231fa-0ecc-44da-9259-a7dc7261127c-kube-api-access-9p9xx\") on node \"crc\" DevicePath \"\"" Dec 01 08:46:25 crc kubenswrapper[4813]: I1201 08:46:25.587109 4813 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a2b231fa-0ecc-44da-9259-a7dc7261127c-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 01 08:46:25 crc kubenswrapper[4813]: I1201 08:46:25.587126 4813 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/a2b231fa-0ecc-44da-9259-a7dc7261127c-registry-tls\") on node \"crc\" DevicePath \"\"" Dec 01 08:46:25 crc kubenswrapper[4813]: I1201 08:46:25.587138 4813 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/a2b231fa-0ecc-44da-9259-a7dc7261127c-registry-certificates\") on node \"crc\" DevicePath \"\"" Dec 01 08:46:25 crc kubenswrapper[4813]: I1201 08:46:25.587149 4813 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a2b231fa-0ecc-44da-9259-a7dc7261127c-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 01 08:46:25 crc kubenswrapper[4813]: I1201 08:46:25.587161 4813 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/a2b231fa-0ecc-44da-9259-a7dc7261127c-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Dec 01 08:46:25 crc kubenswrapper[4813]: I1201 08:46:25.587177 4813 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/a2b231fa-0ecc-44da-9259-a7dc7261127c-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Dec 01 08:46:26 crc kubenswrapper[4813]: I1201 08:46:26.120587 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-mj64s" event={"ID":"a2b231fa-0ecc-44da-9259-a7dc7261127c","Type":"ContainerDied","Data":"824177aa29c11338724cc5b8be3c5968e74b376b652172baad8b598d76606b57"} Dec 01 08:46:26 crc kubenswrapper[4813]: I1201 08:46:26.120765 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-mj64s" Dec 01 08:46:26 crc kubenswrapper[4813]: I1201 08:46:26.121271 4813 scope.go:117] "RemoveContainer" containerID="ac60085c4717fdae779b48e0b9b874818666761bf46e67459bdf3c2cd59f37de" Dec 01 08:46:26 crc kubenswrapper[4813]: I1201 08:46:26.184015 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-mj64s"] Dec 01 08:46:26 crc kubenswrapper[4813]: I1201 08:46:26.193908 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-mj64s"] Dec 01 08:46:26 crc kubenswrapper[4813]: I1201 08:46:26.408770 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a2b231fa-0ecc-44da-9259-a7dc7261127c" path="/var/lib/kubelet/pods/a2b231fa-0ecc-44da-9259-a7dc7261127c/volumes" Dec 01 08:46:47 crc kubenswrapper[4813]: I1201 08:46:47.221845 4813 patch_prober.go:28] interesting pod/machine-config-daemon-9wn4q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 08:46:47 crc kubenswrapper[4813]: I1201 08:46:47.222508 4813 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9wn4q" podUID="3d423402-e08f-42e9-b60c-6eb9bf2f6d48" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 08:46:47 crc kubenswrapper[4813]: I1201 08:46:47.222625 4813 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-9wn4q" Dec 01 08:46:47 crc kubenswrapper[4813]: I1201 08:46:47.223527 4813 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"5899e07de4e292f45ea6ae9799d87cd4fae3236f41b153e0795a33f5f1ed93ff"} pod="openshift-machine-config-operator/machine-config-daemon-9wn4q" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 01 08:46:47 crc kubenswrapper[4813]: I1201 08:46:47.223643 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-9wn4q" podUID="3d423402-e08f-42e9-b60c-6eb9bf2f6d48" containerName="machine-config-daemon" containerID="cri-o://5899e07de4e292f45ea6ae9799d87cd4fae3236f41b153e0795a33f5f1ed93ff" gracePeriod=600 Dec 01 08:46:48 crc kubenswrapper[4813]: I1201 08:46:48.310258 4813 generic.go:334] "Generic (PLEG): container finished" podID="3d423402-e08f-42e9-b60c-6eb9bf2f6d48" containerID="5899e07de4e292f45ea6ae9799d87cd4fae3236f41b153e0795a33f5f1ed93ff" exitCode=0 Dec 01 08:46:48 crc kubenswrapper[4813]: I1201 08:46:48.310341 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-9wn4q" event={"ID":"3d423402-e08f-42e9-b60c-6eb9bf2f6d48","Type":"ContainerDied","Data":"5899e07de4e292f45ea6ae9799d87cd4fae3236f41b153e0795a33f5f1ed93ff"} Dec 01 08:46:48 crc kubenswrapper[4813]: I1201 08:46:48.310662 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-9wn4q" event={"ID":"3d423402-e08f-42e9-b60c-6eb9bf2f6d48","Type":"ContainerStarted","Data":"e6f5f635caf35fdd2a0d88216d0cd5f28601a09788d14975d2ea2dca78233ac3"} Dec 01 08:46:48 crc kubenswrapper[4813]: I1201 08:46:48.310829 4813 scope.go:117] "RemoveContainer" containerID="1c915733072b32a63237e2401cbf84b059aca396c2c0db884f0f439d599a35d2" Dec 01 08:48:47 crc kubenswrapper[4813]: I1201 08:48:47.222123 4813 patch_prober.go:28] interesting pod/machine-config-daemon-9wn4q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 08:48:47 crc kubenswrapper[4813]: I1201 08:48:47.222873 4813 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9wn4q" podUID="3d423402-e08f-42e9-b60c-6eb9bf2f6d48" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 08:49:17 crc kubenswrapper[4813]: I1201 08:49:17.222102 4813 patch_prober.go:28] interesting pod/machine-config-daemon-9wn4q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 08:49:17 crc kubenswrapper[4813]: I1201 08:49:17.222772 4813 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9wn4q" podUID="3d423402-e08f-42e9-b60c-6eb9bf2f6d48" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 08:49:47 crc kubenswrapper[4813]: I1201 08:49:47.221586 4813 patch_prober.go:28] interesting pod/machine-config-daemon-9wn4q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 08:49:47 crc kubenswrapper[4813]: I1201 08:49:47.222389 4813 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9wn4q" podUID="3d423402-e08f-42e9-b60c-6eb9bf2f6d48" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 08:49:47 crc kubenswrapper[4813]: I1201 08:49:47.222559 4813 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-9wn4q" Dec 01 08:49:47 crc kubenswrapper[4813]: I1201 08:49:47.223508 4813 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"e6f5f635caf35fdd2a0d88216d0cd5f28601a09788d14975d2ea2dca78233ac3"} pod="openshift-machine-config-operator/machine-config-daemon-9wn4q" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 01 08:49:47 crc kubenswrapper[4813]: I1201 08:49:47.223661 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-9wn4q" podUID="3d423402-e08f-42e9-b60c-6eb9bf2f6d48" containerName="machine-config-daemon" containerID="cri-o://e6f5f635caf35fdd2a0d88216d0cd5f28601a09788d14975d2ea2dca78233ac3" gracePeriod=600 Dec 01 08:49:47 crc kubenswrapper[4813]: I1201 08:49:47.862919 4813 generic.go:334] "Generic (PLEG): container finished" podID="3d423402-e08f-42e9-b60c-6eb9bf2f6d48" containerID="e6f5f635caf35fdd2a0d88216d0cd5f28601a09788d14975d2ea2dca78233ac3" exitCode=0 Dec 01 08:49:47 crc kubenswrapper[4813]: I1201 08:49:47.863163 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-9wn4q" event={"ID":"3d423402-e08f-42e9-b60c-6eb9bf2f6d48","Type":"ContainerDied","Data":"e6f5f635caf35fdd2a0d88216d0cd5f28601a09788d14975d2ea2dca78233ac3"} Dec 01 08:49:47 crc kubenswrapper[4813]: I1201 08:49:47.863346 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-9wn4q" event={"ID":"3d423402-e08f-42e9-b60c-6eb9bf2f6d48","Type":"ContainerStarted","Data":"c9bc69c1312504b715819da3ef57a1ae8b88ab55017f5060cf77c80a5fa0c1e8"} Dec 01 08:49:47 crc kubenswrapper[4813]: I1201 08:49:47.863417 4813 scope.go:117] "RemoveContainer" containerID="5899e07de4e292f45ea6ae9799d87cd4fae3236f41b153e0795a33f5f1ed93ff" Dec 01 08:50:50 crc kubenswrapper[4813]: I1201 08:50:50.204411 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-j4pcc"] Dec 01 08:50:50 crc kubenswrapper[4813]: I1201 08:50:50.205913 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-j4pcc" podUID="b7505e14-3560-4d0d-be2c-d6a7a5953865" containerName="ovn-controller" containerID="cri-o://ce8963bf1e9ee66926907aaf3a03d18977f4d0a1983d7121b62fbc8e02a37e4d" gracePeriod=30 Dec 01 08:50:50 crc kubenswrapper[4813]: I1201 08:50:50.206122 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-j4pcc" podUID="b7505e14-3560-4d0d-be2c-d6a7a5953865" containerName="kube-rbac-proxy-node" containerID="cri-o://ccc30a63dc1c23ff9880fb04af1f31f0656b2696d359988e23deed23989f480a" gracePeriod=30 Dec 01 08:50:50 crc kubenswrapper[4813]: I1201 08:50:50.205928 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-j4pcc" podUID="b7505e14-3560-4d0d-be2c-d6a7a5953865" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://944977e9722025088720107530d0c91fc693593f380e4b49678f28a56da180cc" gracePeriod=30 Dec 01 08:50:50 crc kubenswrapper[4813]: I1201 08:50:50.206207 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-j4pcc" podUID="b7505e14-3560-4d0d-be2c-d6a7a5953865" containerName="northd" containerID="cri-o://7f19c67b1c6c80c7f3223512102810d46ede1445464d21ae3bbea83424539b2b" gracePeriod=30 Dec 01 08:50:50 crc kubenswrapper[4813]: I1201 08:50:50.206173 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-j4pcc" podUID="b7505e14-3560-4d0d-be2c-d6a7a5953865" containerName="ovn-acl-logging" containerID="cri-o://9d984c90302e57f3e5a6f3ca3613b415db04ac74a5a228582f1b52ef1d59f93e" gracePeriod=30 Dec 01 08:50:50 crc kubenswrapper[4813]: I1201 08:50:50.206098 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-j4pcc" podUID="b7505e14-3560-4d0d-be2c-d6a7a5953865" containerName="nbdb" containerID="cri-o://dc32d930ab23f30103d8fa42291909cbec89ebb75c5c53cd783cd27fe1a3f09e" gracePeriod=30 Dec 01 08:50:50 crc kubenswrapper[4813]: I1201 08:50:50.206479 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-j4pcc" podUID="b7505e14-3560-4d0d-be2c-d6a7a5953865" containerName="sbdb" containerID="cri-o://b45454b440c5108fa150a6b856afa1778ba6e9f89633433cb3117b68f00f72c3" gracePeriod=30 Dec 01 08:50:50 crc kubenswrapper[4813]: I1201 08:50:50.252883 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-j4pcc" podUID="b7505e14-3560-4d0d-be2c-d6a7a5953865" containerName="ovnkube-controller" containerID="cri-o://6aa74106386539915851c80789dac37ef865908f24cb6a67ffa406c2599cb5d7" gracePeriod=30 Dec 01 08:50:50 crc kubenswrapper[4813]: I1201 08:50:50.389384 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-j4pcc_b7505e14-3560-4d0d-be2c-d6a7a5953865/ovnkube-controller/2.log" Dec 01 08:50:50 crc kubenswrapper[4813]: I1201 08:50:50.391677 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-j4pcc_b7505e14-3560-4d0d-be2c-d6a7a5953865/ovn-acl-logging/0.log" Dec 01 08:50:50 crc kubenswrapper[4813]: I1201 08:50:50.392205 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-j4pcc_b7505e14-3560-4d0d-be2c-d6a7a5953865/ovn-controller/0.log" Dec 01 08:50:50 crc kubenswrapper[4813]: I1201 08:50:50.392670 4813 generic.go:334] "Generic (PLEG): container finished" podID="b7505e14-3560-4d0d-be2c-d6a7a5953865" containerID="6aa74106386539915851c80789dac37ef865908f24cb6a67ffa406c2599cb5d7" exitCode=0 Dec 01 08:50:50 crc kubenswrapper[4813]: I1201 08:50:50.392730 4813 generic.go:334] "Generic (PLEG): container finished" podID="b7505e14-3560-4d0d-be2c-d6a7a5953865" containerID="944977e9722025088720107530d0c91fc693593f380e4b49678f28a56da180cc" exitCode=0 Dec 01 08:50:50 crc kubenswrapper[4813]: I1201 08:50:50.392746 4813 generic.go:334] "Generic (PLEG): container finished" podID="b7505e14-3560-4d0d-be2c-d6a7a5953865" containerID="ccc30a63dc1c23ff9880fb04af1f31f0656b2696d359988e23deed23989f480a" exitCode=0 Dec 01 08:50:50 crc kubenswrapper[4813]: I1201 08:50:50.392759 4813 generic.go:334] "Generic (PLEG): container finished" podID="b7505e14-3560-4d0d-be2c-d6a7a5953865" containerID="9d984c90302e57f3e5a6f3ca3613b415db04ac74a5a228582f1b52ef1d59f93e" exitCode=143 Dec 01 08:50:50 crc kubenswrapper[4813]: I1201 08:50:50.392772 4813 generic.go:334] "Generic (PLEG): container finished" podID="b7505e14-3560-4d0d-be2c-d6a7a5953865" containerID="ce8963bf1e9ee66926907aaf3a03d18977f4d0a1983d7121b62fbc8e02a37e4d" exitCode=143 Dec 01 08:50:50 crc kubenswrapper[4813]: I1201 08:50:50.394675 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-b6qqv_b9656a03-0096-4c89-b0da-e37e4103611b/kube-multus/1.log" Dec 01 08:50:50 crc kubenswrapper[4813]: I1201 08:50:50.395196 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-b6qqv_b9656a03-0096-4c89-b0da-e37e4103611b/kube-multus/0.log" Dec 01 08:50:50 crc kubenswrapper[4813]: I1201 08:50:50.395231 4813 generic.go:334] "Generic (PLEG): container finished" podID="b9656a03-0096-4c89-b0da-e37e4103611b" containerID="2b730cdf02e6456d9a1d02d66b94064ca71052eb8f373eaabfe327369eb03a15" exitCode=2 Dec 01 08:50:50 crc kubenswrapper[4813]: I1201 08:50:50.400471 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-j4pcc" event={"ID":"b7505e14-3560-4d0d-be2c-d6a7a5953865","Type":"ContainerDied","Data":"6aa74106386539915851c80789dac37ef865908f24cb6a67ffa406c2599cb5d7"} Dec 01 08:50:50 crc kubenswrapper[4813]: I1201 08:50:50.400527 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-j4pcc" event={"ID":"b7505e14-3560-4d0d-be2c-d6a7a5953865","Type":"ContainerDied","Data":"944977e9722025088720107530d0c91fc693593f380e4b49678f28a56da180cc"} Dec 01 08:50:50 crc kubenswrapper[4813]: I1201 08:50:50.400539 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-j4pcc" event={"ID":"b7505e14-3560-4d0d-be2c-d6a7a5953865","Type":"ContainerDied","Data":"ccc30a63dc1c23ff9880fb04af1f31f0656b2696d359988e23deed23989f480a"} Dec 01 08:50:50 crc kubenswrapper[4813]: I1201 08:50:50.400549 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-j4pcc" event={"ID":"b7505e14-3560-4d0d-be2c-d6a7a5953865","Type":"ContainerDied","Data":"9d984c90302e57f3e5a6f3ca3613b415db04ac74a5a228582f1b52ef1d59f93e"} Dec 01 08:50:50 crc kubenswrapper[4813]: I1201 08:50:50.400557 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-j4pcc" event={"ID":"b7505e14-3560-4d0d-be2c-d6a7a5953865","Type":"ContainerDied","Data":"ce8963bf1e9ee66926907aaf3a03d18977f4d0a1983d7121b62fbc8e02a37e4d"} Dec 01 08:50:50 crc kubenswrapper[4813]: I1201 08:50:50.400566 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-b6qqv" event={"ID":"b9656a03-0096-4c89-b0da-e37e4103611b","Type":"ContainerDied","Data":"2b730cdf02e6456d9a1d02d66b94064ca71052eb8f373eaabfe327369eb03a15"} Dec 01 08:50:50 crc kubenswrapper[4813]: I1201 08:50:50.401088 4813 scope.go:117] "RemoveContainer" containerID="2b730cdf02e6456d9a1d02d66b94064ca71052eb8f373eaabfe327369eb03a15" Dec 01 08:50:50 crc kubenswrapper[4813]: E1201 08:50:50.401302 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-multus pod=multus-b6qqv_openshift-multus(b9656a03-0096-4c89-b0da-e37e4103611b)\"" pod="openshift-multus/multus-b6qqv" podUID="b9656a03-0096-4c89-b0da-e37e4103611b" Dec 01 08:50:50 crc kubenswrapper[4813]: I1201 08:50:50.401514 4813 scope.go:117] "RemoveContainer" containerID="fb17fb9ae11fa62235328567418fb259f4f8540ebac635842d3afe751dcbde8c" Dec 01 08:50:50 crc kubenswrapper[4813]: I1201 08:50:50.515423 4813 scope.go:117] "RemoveContainer" containerID="c4195aebc1b2e388d480920a7f72257f6660afc99ad8c1a09478ef445739b272" Dec 01 08:50:50 crc kubenswrapper[4813]: I1201 08:50:50.549587 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-j4pcc_b7505e14-3560-4d0d-be2c-d6a7a5953865/ovn-acl-logging/0.log" Dec 01 08:50:50 crc kubenswrapper[4813]: I1201 08:50:50.550347 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-j4pcc_b7505e14-3560-4d0d-be2c-d6a7a5953865/ovn-controller/0.log" Dec 01 08:50:50 crc kubenswrapper[4813]: I1201 08:50:50.550806 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-j4pcc" Dec 01 08:50:50 crc kubenswrapper[4813]: I1201 08:50:50.614191 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-qnss9"] Dec 01 08:50:50 crc kubenswrapper[4813]: E1201 08:50:50.614670 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b7505e14-3560-4d0d-be2c-d6a7a5953865" containerName="kube-rbac-proxy-ovn-metrics" Dec 01 08:50:50 crc kubenswrapper[4813]: I1201 08:50:50.614712 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="b7505e14-3560-4d0d-be2c-d6a7a5953865" containerName="kube-rbac-proxy-ovn-metrics" Dec 01 08:50:50 crc kubenswrapper[4813]: E1201 08:50:50.614741 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b7505e14-3560-4d0d-be2c-d6a7a5953865" containerName="ovn-acl-logging" Dec 01 08:50:50 crc kubenswrapper[4813]: I1201 08:50:50.614754 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="b7505e14-3560-4d0d-be2c-d6a7a5953865" containerName="ovn-acl-logging" Dec 01 08:50:50 crc kubenswrapper[4813]: E1201 08:50:50.614776 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b7505e14-3560-4d0d-be2c-d6a7a5953865" containerName="kubecfg-setup" Dec 01 08:50:50 crc kubenswrapper[4813]: I1201 08:50:50.614789 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="b7505e14-3560-4d0d-be2c-d6a7a5953865" containerName="kubecfg-setup" Dec 01 08:50:50 crc kubenswrapper[4813]: E1201 08:50:50.614803 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b7505e14-3560-4d0d-be2c-d6a7a5953865" containerName="kube-rbac-proxy-node" Dec 01 08:50:50 crc kubenswrapper[4813]: I1201 08:50:50.614815 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="b7505e14-3560-4d0d-be2c-d6a7a5953865" containerName="kube-rbac-proxy-node" Dec 01 08:50:50 crc kubenswrapper[4813]: E1201 08:50:50.614834 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b7505e14-3560-4d0d-be2c-d6a7a5953865" containerName="ovn-controller" Dec 01 08:50:50 crc kubenswrapper[4813]: I1201 08:50:50.614845 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="b7505e14-3560-4d0d-be2c-d6a7a5953865" containerName="ovn-controller" Dec 01 08:50:50 crc kubenswrapper[4813]: E1201 08:50:50.614863 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b7505e14-3560-4d0d-be2c-d6a7a5953865" containerName="ovnkube-controller" Dec 01 08:50:50 crc kubenswrapper[4813]: I1201 08:50:50.614875 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="b7505e14-3560-4d0d-be2c-d6a7a5953865" containerName="ovnkube-controller" Dec 01 08:50:50 crc kubenswrapper[4813]: E1201 08:50:50.614895 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a2b231fa-0ecc-44da-9259-a7dc7261127c" containerName="registry" Dec 01 08:50:50 crc kubenswrapper[4813]: I1201 08:50:50.614907 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="a2b231fa-0ecc-44da-9259-a7dc7261127c" containerName="registry" Dec 01 08:50:50 crc kubenswrapper[4813]: E1201 08:50:50.614928 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b7505e14-3560-4d0d-be2c-d6a7a5953865" containerName="sbdb" Dec 01 08:50:50 crc kubenswrapper[4813]: I1201 08:50:50.614940 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="b7505e14-3560-4d0d-be2c-d6a7a5953865" containerName="sbdb" Dec 01 08:50:50 crc kubenswrapper[4813]: E1201 08:50:50.614959 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b7505e14-3560-4d0d-be2c-d6a7a5953865" containerName="ovnkube-controller" Dec 01 08:50:50 crc kubenswrapper[4813]: I1201 08:50:50.614997 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="b7505e14-3560-4d0d-be2c-d6a7a5953865" containerName="ovnkube-controller" Dec 01 08:50:50 crc kubenswrapper[4813]: E1201 08:50:50.615010 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b7505e14-3560-4d0d-be2c-d6a7a5953865" containerName="northd" Dec 01 08:50:50 crc kubenswrapper[4813]: I1201 08:50:50.615022 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="b7505e14-3560-4d0d-be2c-d6a7a5953865" containerName="northd" Dec 01 08:50:50 crc kubenswrapper[4813]: E1201 08:50:50.615043 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b7505e14-3560-4d0d-be2c-d6a7a5953865" containerName="nbdb" Dec 01 08:50:50 crc kubenswrapper[4813]: I1201 08:50:50.615054 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="b7505e14-3560-4d0d-be2c-d6a7a5953865" containerName="nbdb" Dec 01 08:50:50 crc kubenswrapper[4813]: E1201 08:50:50.615072 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b7505e14-3560-4d0d-be2c-d6a7a5953865" containerName="ovnkube-controller" Dec 01 08:50:50 crc kubenswrapper[4813]: I1201 08:50:50.615083 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="b7505e14-3560-4d0d-be2c-d6a7a5953865" containerName="ovnkube-controller" Dec 01 08:50:50 crc kubenswrapper[4813]: I1201 08:50:50.615321 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="b7505e14-3560-4d0d-be2c-d6a7a5953865" containerName="ovnkube-controller" Dec 01 08:50:50 crc kubenswrapper[4813]: I1201 08:50:50.615349 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="b7505e14-3560-4d0d-be2c-d6a7a5953865" containerName="ovn-controller" Dec 01 08:50:50 crc kubenswrapper[4813]: I1201 08:50:50.615364 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="b7505e14-3560-4d0d-be2c-d6a7a5953865" containerName="ovnkube-controller" Dec 01 08:50:50 crc kubenswrapper[4813]: I1201 08:50:50.615380 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="b7505e14-3560-4d0d-be2c-d6a7a5953865" containerName="sbdb" Dec 01 08:50:50 crc kubenswrapper[4813]: I1201 08:50:50.615400 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="b7505e14-3560-4d0d-be2c-d6a7a5953865" containerName="kube-rbac-proxy-node" Dec 01 08:50:50 crc kubenswrapper[4813]: I1201 08:50:50.615413 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="b7505e14-3560-4d0d-be2c-d6a7a5953865" containerName="kube-rbac-proxy-ovn-metrics" Dec 01 08:50:50 crc kubenswrapper[4813]: I1201 08:50:50.615431 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="a2b231fa-0ecc-44da-9259-a7dc7261127c" containerName="registry" Dec 01 08:50:50 crc kubenswrapper[4813]: I1201 08:50:50.615448 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="b7505e14-3560-4d0d-be2c-d6a7a5953865" containerName="northd" Dec 01 08:50:50 crc kubenswrapper[4813]: I1201 08:50:50.615461 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="b7505e14-3560-4d0d-be2c-d6a7a5953865" containerName="ovnkube-controller" Dec 01 08:50:50 crc kubenswrapper[4813]: I1201 08:50:50.615473 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="b7505e14-3560-4d0d-be2c-d6a7a5953865" containerName="ovn-acl-logging" Dec 01 08:50:50 crc kubenswrapper[4813]: I1201 08:50:50.615487 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="b7505e14-3560-4d0d-be2c-d6a7a5953865" containerName="nbdb" Dec 01 08:50:50 crc kubenswrapper[4813]: E1201 08:50:50.615659 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b7505e14-3560-4d0d-be2c-d6a7a5953865" containerName="ovnkube-controller" Dec 01 08:50:50 crc kubenswrapper[4813]: I1201 08:50:50.615675 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="b7505e14-3560-4d0d-be2c-d6a7a5953865" containerName="ovnkube-controller" Dec 01 08:50:50 crc kubenswrapper[4813]: I1201 08:50:50.615901 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="b7505e14-3560-4d0d-be2c-d6a7a5953865" containerName="ovnkube-controller" Dec 01 08:50:50 crc kubenswrapper[4813]: I1201 08:50:50.619292 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-qnss9" Dec 01 08:50:50 crc kubenswrapper[4813]: I1201 08:50:50.665879 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/b7505e14-3560-4d0d-be2c-d6a7a5953865-host-cni-bin\") pod \"b7505e14-3560-4d0d-be2c-d6a7a5953865\" (UID: \"b7505e14-3560-4d0d-be2c-d6a7a5953865\") " Dec 01 08:50:50 crc kubenswrapper[4813]: I1201 08:50:50.665924 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/b7505e14-3560-4d0d-be2c-d6a7a5953865-host-run-ovn-kubernetes\") pod \"b7505e14-3560-4d0d-be2c-d6a7a5953865\" (UID: \"b7505e14-3560-4d0d-be2c-d6a7a5953865\") " Dec 01 08:50:50 crc kubenswrapper[4813]: I1201 08:50:50.665948 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/b7505e14-3560-4d0d-be2c-d6a7a5953865-node-log\") pod \"b7505e14-3560-4d0d-be2c-d6a7a5953865\" (UID: \"b7505e14-3560-4d0d-be2c-d6a7a5953865\") " Dec 01 08:50:50 crc kubenswrapper[4813]: I1201 08:50:50.665999 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/b7505e14-3560-4d0d-be2c-d6a7a5953865-host-slash\") pod \"b7505e14-3560-4d0d-be2c-d6a7a5953865\" (UID: \"b7505e14-3560-4d0d-be2c-d6a7a5953865\") " Dec 01 08:50:50 crc kubenswrapper[4813]: I1201 08:50:50.666024 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/b7505e14-3560-4d0d-be2c-d6a7a5953865-host-run-netns\") pod \"b7505e14-3560-4d0d-be2c-d6a7a5953865\" (UID: \"b7505e14-3560-4d0d-be2c-d6a7a5953865\") " Dec 01 08:50:50 crc kubenswrapper[4813]: I1201 08:50:50.666030 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b7505e14-3560-4d0d-be2c-d6a7a5953865-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "b7505e14-3560-4d0d-be2c-d6a7a5953865" (UID: "b7505e14-3560-4d0d-be2c-d6a7a5953865"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 08:50:50 crc kubenswrapper[4813]: I1201 08:50:50.666054 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/b7505e14-3560-4d0d-be2c-d6a7a5953865-host-cni-netd\") pod \"b7505e14-3560-4d0d-be2c-d6a7a5953865\" (UID: \"b7505e14-3560-4d0d-be2c-d6a7a5953865\") " Dec 01 08:50:50 crc kubenswrapper[4813]: I1201 08:50:50.666081 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b7505e14-3560-4d0d-be2c-d6a7a5953865-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "b7505e14-3560-4d0d-be2c-d6a7a5953865" (UID: "b7505e14-3560-4d0d-be2c-d6a7a5953865"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 08:50:50 crc kubenswrapper[4813]: I1201 08:50:50.666114 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b7505e14-3560-4d0d-be2c-d6a7a5953865-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "b7505e14-3560-4d0d-be2c-d6a7a5953865" (UID: "b7505e14-3560-4d0d-be2c-d6a7a5953865"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 08:50:50 crc kubenswrapper[4813]: I1201 08:50:50.666122 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/b7505e14-3560-4d0d-be2c-d6a7a5953865-run-ovn\") pod \"b7505e14-3560-4d0d-be2c-d6a7a5953865\" (UID: \"b7505e14-3560-4d0d-be2c-d6a7a5953865\") " Dec 01 08:50:50 crc kubenswrapper[4813]: I1201 08:50:50.666122 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b7505e14-3560-4d0d-be2c-d6a7a5953865-node-log" (OuterVolumeSpecName: "node-log") pod "b7505e14-3560-4d0d-be2c-d6a7a5953865" (UID: "b7505e14-3560-4d0d-be2c-d6a7a5953865"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 08:50:50 crc kubenswrapper[4813]: I1201 08:50:50.666134 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b7505e14-3560-4d0d-be2c-d6a7a5953865-host-slash" (OuterVolumeSpecName: "host-slash") pod "b7505e14-3560-4d0d-be2c-d6a7a5953865" (UID: "b7505e14-3560-4d0d-be2c-d6a7a5953865"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 08:50:50 crc kubenswrapper[4813]: I1201 08:50:50.666141 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b7505e14-3560-4d0d-be2c-d6a7a5953865-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "b7505e14-3560-4d0d-be2c-d6a7a5953865" (UID: "b7505e14-3560-4d0d-be2c-d6a7a5953865"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 08:50:50 crc kubenswrapper[4813]: I1201 08:50:50.666167 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b7505e14-3560-4d0d-be2c-d6a7a5953865-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "b7505e14-3560-4d0d-be2c-d6a7a5953865" (UID: "b7505e14-3560-4d0d-be2c-d6a7a5953865"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 08:50:50 crc kubenswrapper[4813]: I1201 08:50:50.666155 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dgqwb\" (UniqueName: \"kubernetes.io/projected/b7505e14-3560-4d0d-be2c-d6a7a5953865-kube-api-access-dgqwb\") pod \"b7505e14-3560-4d0d-be2c-d6a7a5953865\" (UID: \"b7505e14-3560-4d0d-be2c-d6a7a5953865\") " Dec 01 08:50:50 crc kubenswrapper[4813]: I1201 08:50:50.666255 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/b7505e14-3560-4d0d-be2c-d6a7a5953865-run-systemd\") pod \"b7505e14-3560-4d0d-be2c-d6a7a5953865\" (UID: \"b7505e14-3560-4d0d-be2c-d6a7a5953865\") " Dec 01 08:50:50 crc kubenswrapper[4813]: I1201 08:50:50.666294 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/b7505e14-3560-4d0d-be2c-d6a7a5953865-etc-openvswitch\") pod \"b7505e14-3560-4d0d-be2c-d6a7a5953865\" (UID: \"b7505e14-3560-4d0d-be2c-d6a7a5953865\") " Dec 01 08:50:50 crc kubenswrapper[4813]: I1201 08:50:50.666329 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/b7505e14-3560-4d0d-be2c-d6a7a5953865-host-kubelet\") pod \"b7505e14-3560-4d0d-be2c-d6a7a5953865\" (UID: \"b7505e14-3560-4d0d-be2c-d6a7a5953865\") " Dec 01 08:50:50 crc kubenswrapper[4813]: I1201 08:50:50.666369 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/b7505e14-3560-4d0d-be2c-d6a7a5953865-ovnkube-config\") pod \"b7505e14-3560-4d0d-be2c-d6a7a5953865\" (UID: \"b7505e14-3560-4d0d-be2c-d6a7a5953865\") " Dec 01 08:50:50 crc kubenswrapper[4813]: I1201 08:50:50.666401 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/b7505e14-3560-4d0d-be2c-d6a7a5953865-run-openvswitch\") pod \"b7505e14-3560-4d0d-be2c-d6a7a5953865\" (UID: \"b7505e14-3560-4d0d-be2c-d6a7a5953865\") " Dec 01 08:50:50 crc kubenswrapper[4813]: I1201 08:50:50.666448 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/b7505e14-3560-4d0d-be2c-d6a7a5953865-log-socket\") pod \"b7505e14-3560-4d0d-be2c-d6a7a5953865\" (UID: \"b7505e14-3560-4d0d-be2c-d6a7a5953865\") " Dec 01 08:50:50 crc kubenswrapper[4813]: I1201 08:50:50.666494 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/b7505e14-3560-4d0d-be2c-d6a7a5953865-host-var-lib-cni-networks-ovn-kubernetes\") pod \"b7505e14-3560-4d0d-be2c-d6a7a5953865\" (UID: \"b7505e14-3560-4d0d-be2c-d6a7a5953865\") " Dec 01 08:50:50 crc kubenswrapper[4813]: I1201 08:50:50.666530 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/b7505e14-3560-4d0d-be2c-d6a7a5953865-env-overrides\") pod \"b7505e14-3560-4d0d-be2c-d6a7a5953865\" (UID: \"b7505e14-3560-4d0d-be2c-d6a7a5953865\") " Dec 01 08:50:50 crc kubenswrapper[4813]: I1201 08:50:50.666589 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/b7505e14-3560-4d0d-be2c-d6a7a5953865-var-lib-openvswitch\") pod \"b7505e14-3560-4d0d-be2c-d6a7a5953865\" (UID: \"b7505e14-3560-4d0d-be2c-d6a7a5953865\") " Dec 01 08:50:50 crc kubenswrapper[4813]: I1201 08:50:50.666628 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/b7505e14-3560-4d0d-be2c-d6a7a5953865-ovn-node-metrics-cert\") pod \"b7505e14-3560-4d0d-be2c-d6a7a5953865\" (UID: \"b7505e14-3560-4d0d-be2c-d6a7a5953865\") " Dec 01 08:50:50 crc kubenswrapper[4813]: I1201 08:50:50.666693 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/b7505e14-3560-4d0d-be2c-d6a7a5953865-systemd-units\") pod \"b7505e14-3560-4d0d-be2c-d6a7a5953865\" (UID: \"b7505e14-3560-4d0d-be2c-d6a7a5953865\") " Dec 01 08:50:50 crc kubenswrapper[4813]: I1201 08:50:50.666728 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/b7505e14-3560-4d0d-be2c-d6a7a5953865-ovnkube-script-lib\") pod \"b7505e14-3560-4d0d-be2c-d6a7a5953865\" (UID: \"b7505e14-3560-4d0d-be2c-d6a7a5953865\") " Dec 01 08:50:50 crc kubenswrapper[4813]: I1201 08:50:50.667002 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b7505e14-3560-4d0d-be2c-d6a7a5953865-log-socket" (OuterVolumeSpecName: "log-socket") pod "b7505e14-3560-4d0d-be2c-d6a7a5953865" (UID: "b7505e14-3560-4d0d-be2c-d6a7a5953865"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 08:50:50 crc kubenswrapper[4813]: I1201 08:50:50.667007 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b7505e14-3560-4d0d-be2c-d6a7a5953865-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "b7505e14-3560-4d0d-be2c-d6a7a5953865" (UID: "b7505e14-3560-4d0d-be2c-d6a7a5953865"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 08:50:50 crc kubenswrapper[4813]: I1201 08:50:50.667237 4813 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/b7505e14-3560-4d0d-be2c-d6a7a5953865-host-run-netns\") on node \"crc\" DevicePath \"\"" Dec 01 08:50:50 crc kubenswrapper[4813]: I1201 08:50:50.667269 4813 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/b7505e14-3560-4d0d-be2c-d6a7a5953865-host-cni-netd\") on node \"crc\" DevicePath \"\"" Dec 01 08:50:50 crc kubenswrapper[4813]: I1201 08:50:50.667288 4813 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/b7505e14-3560-4d0d-be2c-d6a7a5953865-run-ovn\") on node \"crc\" DevicePath \"\"" Dec 01 08:50:50 crc kubenswrapper[4813]: I1201 08:50:50.667305 4813 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/b7505e14-3560-4d0d-be2c-d6a7a5953865-host-kubelet\") on node \"crc\" DevicePath \"\"" Dec 01 08:50:50 crc kubenswrapper[4813]: I1201 08:50:50.667319 4813 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/b7505e14-3560-4d0d-be2c-d6a7a5953865-log-socket\") on node \"crc\" DevicePath \"\"" Dec 01 08:50:50 crc kubenswrapper[4813]: I1201 08:50:50.667334 4813 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/b7505e14-3560-4d0d-be2c-d6a7a5953865-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Dec 01 08:50:50 crc kubenswrapper[4813]: I1201 08:50:50.667349 4813 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/b7505e14-3560-4d0d-be2c-d6a7a5953865-host-cni-bin\") on node \"crc\" DevicePath \"\"" Dec 01 08:50:50 crc kubenswrapper[4813]: I1201 08:50:50.667364 4813 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/b7505e14-3560-4d0d-be2c-d6a7a5953865-node-log\") on node \"crc\" DevicePath \"\"" Dec 01 08:50:50 crc kubenswrapper[4813]: I1201 08:50:50.667379 4813 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/b7505e14-3560-4d0d-be2c-d6a7a5953865-host-slash\") on node \"crc\" DevicePath \"\"" Dec 01 08:50:50 crc kubenswrapper[4813]: I1201 08:50:50.667423 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b7505e14-3560-4d0d-be2c-d6a7a5953865-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "b7505e14-3560-4d0d-be2c-d6a7a5953865" (UID: "b7505e14-3560-4d0d-be2c-d6a7a5953865"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 08:50:50 crc kubenswrapper[4813]: I1201 08:50:50.667542 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b7505e14-3560-4d0d-be2c-d6a7a5953865-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "b7505e14-3560-4d0d-be2c-d6a7a5953865" (UID: "b7505e14-3560-4d0d-be2c-d6a7a5953865"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:50:50 crc kubenswrapper[4813]: I1201 08:50:50.667606 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b7505e14-3560-4d0d-be2c-d6a7a5953865-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "b7505e14-3560-4d0d-be2c-d6a7a5953865" (UID: "b7505e14-3560-4d0d-be2c-d6a7a5953865"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 08:50:50 crc kubenswrapper[4813]: I1201 08:50:50.668006 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b7505e14-3560-4d0d-be2c-d6a7a5953865-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "b7505e14-3560-4d0d-be2c-d6a7a5953865" (UID: "b7505e14-3560-4d0d-be2c-d6a7a5953865"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:50:50 crc kubenswrapper[4813]: I1201 08:50:50.668067 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b7505e14-3560-4d0d-be2c-d6a7a5953865-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "b7505e14-3560-4d0d-be2c-d6a7a5953865" (UID: "b7505e14-3560-4d0d-be2c-d6a7a5953865"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 08:50:50 crc kubenswrapper[4813]: I1201 08:50:50.668566 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b7505e14-3560-4d0d-be2c-d6a7a5953865-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "b7505e14-3560-4d0d-be2c-d6a7a5953865" (UID: "b7505e14-3560-4d0d-be2c-d6a7a5953865"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:50:50 crc kubenswrapper[4813]: I1201 08:50:50.668623 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b7505e14-3560-4d0d-be2c-d6a7a5953865-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "b7505e14-3560-4d0d-be2c-d6a7a5953865" (UID: "b7505e14-3560-4d0d-be2c-d6a7a5953865"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 08:50:50 crc kubenswrapper[4813]: I1201 08:50:50.668658 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b7505e14-3560-4d0d-be2c-d6a7a5953865-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "b7505e14-3560-4d0d-be2c-d6a7a5953865" (UID: "b7505e14-3560-4d0d-be2c-d6a7a5953865"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 08:50:50 crc kubenswrapper[4813]: I1201 08:50:50.671295 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b7505e14-3560-4d0d-be2c-d6a7a5953865-kube-api-access-dgqwb" (OuterVolumeSpecName: "kube-api-access-dgqwb") pod "b7505e14-3560-4d0d-be2c-d6a7a5953865" (UID: "b7505e14-3560-4d0d-be2c-d6a7a5953865"). InnerVolumeSpecName "kube-api-access-dgqwb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:50:50 crc kubenswrapper[4813]: I1201 08:50:50.671812 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b7505e14-3560-4d0d-be2c-d6a7a5953865-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "b7505e14-3560-4d0d-be2c-d6a7a5953865" (UID: "b7505e14-3560-4d0d-be2c-d6a7a5953865"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:50:50 crc kubenswrapper[4813]: I1201 08:50:50.678686 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b7505e14-3560-4d0d-be2c-d6a7a5953865-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "b7505e14-3560-4d0d-be2c-d6a7a5953865" (UID: "b7505e14-3560-4d0d-be2c-d6a7a5953865"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 08:50:50 crc kubenswrapper[4813]: I1201 08:50:50.768435 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/5a004969-7056-4b56-93ad-f93d8736c43f-ovnkube-config\") pod \"ovnkube-node-qnss9\" (UID: \"5a004969-7056-4b56-93ad-f93d8736c43f\") " pod="openshift-ovn-kubernetes/ovnkube-node-qnss9" Dec 01 08:50:50 crc kubenswrapper[4813]: I1201 08:50:50.768702 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/5a004969-7056-4b56-93ad-f93d8736c43f-env-overrides\") pod \"ovnkube-node-qnss9\" (UID: \"5a004969-7056-4b56-93ad-f93d8736c43f\") " pod="openshift-ovn-kubernetes/ovnkube-node-qnss9" Dec 01 08:50:50 crc kubenswrapper[4813]: I1201 08:50:50.768723 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/5a004969-7056-4b56-93ad-f93d8736c43f-log-socket\") pod \"ovnkube-node-qnss9\" (UID: \"5a004969-7056-4b56-93ad-f93d8736c43f\") " pod="openshift-ovn-kubernetes/ovnkube-node-qnss9" Dec 01 08:50:50 crc kubenswrapper[4813]: I1201 08:50:50.768744 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/5a004969-7056-4b56-93ad-f93d8736c43f-host-slash\") pod \"ovnkube-node-qnss9\" (UID: \"5a004969-7056-4b56-93ad-f93d8736c43f\") " pod="openshift-ovn-kubernetes/ovnkube-node-qnss9" Dec 01 08:50:50 crc kubenswrapper[4813]: I1201 08:50:50.768758 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/5a004969-7056-4b56-93ad-f93d8736c43f-systemd-units\") pod \"ovnkube-node-qnss9\" (UID: \"5a004969-7056-4b56-93ad-f93d8736c43f\") " pod="openshift-ovn-kubernetes/ovnkube-node-qnss9" Dec 01 08:50:50 crc kubenswrapper[4813]: I1201 08:50:50.768781 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/5a004969-7056-4b56-93ad-f93d8736c43f-var-lib-openvswitch\") pod \"ovnkube-node-qnss9\" (UID: \"5a004969-7056-4b56-93ad-f93d8736c43f\") " pod="openshift-ovn-kubernetes/ovnkube-node-qnss9" Dec 01 08:50:50 crc kubenswrapper[4813]: I1201 08:50:50.768799 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/5a004969-7056-4b56-93ad-f93d8736c43f-run-openvswitch\") pod \"ovnkube-node-qnss9\" (UID: \"5a004969-7056-4b56-93ad-f93d8736c43f\") " pod="openshift-ovn-kubernetes/ovnkube-node-qnss9" Dec 01 08:50:50 crc kubenswrapper[4813]: I1201 08:50:50.768820 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/5a004969-7056-4b56-93ad-f93d8736c43f-host-run-ovn-kubernetes\") pod \"ovnkube-node-qnss9\" (UID: \"5a004969-7056-4b56-93ad-f93d8736c43f\") " pod="openshift-ovn-kubernetes/ovnkube-node-qnss9" Dec 01 08:50:50 crc kubenswrapper[4813]: I1201 08:50:50.769078 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/5a004969-7056-4b56-93ad-f93d8736c43f-host-kubelet\") pod \"ovnkube-node-qnss9\" (UID: \"5a004969-7056-4b56-93ad-f93d8736c43f\") " pod="openshift-ovn-kubernetes/ovnkube-node-qnss9" Dec 01 08:50:50 crc kubenswrapper[4813]: I1201 08:50:50.769198 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/5a004969-7056-4b56-93ad-f93d8736c43f-run-ovn\") pod \"ovnkube-node-qnss9\" (UID: \"5a004969-7056-4b56-93ad-f93d8736c43f\") " pod="openshift-ovn-kubernetes/ovnkube-node-qnss9" Dec 01 08:50:50 crc kubenswrapper[4813]: I1201 08:50:50.769266 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/5a004969-7056-4b56-93ad-f93d8736c43f-host-cni-bin\") pod \"ovnkube-node-qnss9\" (UID: \"5a004969-7056-4b56-93ad-f93d8736c43f\") " pod="openshift-ovn-kubernetes/ovnkube-node-qnss9" Dec 01 08:50:50 crc kubenswrapper[4813]: I1201 08:50:50.769299 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/5a004969-7056-4b56-93ad-f93d8736c43f-host-run-netns\") pod \"ovnkube-node-qnss9\" (UID: \"5a004969-7056-4b56-93ad-f93d8736c43f\") " pod="openshift-ovn-kubernetes/ovnkube-node-qnss9" Dec 01 08:50:50 crc kubenswrapper[4813]: I1201 08:50:50.769321 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/5a004969-7056-4b56-93ad-f93d8736c43f-host-cni-netd\") pod \"ovnkube-node-qnss9\" (UID: \"5a004969-7056-4b56-93ad-f93d8736c43f\") " pod="openshift-ovn-kubernetes/ovnkube-node-qnss9" Dec 01 08:50:50 crc kubenswrapper[4813]: I1201 08:50:50.769335 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/5a004969-7056-4b56-93ad-f93d8736c43f-etc-openvswitch\") pod \"ovnkube-node-qnss9\" (UID: \"5a004969-7056-4b56-93ad-f93d8736c43f\") " pod="openshift-ovn-kubernetes/ovnkube-node-qnss9" Dec 01 08:50:50 crc kubenswrapper[4813]: I1201 08:50:50.769392 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ntj5p\" (UniqueName: \"kubernetes.io/projected/5a004969-7056-4b56-93ad-f93d8736c43f-kube-api-access-ntj5p\") pod \"ovnkube-node-qnss9\" (UID: \"5a004969-7056-4b56-93ad-f93d8736c43f\") " pod="openshift-ovn-kubernetes/ovnkube-node-qnss9" Dec 01 08:50:50 crc kubenswrapper[4813]: I1201 08:50:50.769486 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/5a004969-7056-4b56-93ad-f93d8736c43f-ovnkube-script-lib\") pod \"ovnkube-node-qnss9\" (UID: \"5a004969-7056-4b56-93ad-f93d8736c43f\") " pod="openshift-ovn-kubernetes/ovnkube-node-qnss9" Dec 01 08:50:50 crc kubenswrapper[4813]: I1201 08:50:50.769533 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/5a004969-7056-4b56-93ad-f93d8736c43f-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-qnss9\" (UID: \"5a004969-7056-4b56-93ad-f93d8736c43f\") " pod="openshift-ovn-kubernetes/ovnkube-node-qnss9" Dec 01 08:50:50 crc kubenswrapper[4813]: I1201 08:50:50.769580 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/5a004969-7056-4b56-93ad-f93d8736c43f-ovn-node-metrics-cert\") pod \"ovnkube-node-qnss9\" (UID: \"5a004969-7056-4b56-93ad-f93d8736c43f\") " pod="openshift-ovn-kubernetes/ovnkube-node-qnss9" Dec 01 08:50:50 crc kubenswrapper[4813]: I1201 08:50:50.769643 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/5a004969-7056-4b56-93ad-f93d8736c43f-node-log\") pod \"ovnkube-node-qnss9\" (UID: \"5a004969-7056-4b56-93ad-f93d8736c43f\") " pod="openshift-ovn-kubernetes/ovnkube-node-qnss9" Dec 01 08:50:50 crc kubenswrapper[4813]: I1201 08:50:50.769714 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/5a004969-7056-4b56-93ad-f93d8736c43f-run-systemd\") pod \"ovnkube-node-qnss9\" (UID: \"5a004969-7056-4b56-93ad-f93d8736c43f\") " pod="openshift-ovn-kubernetes/ovnkube-node-qnss9" Dec 01 08:50:50 crc kubenswrapper[4813]: I1201 08:50:50.769810 4813 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/b7505e14-3560-4d0d-be2c-d6a7a5953865-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Dec 01 08:50:50 crc kubenswrapper[4813]: I1201 08:50:50.769836 4813 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/b7505e14-3560-4d0d-be2c-d6a7a5953865-env-overrides\") on node \"crc\" DevicePath \"\"" Dec 01 08:50:50 crc kubenswrapper[4813]: I1201 08:50:50.769854 4813 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/b7505e14-3560-4d0d-be2c-d6a7a5953865-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Dec 01 08:50:50 crc kubenswrapper[4813]: I1201 08:50:50.769871 4813 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/b7505e14-3560-4d0d-be2c-d6a7a5953865-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Dec 01 08:50:50 crc kubenswrapper[4813]: I1201 08:50:50.769907 4813 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/b7505e14-3560-4d0d-be2c-d6a7a5953865-systemd-units\") on node \"crc\" DevicePath \"\"" Dec 01 08:50:50 crc kubenswrapper[4813]: I1201 08:50:50.769923 4813 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/b7505e14-3560-4d0d-be2c-d6a7a5953865-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Dec 01 08:50:50 crc kubenswrapper[4813]: I1201 08:50:50.769941 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dgqwb\" (UniqueName: \"kubernetes.io/projected/b7505e14-3560-4d0d-be2c-d6a7a5953865-kube-api-access-dgqwb\") on node \"crc\" DevicePath \"\"" Dec 01 08:50:50 crc kubenswrapper[4813]: I1201 08:50:50.769957 4813 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/b7505e14-3560-4d0d-be2c-d6a7a5953865-run-systemd\") on node \"crc\" DevicePath \"\"" Dec 01 08:50:50 crc kubenswrapper[4813]: I1201 08:50:50.769996 4813 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/b7505e14-3560-4d0d-be2c-d6a7a5953865-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Dec 01 08:50:50 crc kubenswrapper[4813]: I1201 08:50:50.770011 4813 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/b7505e14-3560-4d0d-be2c-d6a7a5953865-ovnkube-config\") on node \"crc\" DevicePath \"\"" Dec 01 08:50:50 crc kubenswrapper[4813]: I1201 08:50:50.770025 4813 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/b7505e14-3560-4d0d-be2c-d6a7a5953865-run-openvswitch\") on node \"crc\" DevicePath \"\"" Dec 01 08:50:50 crc kubenswrapper[4813]: I1201 08:50:50.870829 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/5a004969-7056-4b56-93ad-f93d8736c43f-ovn-node-metrics-cert\") pod \"ovnkube-node-qnss9\" (UID: \"5a004969-7056-4b56-93ad-f93d8736c43f\") " pod="openshift-ovn-kubernetes/ovnkube-node-qnss9" Dec 01 08:50:50 crc kubenswrapper[4813]: I1201 08:50:50.870911 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/5a004969-7056-4b56-93ad-f93d8736c43f-node-log\") pod \"ovnkube-node-qnss9\" (UID: \"5a004969-7056-4b56-93ad-f93d8736c43f\") " pod="openshift-ovn-kubernetes/ovnkube-node-qnss9" Dec 01 08:50:50 crc kubenswrapper[4813]: I1201 08:50:50.870951 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/5a004969-7056-4b56-93ad-f93d8736c43f-run-systemd\") pod \"ovnkube-node-qnss9\" (UID: \"5a004969-7056-4b56-93ad-f93d8736c43f\") " pod="openshift-ovn-kubernetes/ovnkube-node-qnss9" Dec 01 08:50:50 crc kubenswrapper[4813]: I1201 08:50:50.871017 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/5a004969-7056-4b56-93ad-f93d8736c43f-ovnkube-config\") pod \"ovnkube-node-qnss9\" (UID: \"5a004969-7056-4b56-93ad-f93d8736c43f\") " pod="openshift-ovn-kubernetes/ovnkube-node-qnss9" Dec 01 08:50:50 crc kubenswrapper[4813]: I1201 08:50:50.871056 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/5a004969-7056-4b56-93ad-f93d8736c43f-env-overrides\") pod \"ovnkube-node-qnss9\" (UID: \"5a004969-7056-4b56-93ad-f93d8736c43f\") " pod="openshift-ovn-kubernetes/ovnkube-node-qnss9" Dec 01 08:50:50 crc kubenswrapper[4813]: I1201 08:50:50.871086 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/5a004969-7056-4b56-93ad-f93d8736c43f-log-socket\") pod \"ovnkube-node-qnss9\" (UID: \"5a004969-7056-4b56-93ad-f93d8736c43f\") " pod="openshift-ovn-kubernetes/ovnkube-node-qnss9" Dec 01 08:50:50 crc kubenswrapper[4813]: I1201 08:50:50.871127 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/5a004969-7056-4b56-93ad-f93d8736c43f-host-slash\") pod \"ovnkube-node-qnss9\" (UID: \"5a004969-7056-4b56-93ad-f93d8736c43f\") " pod="openshift-ovn-kubernetes/ovnkube-node-qnss9" Dec 01 08:50:50 crc kubenswrapper[4813]: I1201 08:50:50.871172 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/5a004969-7056-4b56-93ad-f93d8736c43f-systemd-units\") pod \"ovnkube-node-qnss9\" (UID: \"5a004969-7056-4b56-93ad-f93d8736c43f\") " pod="openshift-ovn-kubernetes/ovnkube-node-qnss9" Dec 01 08:50:50 crc kubenswrapper[4813]: I1201 08:50:50.871221 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/5a004969-7056-4b56-93ad-f93d8736c43f-var-lib-openvswitch\") pod \"ovnkube-node-qnss9\" (UID: \"5a004969-7056-4b56-93ad-f93d8736c43f\") " pod="openshift-ovn-kubernetes/ovnkube-node-qnss9" Dec 01 08:50:50 crc kubenswrapper[4813]: I1201 08:50:50.871255 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/5a004969-7056-4b56-93ad-f93d8736c43f-run-openvswitch\") pod \"ovnkube-node-qnss9\" (UID: \"5a004969-7056-4b56-93ad-f93d8736c43f\") " pod="openshift-ovn-kubernetes/ovnkube-node-qnss9" Dec 01 08:50:50 crc kubenswrapper[4813]: I1201 08:50:50.871288 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/5a004969-7056-4b56-93ad-f93d8736c43f-host-run-ovn-kubernetes\") pod \"ovnkube-node-qnss9\" (UID: \"5a004969-7056-4b56-93ad-f93d8736c43f\") " pod="openshift-ovn-kubernetes/ovnkube-node-qnss9" Dec 01 08:50:50 crc kubenswrapper[4813]: I1201 08:50:50.871348 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/5a004969-7056-4b56-93ad-f93d8736c43f-host-kubelet\") pod \"ovnkube-node-qnss9\" (UID: \"5a004969-7056-4b56-93ad-f93d8736c43f\") " pod="openshift-ovn-kubernetes/ovnkube-node-qnss9" Dec 01 08:50:50 crc kubenswrapper[4813]: I1201 08:50:50.871380 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/5a004969-7056-4b56-93ad-f93d8736c43f-run-ovn\") pod \"ovnkube-node-qnss9\" (UID: \"5a004969-7056-4b56-93ad-f93d8736c43f\") " pod="openshift-ovn-kubernetes/ovnkube-node-qnss9" Dec 01 08:50:50 crc kubenswrapper[4813]: I1201 08:50:50.871586 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/5a004969-7056-4b56-93ad-f93d8736c43f-host-slash\") pod \"ovnkube-node-qnss9\" (UID: \"5a004969-7056-4b56-93ad-f93d8736c43f\") " pod="openshift-ovn-kubernetes/ovnkube-node-qnss9" Dec 01 08:50:50 crc kubenswrapper[4813]: I1201 08:50:50.871428 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/5a004969-7056-4b56-93ad-f93d8736c43f-host-cni-bin\") pod \"ovnkube-node-qnss9\" (UID: \"5a004969-7056-4b56-93ad-f93d8736c43f\") " pod="openshift-ovn-kubernetes/ovnkube-node-qnss9" Dec 01 08:50:50 crc kubenswrapper[4813]: I1201 08:50:50.871633 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/5a004969-7056-4b56-93ad-f93d8736c43f-node-log\") pod \"ovnkube-node-qnss9\" (UID: \"5a004969-7056-4b56-93ad-f93d8736c43f\") " pod="openshift-ovn-kubernetes/ovnkube-node-qnss9" Dec 01 08:50:50 crc kubenswrapper[4813]: I1201 08:50:50.871656 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/5a004969-7056-4b56-93ad-f93d8736c43f-run-systemd\") pod \"ovnkube-node-qnss9\" (UID: \"5a004969-7056-4b56-93ad-f93d8736c43f\") " pod="openshift-ovn-kubernetes/ovnkube-node-qnss9" Dec 01 08:50:50 crc kubenswrapper[4813]: I1201 08:50:50.871674 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/5a004969-7056-4b56-93ad-f93d8736c43f-host-run-netns\") pod \"ovnkube-node-qnss9\" (UID: \"5a004969-7056-4b56-93ad-f93d8736c43f\") " pod="openshift-ovn-kubernetes/ovnkube-node-qnss9" Dec 01 08:50:50 crc kubenswrapper[4813]: I1201 08:50:50.871713 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/5a004969-7056-4b56-93ad-f93d8736c43f-host-cni-netd\") pod \"ovnkube-node-qnss9\" (UID: \"5a004969-7056-4b56-93ad-f93d8736c43f\") " pod="openshift-ovn-kubernetes/ovnkube-node-qnss9" Dec 01 08:50:50 crc kubenswrapper[4813]: I1201 08:50:50.871744 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/5a004969-7056-4b56-93ad-f93d8736c43f-etc-openvswitch\") pod \"ovnkube-node-qnss9\" (UID: \"5a004969-7056-4b56-93ad-f93d8736c43f\") " pod="openshift-ovn-kubernetes/ovnkube-node-qnss9" Dec 01 08:50:50 crc kubenswrapper[4813]: I1201 08:50:50.871778 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ntj5p\" (UniqueName: \"kubernetes.io/projected/5a004969-7056-4b56-93ad-f93d8736c43f-kube-api-access-ntj5p\") pod \"ovnkube-node-qnss9\" (UID: \"5a004969-7056-4b56-93ad-f93d8736c43f\") " pod="openshift-ovn-kubernetes/ovnkube-node-qnss9" Dec 01 08:50:50 crc kubenswrapper[4813]: I1201 08:50:50.871822 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/5a004969-7056-4b56-93ad-f93d8736c43f-ovnkube-script-lib\") pod \"ovnkube-node-qnss9\" (UID: \"5a004969-7056-4b56-93ad-f93d8736c43f\") " pod="openshift-ovn-kubernetes/ovnkube-node-qnss9" Dec 01 08:50:50 crc kubenswrapper[4813]: I1201 08:50:50.871873 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/5a004969-7056-4b56-93ad-f93d8736c43f-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-qnss9\" (UID: \"5a004969-7056-4b56-93ad-f93d8736c43f\") " pod="openshift-ovn-kubernetes/ovnkube-node-qnss9" Dec 01 08:50:50 crc kubenswrapper[4813]: I1201 08:50:50.872012 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/5a004969-7056-4b56-93ad-f93d8736c43f-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-qnss9\" (UID: \"5a004969-7056-4b56-93ad-f93d8736c43f\") " pod="openshift-ovn-kubernetes/ovnkube-node-qnss9" Dec 01 08:50:50 crc kubenswrapper[4813]: I1201 08:50:50.872101 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/5a004969-7056-4b56-93ad-f93d8736c43f-systemd-units\") pod \"ovnkube-node-qnss9\" (UID: \"5a004969-7056-4b56-93ad-f93d8736c43f\") " pod="openshift-ovn-kubernetes/ovnkube-node-qnss9" Dec 01 08:50:50 crc kubenswrapper[4813]: I1201 08:50:50.872146 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/5a004969-7056-4b56-93ad-f93d8736c43f-var-lib-openvswitch\") pod \"ovnkube-node-qnss9\" (UID: \"5a004969-7056-4b56-93ad-f93d8736c43f\") " pod="openshift-ovn-kubernetes/ovnkube-node-qnss9" Dec 01 08:50:50 crc kubenswrapper[4813]: I1201 08:50:50.872187 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/5a004969-7056-4b56-93ad-f93d8736c43f-run-openvswitch\") pod \"ovnkube-node-qnss9\" (UID: \"5a004969-7056-4b56-93ad-f93d8736c43f\") " pod="openshift-ovn-kubernetes/ovnkube-node-qnss9" Dec 01 08:50:50 crc kubenswrapper[4813]: I1201 08:50:50.872232 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/5a004969-7056-4b56-93ad-f93d8736c43f-host-run-ovn-kubernetes\") pod \"ovnkube-node-qnss9\" (UID: \"5a004969-7056-4b56-93ad-f93d8736c43f\") " pod="openshift-ovn-kubernetes/ovnkube-node-qnss9" Dec 01 08:50:50 crc kubenswrapper[4813]: I1201 08:50:50.872272 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/5a004969-7056-4b56-93ad-f93d8736c43f-host-kubelet\") pod \"ovnkube-node-qnss9\" (UID: \"5a004969-7056-4b56-93ad-f93d8736c43f\") " pod="openshift-ovn-kubernetes/ovnkube-node-qnss9" Dec 01 08:50:50 crc kubenswrapper[4813]: I1201 08:50:50.872345 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/5a004969-7056-4b56-93ad-f93d8736c43f-run-ovn\") pod \"ovnkube-node-qnss9\" (UID: \"5a004969-7056-4b56-93ad-f93d8736c43f\") " pod="openshift-ovn-kubernetes/ovnkube-node-qnss9" Dec 01 08:50:50 crc kubenswrapper[4813]: I1201 08:50:50.872389 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/5a004969-7056-4b56-93ad-f93d8736c43f-host-cni-bin\") pod \"ovnkube-node-qnss9\" (UID: \"5a004969-7056-4b56-93ad-f93d8736c43f\") " pod="openshift-ovn-kubernetes/ovnkube-node-qnss9" Dec 01 08:50:50 crc kubenswrapper[4813]: I1201 08:50:50.872404 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/5a004969-7056-4b56-93ad-f93d8736c43f-ovnkube-config\") pod \"ovnkube-node-qnss9\" (UID: \"5a004969-7056-4b56-93ad-f93d8736c43f\") " pod="openshift-ovn-kubernetes/ovnkube-node-qnss9" Dec 01 08:50:50 crc kubenswrapper[4813]: I1201 08:50:50.872429 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/5a004969-7056-4b56-93ad-f93d8736c43f-host-run-netns\") pod \"ovnkube-node-qnss9\" (UID: \"5a004969-7056-4b56-93ad-f93d8736c43f\") " pod="openshift-ovn-kubernetes/ovnkube-node-qnss9" Dec 01 08:50:50 crc kubenswrapper[4813]: I1201 08:50:50.872470 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/5a004969-7056-4b56-93ad-f93d8736c43f-host-cni-netd\") pod \"ovnkube-node-qnss9\" (UID: \"5a004969-7056-4b56-93ad-f93d8736c43f\") " pod="openshift-ovn-kubernetes/ovnkube-node-qnss9" Dec 01 08:50:50 crc kubenswrapper[4813]: I1201 08:50:50.872552 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/5a004969-7056-4b56-93ad-f93d8736c43f-etc-openvswitch\") pod \"ovnkube-node-qnss9\" (UID: \"5a004969-7056-4b56-93ad-f93d8736c43f\") " pod="openshift-ovn-kubernetes/ovnkube-node-qnss9" Dec 01 08:50:50 crc kubenswrapper[4813]: I1201 08:50:50.872834 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/5a004969-7056-4b56-93ad-f93d8736c43f-env-overrides\") pod \"ovnkube-node-qnss9\" (UID: \"5a004969-7056-4b56-93ad-f93d8736c43f\") " pod="openshift-ovn-kubernetes/ovnkube-node-qnss9" Dec 01 08:50:50 crc kubenswrapper[4813]: I1201 08:50:50.873349 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/5a004969-7056-4b56-93ad-f93d8736c43f-log-socket\") pod \"ovnkube-node-qnss9\" (UID: \"5a004969-7056-4b56-93ad-f93d8736c43f\") " pod="openshift-ovn-kubernetes/ovnkube-node-qnss9" Dec 01 08:50:50 crc kubenswrapper[4813]: I1201 08:50:50.874026 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/5a004969-7056-4b56-93ad-f93d8736c43f-ovnkube-script-lib\") pod \"ovnkube-node-qnss9\" (UID: \"5a004969-7056-4b56-93ad-f93d8736c43f\") " pod="openshift-ovn-kubernetes/ovnkube-node-qnss9" Dec 01 08:50:50 crc kubenswrapper[4813]: I1201 08:50:50.890877 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/5a004969-7056-4b56-93ad-f93d8736c43f-ovn-node-metrics-cert\") pod \"ovnkube-node-qnss9\" (UID: \"5a004969-7056-4b56-93ad-f93d8736c43f\") " pod="openshift-ovn-kubernetes/ovnkube-node-qnss9" Dec 01 08:50:50 crc kubenswrapper[4813]: I1201 08:50:50.910611 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ntj5p\" (UniqueName: \"kubernetes.io/projected/5a004969-7056-4b56-93ad-f93d8736c43f-kube-api-access-ntj5p\") pod \"ovnkube-node-qnss9\" (UID: \"5a004969-7056-4b56-93ad-f93d8736c43f\") " pod="openshift-ovn-kubernetes/ovnkube-node-qnss9" Dec 01 08:50:50 crc kubenswrapper[4813]: I1201 08:50:50.932943 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-qnss9" Dec 01 08:50:51 crc kubenswrapper[4813]: I1201 08:50:51.403182 4813 generic.go:334] "Generic (PLEG): container finished" podID="5a004969-7056-4b56-93ad-f93d8736c43f" containerID="315193de2b8e25bf214ac232152c88b9e8e852fc26048c44637da595a2346a54" exitCode=0 Dec 01 08:50:51 crc kubenswrapper[4813]: I1201 08:50:51.403280 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qnss9" event={"ID":"5a004969-7056-4b56-93ad-f93d8736c43f","Type":"ContainerDied","Data":"315193de2b8e25bf214ac232152c88b9e8e852fc26048c44637da595a2346a54"} Dec 01 08:50:51 crc kubenswrapper[4813]: I1201 08:50:51.403591 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qnss9" event={"ID":"5a004969-7056-4b56-93ad-f93d8736c43f","Type":"ContainerStarted","Data":"db89a3ef30296074be982f19d4c287e0ca80f95c5f01748ee23c68c6b92006c3"} Dec 01 08:50:51 crc kubenswrapper[4813]: I1201 08:50:51.412770 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-j4pcc_b7505e14-3560-4d0d-be2c-d6a7a5953865/ovn-acl-logging/0.log" Dec 01 08:50:51 crc kubenswrapper[4813]: I1201 08:50:51.413840 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-j4pcc_b7505e14-3560-4d0d-be2c-d6a7a5953865/ovn-controller/0.log" Dec 01 08:50:51 crc kubenswrapper[4813]: I1201 08:50:51.414325 4813 generic.go:334] "Generic (PLEG): container finished" podID="b7505e14-3560-4d0d-be2c-d6a7a5953865" containerID="b45454b440c5108fa150a6b856afa1778ba6e9f89633433cb3117b68f00f72c3" exitCode=0 Dec 01 08:50:51 crc kubenswrapper[4813]: I1201 08:50:51.414444 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-j4pcc" Dec 01 08:50:51 crc kubenswrapper[4813]: I1201 08:50:51.414445 4813 generic.go:334] "Generic (PLEG): container finished" podID="b7505e14-3560-4d0d-be2c-d6a7a5953865" containerID="dc32d930ab23f30103d8fa42291909cbec89ebb75c5c53cd783cd27fe1a3f09e" exitCode=0 Dec 01 08:50:51 crc kubenswrapper[4813]: I1201 08:50:51.414626 4813 generic.go:334] "Generic (PLEG): container finished" podID="b7505e14-3560-4d0d-be2c-d6a7a5953865" containerID="7f19c67b1c6c80c7f3223512102810d46ede1445464d21ae3bbea83424539b2b" exitCode=0 Dec 01 08:50:51 crc kubenswrapper[4813]: I1201 08:50:51.414398 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-j4pcc" event={"ID":"b7505e14-3560-4d0d-be2c-d6a7a5953865","Type":"ContainerDied","Data":"b45454b440c5108fa150a6b856afa1778ba6e9f89633433cb3117b68f00f72c3"} Dec 01 08:50:51 crc kubenswrapper[4813]: I1201 08:50:51.414988 4813 scope.go:117] "RemoveContainer" containerID="6aa74106386539915851c80789dac37ef865908f24cb6a67ffa406c2599cb5d7" Dec 01 08:50:51 crc kubenswrapper[4813]: I1201 08:50:51.415026 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-j4pcc" event={"ID":"b7505e14-3560-4d0d-be2c-d6a7a5953865","Type":"ContainerDied","Data":"dc32d930ab23f30103d8fa42291909cbec89ebb75c5c53cd783cd27fe1a3f09e"} Dec 01 08:50:51 crc kubenswrapper[4813]: I1201 08:50:51.415204 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-j4pcc" event={"ID":"b7505e14-3560-4d0d-be2c-d6a7a5953865","Type":"ContainerDied","Data":"7f19c67b1c6c80c7f3223512102810d46ede1445464d21ae3bbea83424539b2b"} Dec 01 08:50:51 crc kubenswrapper[4813]: I1201 08:50:51.415347 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-j4pcc" event={"ID":"b7505e14-3560-4d0d-be2c-d6a7a5953865","Type":"ContainerDied","Data":"4418f2beb2999b7488df1221f849f6a9c01c3a4439ea419a6462443d759f7309"} Dec 01 08:50:51 crc kubenswrapper[4813]: I1201 08:50:51.419274 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-b6qqv_b9656a03-0096-4c89-b0da-e37e4103611b/kube-multus/1.log" Dec 01 08:50:51 crc kubenswrapper[4813]: I1201 08:50:51.440887 4813 scope.go:117] "RemoveContainer" containerID="b45454b440c5108fa150a6b856afa1778ba6e9f89633433cb3117b68f00f72c3" Dec 01 08:50:51 crc kubenswrapper[4813]: I1201 08:50:51.466581 4813 scope.go:117] "RemoveContainer" containerID="dc32d930ab23f30103d8fa42291909cbec89ebb75c5c53cd783cd27fe1a3f09e" Dec 01 08:50:51 crc kubenswrapper[4813]: I1201 08:50:51.503670 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-j4pcc"] Dec 01 08:50:51 crc kubenswrapper[4813]: I1201 08:50:51.509482 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-j4pcc"] Dec 01 08:50:51 crc kubenswrapper[4813]: I1201 08:50:51.513403 4813 scope.go:117] "RemoveContainer" containerID="7f19c67b1c6c80c7f3223512102810d46ede1445464d21ae3bbea83424539b2b" Dec 01 08:50:51 crc kubenswrapper[4813]: I1201 08:50:51.534632 4813 scope.go:117] "RemoveContainer" containerID="944977e9722025088720107530d0c91fc693593f380e4b49678f28a56da180cc" Dec 01 08:50:51 crc kubenswrapper[4813]: I1201 08:50:51.562665 4813 scope.go:117] "RemoveContainer" containerID="ccc30a63dc1c23ff9880fb04af1f31f0656b2696d359988e23deed23989f480a" Dec 01 08:50:51 crc kubenswrapper[4813]: I1201 08:50:51.585871 4813 scope.go:117] "RemoveContainer" containerID="9d984c90302e57f3e5a6f3ca3613b415db04ac74a5a228582f1b52ef1d59f93e" Dec 01 08:50:51 crc kubenswrapper[4813]: I1201 08:50:51.601695 4813 scope.go:117] "RemoveContainer" containerID="ce8963bf1e9ee66926907aaf3a03d18977f4d0a1983d7121b62fbc8e02a37e4d" Dec 01 08:50:51 crc kubenswrapper[4813]: I1201 08:50:51.624852 4813 scope.go:117] "RemoveContainer" containerID="dd67634e0c806323334495401a4e20b48fea4bc5013acd973d2abd3d70870f99" Dec 01 08:50:51 crc kubenswrapper[4813]: I1201 08:50:51.675168 4813 scope.go:117] "RemoveContainer" containerID="6aa74106386539915851c80789dac37ef865908f24cb6a67ffa406c2599cb5d7" Dec 01 08:50:51 crc kubenswrapper[4813]: E1201 08:50:51.675961 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6aa74106386539915851c80789dac37ef865908f24cb6a67ffa406c2599cb5d7\": container with ID starting with 6aa74106386539915851c80789dac37ef865908f24cb6a67ffa406c2599cb5d7 not found: ID does not exist" containerID="6aa74106386539915851c80789dac37ef865908f24cb6a67ffa406c2599cb5d7" Dec 01 08:50:51 crc kubenswrapper[4813]: I1201 08:50:51.676128 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6aa74106386539915851c80789dac37ef865908f24cb6a67ffa406c2599cb5d7"} err="failed to get container status \"6aa74106386539915851c80789dac37ef865908f24cb6a67ffa406c2599cb5d7\": rpc error: code = NotFound desc = could not find container \"6aa74106386539915851c80789dac37ef865908f24cb6a67ffa406c2599cb5d7\": container with ID starting with 6aa74106386539915851c80789dac37ef865908f24cb6a67ffa406c2599cb5d7 not found: ID does not exist" Dec 01 08:50:51 crc kubenswrapper[4813]: I1201 08:50:51.676202 4813 scope.go:117] "RemoveContainer" containerID="b45454b440c5108fa150a6b856afa1778ba6e9f89633433cb3117b68f00f72c3" Dec 01 08:50:51 crc kubenswrapper[4813]: E1201 08:50:51.677122 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b45454b440c5108fa150a6b856afa1778ba6e9f89633433cb3117b68f00f72c3\": container with ID starting with b45454b440c5108fa150a6b856afa1778ba6e9f89633433cb3117b68f00f72c3 not found: ID does not exist" containerID="b45454b440c5108fa150a6b856afa1778ba6e9f89633433cb3117b68f00f72c3" Dec 01 08:50:51 crc kubenswrapper[4813]: I1201 08:50:51.677176 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b45454b440c5108fa150a6b856afa1778ba6e9f89633433cb3117b68f00f72c3"} err="failed to get container status \"b45454b440c5108fa150a6b856afa1778ba6e9f89633433cb3117b68f00f72c3\": rpc error: code = NotFound desc = could not find container \"b45454b440c5108fa150a6b856afa1778ba6e9f89633433cb3117b68f00f72c3\": container with ID starting with b45454b440c5108fa150a6b856afa1778ba6e9f89633433cb3117b68f00f72c3 not found: ID does not exist" Dec 01 08:50:51 crc kubenswrapper[4813]: I1201 08:50:51.677214 4813 scope.go:117] "RemoveContainer" containerID="dc32d930ab23f30103d8fa42291909cbec89ebb75c5c53cd783cd27fe1a3f09e" Dec 01 08:50:51 crc kubenswrapper[4813]: E1201 08:50:51.677832 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dc32d930ab23f30103d8fa42291909cbec89ebb75c5c53cd783cd27fe1a3f09e\": container with ID starting with dc32d930ab23f30103d8fa42291909cbec89ebb75c5c53cd783cd27fe1a3f09e not found: ID does not exist" containerID="dc32d930ab23f30103d8fa42291909cbec89ebb75c5c53cd783cd27fe1a3f09e" Dec 01 08:50:51 crc kubenswrapper[4813]: I1201 08:50:51.677888 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dc32d930ab23f30103d8fa42291909cbec89ebb75c5c53cd783cd27fe1a3f09e"} err="failed to get container status \"dc32d930ab23f30103d8fa42291909cbec89ebb75c5c53cd783cd27fe1a3f09e\": rpc error: code = NotFound desc = could not find container \"dc32d930ab23f30103d8fa42291909cbec89ebb75c5c53cd783cd27fe1a3f09e\": container with ID starting with dc32d930ab23f30103d8fa42291909cbec89ebb75c5c53cd783cd27fe1a3f09e not found: ID does not exist" Dec 01 08:50:51 crc kubenswrapper[4813]: I1201 08:50:51.677914 4813 scope.go:117] "RemoveContainer" containerID="7f19c67b1c6c80c7f3223512102810d46ede1445464d21ae3bbea83424539b2b" Dec 01 08:50:51 crc kubenswrapper[4813]: E1201 08:50:51.678487 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7f19c67b1c6c80c7f3223512102810d46ede1445464d21ae3bbea83424539b2b\": container with ID starting with 7f19c67b1c6c80c7f3223512102810d46ede1445464d21ae3bbea83424539b2b not found: ID does not exist" containerID="7f19c67b1c6c80c7f3223512102810d46ede1445464d21ae3bbea83424539b2b" Dec 01 08:50:51 crc kubenswrapper[4813]: I1201 08:50:51.678533 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7f19c67b1c6c80c7f3223512102810d46ede1445464d21ae3bbea83424539b2b"} err="failed to get container status \"7f19c67b1c6c80c7f3223512102810d46ede1445464d21ae3bbea83424539b2b\": rpc error: code = NotFound desc = could not find container \"7f19c67b1c6c80c7f3223512102810d46ede1445464d21ae3bbea83424539b2b\": container with ID starting with 7f19c67b1c6c80c7f3223512102810d46ede1445464d21ae3bbea83424539b2b not found: ID does not exist" Dec 01 08:50:51 crc kubenswrapper[4813]: I1201 08:50:51.678557 4813 scope.go:117] "RemoveContainer" containerID="944977e9722025088720107530d0c91fc693593f380e4b49678f28a56da180cc" Dec 01 08:50:51 crc kubenswrapper[4813]: E1201 08:50:51.679002 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"944977e9722025088720107530d0c91fc693593f380e4b49678f28a56da180cc\": container with ID starting with 944977e9722025088720107530d0c91fc693593f380e4b49678f28a56da180cc not found: ID does not exist" containerID="944977e9722025088720107530d0c91fc693593f380e4b49678f28a56da180cc" Dec 01 08:50:51 crc kubenswrapper[4813]: I1201 08:50:51.679085 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"944977e9722025088720107530d0c91fc693593f380e4b49678f28a56da180cc"} err="failed to get container status \"944977e9722025088720107530d0c91fc693593f380e4b49678f28a56da180cc\": rpc error: code = NotFound desc = could not find container \"944977e9722025088720107530d0c91fc693593f380e4b49678f28a56da180cc\": container with ID starting with 944977e9722025088720107530d0c91fc693593f380e4b49678f28a56da180cc not found: ID does not exist" Dec 01 08:50:51 crc kubenswrapper[4813]: I1201 08:50:51.679134 4813 scope.go:117] "RemoveContainer" containerID="ccc30a63dc1c23ff9880fb04af1f31f0656b2696d359988e23deed23989f480a" Dec 01 08:50:51 crc kubenswrapper[4813]: E1201 08:50:51.679471 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ccc30a63dc1c23ff9880fb04af1f31f0656b2696d359988e23deed23989f480a\": container with ID starting with ccc30a63dc1c23ff9880fb04af1f31f0656b2696d359988e23deed23989f480a not found: ID does not exist" containerID="ccc30a63dc1c23ff9880fb04af1f31f0656b2696d359988e23deed23989f480a" Dec 01 08:50:51 crc kubenswrapper[4813]: I1201 08:50:51.679507 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ccc30a63dc1c23ff9880fb04af1f31f0656b2696d359988e23deed23989f480a"} err="failed to get container status \"ccc30a63dc1c23ff9880fb04af1f31f0656b2696d359988e23deed23989f480a\": rpc error: code = NotFound desc = could not find container \"ccc30a63dc1c23ff9880fb04af1f31f0656b2696d359988e23deed23989f480a\": container with ID starting with ccc30a63dc1c23ff9880fb04af1f31f0656b2696d359988e23deed23989f480a not found: ID does not exist" Dec 01 08:50:51 crc kubenswrapper[4813]: I1201 08:50:51.679531 4813 scope.go:117] "RemoveContainer" containerID="9d984c90302e57f3e5a6f3ca3613b415db04ac74a5a228582f1b52ef1d59f93e" Dec 01 08:50:51 crc kubenswrapper[4813]: E1201 08:50:51.679928 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9d984c90302e57f3e5a6f3ca3613b415db04ac74a5a228582f1b52ef1d59f93e\": container with ID starting with 9d984c90302e57f3e5a6f3ca3613b415db04ac74a5a228582f1b52ef1d59f93e not found: ID does not exist" containerID="9d984c90302e57f3e5a6f3ca3613b415db04ac74a5a228582f1b52ef1d59f93e" Dec 01 08:50:51 crc kubenswrapper[4813]: I1201 08:50:51.680077 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9d984c90302e57f3e5a6f3ca3613b415db04ac74a5a228582f1b52ef1d59f93e"} err="failed to get container status \"9d984c90302e57f3e5a6f3ca3613b415db04ac74a5a228582f1b52ef1d59f93e\": rpc error: code = NotFound desc = could not find container \"9d984c90302e57f3e5a6f3ca3613b415db04ac74a5a228582f1b52ef1d59f93e\": container with ID starting with 9d984c90302e57f3e5a6f3ca3613b415db04ac74a5a228582f1b52ef1d59f93e not found: ID does not exist" Dec 01 08:50:51 crc kubenswrapper[4813]: I1201 08:50:51.680102 4813 scope.go:117] "RemoveContainer" containerID="ce8963bf1e9ee66926907aaf3a03d18977f4d0a1983d7121b62fbc8e02a37e4d" Dec 01 08:50:51 crc kubenswrapper[4813]: E1201 08:50:51.680399 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ce8963bf1e9ee66926907aaf3a03d18977f4d0a1983d7121b62fbc8e02a37e4d\": container with ID starting with ce8963bf1e9ee66926907aaf3a03d18977f4d0a1983d7121b62fbc8e02a37e4d not found: ID does not exist" containerID="ce8963bf1e9ee66926907aaf3a03d18977f4d0a1983d7121b62fbc8e02a37e4d" Dec 01 08:50:51 crc kubenswrapper[4813]: I1201 08:50:51.680436 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ce8963bf1e9ee66926907aaf3a03d18977f4d0a1983d7121b62fbc8e02a37e4d"} err="failed to get container status \"ce8963bf1e9ee66926907aaf3a03d18977f4d0a1983d7121b62fbc8e02a37e4d\": rpc error: code = NotFound desc = could not find container \"ce8963bf1e9ee66926907aaf3a03d18977f4d0a1983d7121b62fbc8e02a37e4d\": container with ID starting with ce8963bf1e9ee66926907aaf3a03d18977f4d0a1983d7121b62fbc8e02a37e4d not found: ID does not exist" Dec 01 08:50:51 crc kubenswrapper[4813]: I1201 08:50:51.680461 4813 scope.go:117] "RemoveContainer" containerID="dd67634e0c806323334495401a4e20b48fea4bc5013acd973d2abd3d70870f99" Dec 01 08:50:51 crc kubenswrapper[4813]: E1201 08:50:51.680832 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dd67634e0c806323334495401a4e20b48fea4bc5013acd973d2abd3d70870f99\": container with ID starting with dd67634e0c806323334495401a4e20b48fea4bc5013acd973d2abd3d70870f99 not found: ID does not exist" containerID="dd67634e0c806323334495401a4e20b48fea4bc5013acd973d2abd3d70870f99" Dec 01 08:50:51 crc kubenswrapper[4813]: I1201 08:50:51.680887 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dd67634e0c806323334495401a4e20b48fea4bc5013acd973d2abd3d70870f99"} err="failed to get container status \"dd67634e0c806323334495401a4e20b48fea4bc5013acd973d2abd3d70870f99\": rpc error: code = NotFound desc = could not find container \"dd67634e0c806323334495401a4e20b48fea4bc5013acd973d2abd3d70870f99\": container with ID starting with dd67634e0c806323334495401a4e20b48fea4bc5013acd973d2abd3d70870f99 not found: ID does not exist" Dec 01 08:50:51 crc kubenswrapper[4813]: I1201 08:50:51.680920 4813 scope.go:117] "RemoveContainer" containerID="6aa74106386539915851c80789dac37ef865908f24cb6a67ffa406c2599cb5d7" Dec 01 08:50:51 crc kubenswrapper[4813]: I1201 08:50:51.681325 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6aa74106386539915851c80789dac37ef865908f24cb6a67ffa406c2599cb5d7"} err="failed to get container status \"6aa74106386539915851c80789dac37ef865908f24cb6a67ffa406c2599cb5d7\": rpc error: code = NotFound desc = could not find container \"6aa74106386539915851c80789dac37ef865908f24cb6a67ffa406c2599cb5d7\": container with ID starting with 6aa74106386539915851c80789dac37ef865908f24cb6a67ffa406c2599cb5d7 not found: ID does not exist" Dec 01 08:50:51 crc kubenswrapper[4813]: I1201 08:50:51.681355 4813 scope.go:117] "RemoveContainer" containerID="b45454b440c5108fa150a6b856afa1778ba6e9f89633433cb3117b68f00f72c3" Dec 01 08:50:51 crc kubenswrapper[4813]: I1201 08:50:51.681744 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b45454b440c5108fa150a6b856afa1778ba6e9f89633433cb3117b68f00f72c3"} err="failed to get container status \"b45454b440c5108fa150a6b856afa1778ba6e9f89633433cb3117b68f00f72c3\": rpc error: code = NotFound desc = could not find container \"b45454b440c5108fa150a6b856afa1778ba6e9f89633433cb3117b68f00f72c3\": container with ID starting with b45454b440c5108fa150a6b856afa1778ba6e9f89633433cb3117b68f00f72c3 not found: ID does not exist" Dec 01 08:50:51 crc kubenswrapper[4813]: I1201 08:50:51.681863 4813 scope.go:117] "RemoveContainer" containerID="dc32d930ab23f30103d8fa42291909cbec89ebb75c5c53cd783cd27fe1a3f09e" Dec 01 08:50:51 crc kubenswrapper[4813]: I1201 08:50:51.682297 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dc32d930ab23f30103d8fa42291909cbec89ebb75c5c53cd783cd27fe1a3f09e"} err="failed to get container status \"dc32d930ab23f30103d8fa42291909cbec89ebb75c5c53cd783cd27fe1a3f09e\": rpc error: code = NotFound desc = could not find container \"dc32d930ab23f30103d8fa42291909cbec89ebb75c5c53cd783cd27fe1a3f09e\": container with ID starting with dc32d930ab23f30103d8fa42291909cbec89ebb75c5c53cd783cd27fe1a3f09e not found: ID does not exist" Dec 01 08:50:51 crc kubenswrapper[4813]: I1201 08:50:51.682331 4813 scope.go:117] "RemoveContainer" containerID="7f19c67b1c6c80c7f3223512102810d46ede1445464d21ae3bbea83424539b2b" Dec 01 08:50:51 crc kubenswrapper[4813]: I1201 08:50:51.682631 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7f19c67b1c6c80c7f3223512102810d46ede1445464d21ae3bbea83424539b2b"} err="failed to get container status \"7f19c67b1c6c80c7f3223512102810d46ede1445464d21ae3bbea83424539b2b\": rpc error: code = NotFound desc = could not find container \"7f19c67b1c6c80c7f3223512102810d46ede1445464d21ae3bbea83424539b2b\": container with ID starting with 7f19c67b1c6c80c7f3223512102810d46ede1445464d21ae3bbea83424539b2b not found: ID does not exist" Dec 01 08:50:51 crc kubenswrapper[4813]: I1201 08:50:51.682661 4813 scope.go:117] "RemoveContainer" containerID="944977e9722025088720107530d0c91fc693593f380e4b49678f28a56da180cc" Dec 01 08:50:51 crc kubenswrapper[4813]: I1201 08:50:51.683128 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"944977e9722025088720107530d0c91fc693593f380e4b49678f28a56da180cc"} err="failed to get container status \"944977e9722025088720107530d0c91fc693593f380e4b49678f28a56da180cc\": rpc error: code = NotFound desc = could not find container \"944977e9722025088720107530d0c91fc693593f380e4b49678f28a56da180cc\": container with ID starting with 944977e9722025088720107530d0c91fc693593f380e4b49678f28a56da180cc not found: ID does not exist" Dec 01 08:50:51 crc kubenswrapper[4813]: I1201 08:50:51.683188 4813 scope.go:117] "RemoveContainer" containerID="ccc30a63dc1c23ff9880fb04af1f31f0656b2696d359988e23deed23989f480a" Dec 01 08:50:51 crc kubenswrapper[4813]: I1201 08:50:51.683665 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ccc30a63dc1c23ff9880fb04af1f31f0656b2696d359988e23deed23989f480a"} err="failed to get container status \"ccc30a63dc1c23ff9880fb04af1f31f0656b2696d359988e23deed23989f480a\": rpc error: code = NotFound desc = could not find container \"ccc30a63dc1c23ff9880fb04af1f31f0656b2696d359988e23deed23989f480a\": container with ID starting with ccc30a63dc1c23ff9880fb04af1f31f0656b2696d359988e23deed23989f480a not found: ID does not exist" Dec 01 08:50:51 crc kubenswrapper[4813]: I1201 08:50:51.683705 4813 scope.go:117] "RemoveContainer" containerID="9d984c90302e57f3e5a6f3ca3613b415db04ac74a5a228582f1b52ef1d59f93e" Dec 01 08:50:51 crc kubenswrapper[4813]: I1201 08:50:51.684142 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9d984c90302e57f3e5a6f3ca3613b415db04ac74a5a228582f1b52ef1d59f93e"} err="failed to get container status \"9d984c90302e57f3e5a6f3ca3613b415db04ac74a5a228582f1b52ef1d59f93e\": rpc error: code = NotFound desc = could not find container \"9d984c90302e57f3e5a6f3ca3613b415db04ac74a5a228582f1b52ef1d59f93e\": container with ID starting with 9d984c90302e57f3e5a6f3ca3613b415db04ac74a5a228582f1b52ef1d59f93e not found: ID does not exist" Dec 01 08:50:51 crc kubenswrapper[4813]: I1201 08:50:51.684188 4813 scope.go:117] "RemoveContainer" containerID="ce8963bf1e9ee66926907aaf3a03d18977f4d0a1983d7121b62fbc8e02a37e4d" Dec 01 08:50:51 crc kubenswrapper[4813]: I1201 08:50:51.684651 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ce8963bf1e9ee66926907aaf3a03d18977f4d0a1983d7121b62fbc8e02a37e4d"} err="failed to get container status \"ce8963bf1e9ee66926907aaf3a03d18977f4d0a1983d7121b62fbc8e02a37e4d\": rpc error: code = NotFound desc = could not find container \"ce8963bf1e9ee66926907aaf3a03d18977f4d0a1983d7121b62fbc8e02a37e4d\": container with ID starting with ce8963bf1e9ee66926907aaf3a03d18977f4d0a1983d7121b62fbc8e02a37e4d not found: ID does not exist" Dec 01 08:50:51 crc kubenswrapper[4813]: I1201 08:50:51.684715 4813 scope.go:117] "RemoveContainer" containerID="dd67634e0c806323334495401a4e20b48fea4bc5013acd973d2abd3d70870f99" Dec 01 08:50:51 crc kubenswrapper[4813]: I1201 08:50:51.685082 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dd67634e0c806323334495401a4e20b48fea4bc5013acd973d2abd3d70870f99"} err="failed to get container status \"dd67634e0c806323334495401a4e20b48fea4bc5013acd973d2abd3d70870f99\": rpc error: code = NotFound desc = could not find container \"dd67634e0c806323334495401a4e20b48fea4bc5013acd973d2abd3d70870f99\": container with ID starting with dd67634e0c806323334495401a4e20b48fea4bc5013acd973d2abd3d70870f99 not found: ID does not exist" Dec 01 08:50:51 crc kubenswrapper[4813]: I1201 08:50:51.685114 4813 scope.go:117] "RemoveContainer" containerID="6aa74106386539915851c80789dac37ef865908f24cb6a67ffa406c2599cb5d7" Dec 01 08:50:51 crc kubenswrapper[4813]: I1201 08:50:51.690183 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6aa74106386539915851c80789dac37ef865908f24cb6a67ffa406c2599cb5d7"} err="failed to get container status \"6aa74106386539915851c80789dac37ef865908f24cb6a67ffa406c2599cb5d7\": rpc error: code = NotFound desc = could not find container \"6aa74106386539915851c80789dac37ef865908f24cb6a67ffa406c2599cb5d7\": container with ID starting with 6aa74106386539915851c80789dac37ef865908f24cb6a67ffa406c2599cb5d7 not found: ID does not exist" Dec 01 08:50:51 crc kubenswrapper[4813]: I1201 08:50:51.690234 4813 scope.go:117] "RemoveContainer" containerID="b45454b440c5108fa150a6b856afa1778ba6e9f89633433cb3117b68f00f72c3" Dec 01 08:50:51 crc kubenswrapper[4813]: I1201 08:50:51.690721 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b45454b440c5108fa150a6b856afa1778ba6e9f89633433cb3117b68f00f72c3"} err="failed to get container status \"b45454b440c5108fa150a6b856afa1778ba6e9f89633433cb3117b68f00f72c3\": rpc error: code = NotFound desc = could not find container \"b45454b440c5108fa150a6b856afa1778ba6e9f89633433cb3117b68f00f72c3\": container with ID starting with b45454b440c5108fa150a6b856afa1778ba6e9f89633433cb3117b68f00f72c3 not found: ID does not exist" Dec 01 08:50:51 crc kubenswrapper[4813]: I1201 08:50:51.690748 4813 scope.go:117] "RemoveContainer" containerID="dc32d930ab23f30103d8fa42291909cbec89ebb75c5c53cd783cd27fe1a3f09e" Dec 01 08:50:51 crc kubenswrapper[4813]: I1201 08:50:51.691063 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dc32d930ab23f30103d8fa42291909cbec89ebb75c5c53cd783cd27fe1a3f09e"} err="failed to get container status \"dc32d930ab23f30103d8fa42291909cbec89ebb75c5c53cd783cd27fe1a3f09e\": rpc error: code = NotFound desc = could not find container \"dc32d930ab23f30103d8fa42291909cbec89ebb75c5c53cd783cd27fe1a3f09e\": container with ID starting with dc32d930ab23f30103d8fa42291909cbec89ebb75c5c53cd783cd27fe1a3f09e not found: ID does not exist" Dec 01 08:50:51 crc kubenswrapper[4813]: I1201 08:50:51.691092 4813 scope.go:117] "RemoveContainer" containerID="7f19c67b1c6c80c7f3223512102810d46ede1445464d21ae3bbea83424539b2b" Dec 01 08:50:51 crc kubenswrapper[4813]: I1201 08:50:51.691362 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7f19c67b1c6c80c7f3223512102810d46ede1445464d21ae3bbea83424539b2b"} err="failed to get container status \"7f19c67b1c6c80c7f3223512102810d46ede1445464d21ae3bbea83424539b2b\": rpc error: code = NotFound desc = could not find container \"7f19c67b1c6c80c7f3223512102810d46ede1445464d21ae3bbea83424539b2b\": container with ID starting with 7f19c67b1c6c80c7f3223512102810d46ede1445464d21ae3bbea83424539b2b not found: ID does not exist" Dec 01 08:50:51 crc kubenswrapper[4813]: I1201 08:50:51.691379 4813 scope.go:117] "RemoveContainer" containerID="944977e9722025088720107530d0c91fc693593f380e4b49678f28a56da180cc" Dec 01 08:50:51 crc kubenswrapper[4813]: I1201 08:50:51.691567 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"944977e9722025088720107530d0c91fc693593f380e4b49678f28a56da180cc"} err="failed to get container status \"944977e9722025088720107530d0c91fc693593f380e4b49678f28a56da180cc\": rpc error: code = NotFound desc = could not find container \"944977e9722025088720107530d0c91fc693593f380e4b49678f28a56da180cc\": container with ID starting with 944977e9722025088720107530d0c91fc693593f380e4b49678f28a56da180cc not found: ID does not exist" Dec 01 08:50:51 crc kubenswrapper[4813]: I1201 08:50:51.691588 4813 scope.go:117] "RemoveContainer" containerID="ccc30a63dc1c23ff9880fb04af1f31f0656b2696d359988e23deed23989f480a" Dec 01 08:50:51 crc kubenswrapper[4813]: I1201 08:50:51.691918 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ccc30a63dc1c23ff9880fb04af1f31f0656b2696d359988e23deed23989f480a"} err="failed to get container status \"ccc30a63dc1c23ff9880fb04af1f31f0656b2696d359988e23deed23989f480a\": rpc error: code = NotFound desc = could not find container \"ccc30a63dc1c23ff9880fb04af1f31f0656b2696d359988e23deed23989f480a\": container with ID starting with ccc30a63dc1c23ff9880fb04af1f31f0656b2696d359988e23deed23989f480a not found: ID does not exist" Dec 01 08:50:51 crc kubenswrapper[4813]: I1201 08:50:51.691942 4813 scope.go:117] "RemoveContainer" containerID="9d984c90302e57f3e5a6f3ca3613b415db04ac74a5a228582f1b52ef1d59f93e" Dec 01 08:50:51 crc kubenswrapper[4813]: I1201 08:50:51.694124 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9d984c90302e57f3e5a6f3ca3613b415db04ac74a5a228582f1b52ef1d59f93e"} err="failed to get container status \"9d984c90302e57f3e5a6f3ca3613b415db04ac74a5a228582f1b52ef1d59f93e\": rpc error: code = NotFound desc = could not find container \"9d984c90302e57f3e5a6f3ca3613b415db04ac74a5a228582f1b52ef1d59f93e\": container with ID starting with 9d984c90302e57f3e5a6f3ca3613b415db04ac74a5a228582f1b52ef1d59f93e not found: ID does not exist" Dec 01 08:50:51 crc kubenswrapper[4813]: I1201 08:50:51.694152 4813 scope.go:117] "RemoveContainer" containerID="ce8963bf1e9ee66926907aaf3a03d18977f4d0a1983d7121b62fbc8e02a37e4d" Dec 01 08:50:51 crc kubenswrapper[4813]: I1201 08:50:51.694792 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ce8963bf1e9ee66926907aaf3a03d18977f4d0a1983d7121b62fbc8e02a37e4d"} err="failed to get container status \"ce8963bf1e9ee66926907aaf3a03d18977f4d0a1983d7121b62fbc8e02a37e4d\": rpc error: code = NotFound desc = could not find container \"ce8963bf1e9ee66926907aaf3a03d18977f4d0a1983d7121b62fbc8e02a37e4d\": container with ID starting with ce8963bf1e9ee66926907aaf3a03d18977f4d0a1983d7121b62fbc8e02a37e4d not found: ID does not exist" Dec 01 08:50:51 crc kubenswrapper[4813]: I1201 08:50:51.694822 4813 scope.go:117] "RemoveContainer" containerID="dd67634e0c806323334495401a4e20b48fea4bc5013acd973d2abd3d70870f99" Dec 01 08:50:51 crc kubenswrapper[4813]: I1201 08:50:51.695197 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dd67634e0c806323334495401a4e20b48fea4bc5013acd973d2abd3d70870f99"} err="failed to get container status \"dd67634e0c806323334495401a4e20b48fea4bc5013acd973d2abd3d70870f99\": rpc error: code = NotFound desc = could not find container \"dd67634e0c806323334495401a4e20b48fea4bc5013acd973d2abd3d70870f99\": container with ID starting with dd67634e0c806323334495401a4e20b48fea4bc5013acd973d2abd3d70870f99 not found: ID does not exist" Dec 01 08:50:52 crc kubenswrapper[4813]: I1201 08:50:52.405698 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b7505e14-3560-4d0d-be2c-d6a7a5953865" path="/var/lib/kubelet/pods/b7505e14-3560-4d0d-be2c-d6a7a5953865/volumes" Dec 01 08:50:52 crc kubenswrapper[4813]: I1201 08:50:52.429286 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qnss9" event={"ID":"5a004969-7056-4b56-93ad-f93d8736c43f","Type":"ContainerStarted","Data":"3044ae24f5d9c8bb0c01408bdff05dc6efee3bb5e24042e4e9d0ff07d22afc32"} Dec 01 08:50:52 crc kubenswrapper[4813]: I1201 08:50:52.429375 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qnss9" event={"ID":"5a004969-7056-4b56-93ad-f93d8736c43f","Type":"ContainerStarted","Data":"a1285e021c80ba05309eb9740a8c6f8f89c07e6556f42e29b451b0ec40ddb3bb"} Dec 01 08:50:52 crc kubenswrapper[4813]: I1201 08:50:52.429409 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qnss9" event={"ID":"5a004969-7056-4b56-93ad-f93d8736c43f","Type":"ContainerStarted","Data":"ef47af54a4cd6a3952072c69f78da16fe3c6509d1839959bb67df11f15334475"} Dec 01 08:50:52 crc kubenswrapper[4813]: I1201 08:50:52.429435 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qnss9" event={"ID":"5a004969-7056-4b56-93ad-f93d8736c43f","Type":"ContainerStarted","Data":"191a8635846378d0501dad601f8a0b416b5ed58afbc1f9778089785fa722e195"} Dec 01 08:50:52 crc kubenswrapper[4813]: I1201 08:50:52.429500 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qnss9" event={"ID":"5a004969-7056-4b56-93ad-f93d8736c43f","Type":"ContainerStarted","Data":"7b30e43c23178fa1084826698d50c7e583c26529784ebb51ce5794bbc5e91332"} Dec 01 08:50:52 crc kubenswrapper[4813]: I1201 08:50:52.429529 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qnss9" event={"ID":"5a004969-7056-4b56-93ad-f93d8736c43f","Type":"ContainerStarted","Data":"e93aa8bf48d626a0f70954e277da7f9eda61cad027f2f28b2aa08b60a5160f76"} Dec 01 08:50:55 crc kubenswrapper[4813]: I1201 08:50:55.458739 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qnss9" event={"ID":"5a004969-7056-4b56-93ad-f93d8736c43f","Type":"ContainerStarted","Data":"684d2d08ffe734abb96cb609708cd6578a26cd126fd77f7f021c7f3bca24e1bd"} Dec 01 08:50:57 crc kubenswrapper[4813]: I1201 08:50:57.472510 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qnss9" event={"ID":"5a004969-7056-4b56-93ad-f93d8736c43f","Type":"ContainerStarted","Data":"5e5adf829a4bf63f604858ceab3ca6409d47c4b78589f1952709b87720cc79ca"} Dec 01 08:50:57 crc kubenswrapper[4813]: I1201 08:50:57.473108 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-qnss9" Dec 01 08:50:57 crc kubenswrapper[4813]: I1201 08:50:57.473149 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-qnss9" Dec 01 08:50:57 crc kubenswrapper[4813]: I1201 08:50:57.506427 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-qnss9" Dec 01 08:50:57 crc kubenswrapper[4813]: I1201 08:50:57.512663 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-qnss9" podStartSLOduration=7.512618755 podStartE2EDuration="7.512618755s" podCreationTimestamp="2025-12-01 08:50:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:50:57.507291803 +0000 UTC m=+642.130213399" watchObservedRunningTime="2025-12-01 08:50:57.512618755 +0000 UTC m=+642.135540341" Dec 01 08:50:58 crc kubenswrapper[4813]: I1201 08:50:58.477572 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-qnss9" Dec 01 08:50:58 crc kubenswrapper[4813]: I1201 08:50:58.517641 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-qnss9" Dec 01 08:51:01 crc kubenswrapper[4813]: I1201 08:51:01.393459 4813 scope.go:117] "RemoveContainer" containerID="2b730cdf02e6456d9a1d02d66b94064ca71052eb8f373eaabfe327369eb03a15" Dec 01 08:51:02 crc kubenswrapper[4813]: I1201 08:51:02.504494 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-b6qqv_b9656a03-0096-4c89-b0da-e37e4103611b/kube-multus/1.log" Dec 01 08:51:02 crc kubenswrapper[4813]: I1201 08:51:02.505219 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-b6qqv" event={"ID":"b9656a03-0096-4c89-b0da-e37e4103611b","Type":"ContainerStarted","Data":"bab16a3d4a5b679b34a8d2e81a3794dcb318f30a8f48e93efe77d19092fc99e4"} Dec 01 08:51:20 crc kubenswrapper[4813]: I1201 08:51:20.958889 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-qnss9" Dec 01 08:51:26 crc kubenswrapper[4813]: I1201 08:51:26.028496 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83qc9k9"] Dec 01 08:51:26 crc kubenswrapper[4813]: I1201 08:51:26.030452 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83qc9k9" Dec 01 08:51:26 crc kubenswrapper[4813]: I1201 08:51:26.032873 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Dec 01 08:51:26 crc kubenswrapper[4813]: I1201 08:51:26.043558 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83qc9k9"] Dec 01 08:51:26 crc kubenswrapper[4813]: I1201 08:51:26.124493 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d7e5aa34-61f6-476a-a08e-39ce74c1f82a-util\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83qc9k9\" (UID: \"d7e5aa34-61f6-476a-a08e-39ce74c1f82a\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83qc9k9" Dec 01 08:51:26 crc kubenswrapper[4813]: I1201 08:51:26.124536 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d7e5aa34-61f6-476a-a08e-39ce74c1f82a-bundle\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83qc9k9\" (UID: \"d7e5aa34-61f6-476a-a08e-39ce74c1f82a\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83qc9k9" Dec 01 08:51:26 crc kubenswrapper[4813]: I1201 08:51:26.124565 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6n9vm\" (UniqueName: \"kubernetes.io/projected/d7e5aa34-61f6-476a-a08e-39ce74c1f82a-kube-api-access-6n9vm\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83qc9k9\" (UID: \"d7e5aa34-61f6-476a-a08e-39ce74c1f82a\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83qc9k9" Dec 01 08:51:26 crc kubenswrapper[4813]: I1201 08:51:26.226324 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d7e5aa34-61f6-476a-a08e-39ce74c1f82a-util\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83qc9k9\" (UID: \"d7e5aa34-61f6-476a-a08e-39ce74c1f82a\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83qc9k9" Dec 01 08:51:26 crc kubenswrapper[4813]: I1201 08:51:26.226380 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d7e5aa34-61f6-476a-a08e-39ce74c1f82a-bundle\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83qc9k9\" (UID: \"d7e5aa34-61f6-476a-a08e-39ce74c1f82a\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83qc9k9" Dec 01 08:51:26 crc kubenswrapper[4813]: I1201 08:51:26.226425 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6n9vm\" (UniqueName: \"kubernetes.io/projected/d7e5aa34-61f6-476a-a08e-39ce74c1f82a-kube-api-access-6n9vm\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83qc9k9\" (UID: \"d7e5aa34-61f6-476a-a08e-39ce74c1f82a\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83qc9k9" Dec 01 08:51:26 crc kubenswrapper[4813]: I1201 08:51:26.227124 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d7e5aa34-61f6-476a-a08e-39ce74c1f82a-util\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83qc9k9\" (UID: \"d7e5aa34-61f6-476a-a08e-39ce74c1f82a\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83qc9k9" Dec 01 08:51:26 crc kubenswrapper[4813]: I1201 08:51:26.227138 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d7e5aa34-61f6-476a-a08e-39ce74c1f82a-bundle\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83qc9k9\" (UID: \"d7e5aa34-61f6-476a-a08e-39ce74c1f82a\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83qc9k9" Dec 01 08:51:26 crc kubenswrapper[4813]: I1201 08:51:26.257356 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6n9vm\" (UniqueName: \"kubernetes.io/projected/d7e5aa34-61f6-476a-a08e-39ce74c1f82a-kube-api-access-6n9vm\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83qc9k9\" (UID: \"d7e5aa34-61f6-476a-a08e-39ce74c1f82a\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83qc9k9" Dec 01 08:51:26 crc kubenswrapper[4813]: I1201 08:51:26.365718 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83qc9k9" Dec 01 08:51:26 crc kubenswrapper[4813]: I1201 08:51:26.868254 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83qc9k9"] Dec 01 08:51:27 crc kubenswrapper[4813]: I1201 08:51:27.721623 4813 generic.go:334] "Generic (PLEG): container finished" podID="d7e5aa34-61f6-476a-a08e-39ce74c1f82a" containerID="f5e815adb1597b213601d2e32a0456e0c402a315e4ed48a3fef33577915f6af9" exitCode=0 Dec 01 08:51:27 crc kubenswrapper[4813]: I1201 08:51:27.721780 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83qc9k9" event={"ID":"d7e5aa34-61f6-476a-a08e-39ce74c1f82a","Type":"ContainerDied","Data":"f5e815adb1597b213601d2e32a0456e0c402a315e4ed48a3fef33577915f6af9"} Dec 01 08:51:27 crc kubenswrapper[4813]: I1201 08:51:27.722101 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83qc9k9" event={"ID":"d7e5aa34-61f6-476a-a08e-39ce74c1f82a","Type":"ContainerStarted","Data":"18f6ddae91e0982d6dd4b080c9ae944e20eeae13fadc313466d7025a43ed1c71"} Dec 01 08:51:27 crc kubenswrapper[4813]: I1201 08:51:27.724603 4813 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 01 08:51:29 crc kubenswrapper[4813]: I1201 08:51:29.737164 4813 generic.go:334] "Generic (PLEG): container finished" podID="d7e5aa34-61f6-476a-a08e-39ce74c1f82a" containerID="4c08fd6c4607fffbe3522aeecf16eafec37b1cedc716bb5763e4f8d9e6a5daf4" exitCode=0 Dec 01 08:51:29 crc kubenswrapper[4813]: I1201 08:51:29.737229 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83qc9k9" event={"ID":"d7e5aa34-61f6-476a-a08e-39ce74c1f82a","Type":"ContainerDied","Data":"4c08fd6c4607fffbe3522aeecf16eafec37b1cedc716bb5763e4f8d9e6a5daf4"} Dec 01 08:51:30 crc kubenswrapper[4813]: I1201 08:51:30.744796 4813 generic.go:334] "Generic (PLEG): container finished" podID="d7e5aa34-61f6-476a-a08e-39ce74c1f82a" containerID="6cb8464562a9ceae9abb90638520839d1a9fb425fc56c4391a5f9cf1ea048608" exitCode=0 Dec 01 08:51:30 crc kubenswrapper[4813]: I1201 08:51:30.744857 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83qc9k9" event={"ID":"d7e5aa34-61f6-476a-a08e-39ce74c1f82a","Type":"ContainerDied","Data":"6cb8464562a9ceae9abb90638520839d1a9fb425fc56c4391a5f9cf1ea048608"} Dec 01 08:51:32 crc kubenswrapper[4813]: I1201 08:51:32.031471 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83qc9k9" Dec 01 08:51:32 crc kubenswrapper[4813]: I1201 08:51:32.218264 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6n9vm\" (UniqueName: \"kubernetes.io/projected/d7e5aa34-61f6-476a-a08e-39ce74c1f82a-kube-api-access-6n9vm\") pod \"d7e5aa34-61f6-476a-a08e-39ce74c1f82a\" (UID: \"d7e5aa34-61f6-476a-a08e-39ce74c1f82a\") " Dec 01 08:51:32 crc kubenswrapper[4813]: I1201 08:51:32.218337 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d7e5aa34-61f6-476a-a08e-39ce74c1f82a-util\") pod \"d7e5aa34-61f6-476a-a08e-39ce74c1f82a\" (UID: \"d7e5aa34-61f6-476a-a08e-39ce74c1f82a\") " Dec 01 08:51:32 crc kubenswrapper[4813]: I1201 08:51:32.218416 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d7e5aa34-61f6-476a-a08e-39ce74c1f82a-bundle\") pod \"d7e5aa34-61f6-476a-a08e-39ce74c1f82a\" (UID: \"d7e5aa34-61f6-476a-a08e-39ce74c1f82a\") " Dec 01 08:51:32 crc kubenswrapper[4813]: I1201 08:51:32.219757 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d7e5aa34-61f6-476a-a08e-39ce74c1f82a-bundle" (OuterVolumeSpecName: "bundle") pod "d7e5aa34-61f6-476a-a08e-39ce74c1f82a" (UID: "d7e5aa34-61f6-476a-a08e-39ce74c1f82a"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 08:51:32 crc kubenswrapper[4813]: I1201 08:51:32.227641 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d7e5aa34-61f6-476a-a08e-39ce74c1f82a-kube-api-access-6n9vm" (OuterVolumeSpecName: "kube-api-access-6n9vm") pod "d7e5aa34-61f6-476a-a08e-39ce74c1f82a" (UID: "d7e5aa34-61f6-476a-a08e-39ce74c1f82a"). InnerVolumeSpecName "kube-api-access-6n9vm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:51:32 crc kubenswrapper[4813]: I1201 08:51:32.320541 4813 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d7e5aa34-61f6-476a-a08e-39ce74c1f82a-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 08:51:32 crc kubenswrapper[4813]: I1201 08:51:32.320585 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6n9vm\" (UniqueName: \"kubernetes.io/projected/d7e5aa34-61f6-476a-a08e-39ce74c1f82a-kube-api-access-6n9vm\") on node \"crc\" DevicePath \"\"" Dec 01 08:51:32 crc kubenswrapper[4813]: I1201 08:51:32.491417 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d7e5aa34-61f6-476a-a08e-39ce74c1f82a-util" (OuterVolumeSpecName: "util") pod "d7e5aa34-61f6-476a-a08e-39ce74c1f82a" (UID: "d7e5aa34-61f6-476a-a08e-39ce74c1f82a"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 08:51:32 crc kubenswrapper[4813]: I1201 08:51:32.524317 4813 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d7e5aa34-61f6-476a-a08e-39ce74c1f82a-util\") on node \"crc\" DevicePath \"\"" Dec 01 08:51:32 crc kubenswrapper[4813]: I1201 08:51:32.769140 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83qc9k9" event={"ID":"d7e5aa34-61f6-476a-a08e-39ce74c1f82a","Type":"ContainerDied","Data":"18f6ddae91e0982d6dd4b080c9ae944e20eeae13fadc313466d7025a43ed1c71"} Dec 01 08:51:32 crc kubenswrapper[4813]: I1201 08:51:32.769264 4813 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="18f6ddae91e0982d6dd4b080c9ae944e20eeae13fadc313466d7025a43ed1c71" Dec 01 08:51:32 crc kubenswrapper[4813]: I1201 08:51:32.769389 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83qc9k9" Dec 01 08:51:40 crc kubenswrapper[4813]: I1201 08:51:40.454798 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-controller-manager-754cc864cb-4d2tv"] Dec 01 08:51:40 crc kubenswrapper[4813]: E1201 08:51:40.455610 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d7e5aa34-61f6-476a-a08e-39ce74c1f82a" containerName="util" Dec 01 08:51:40 crc kubenswrapper[4813]: I1201 08:51:40.455641 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="d7e5aa34-61f6-476a-a08e-39ce74c1f82a" containerName="util" Dec 01 08:51:40 crc kubenswrapper[4813]: E1201 08:51:40.455664 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d7e5aa34-61f6-476a-a08e-39ce74c1f82a" containerName="pull" Dec 01 08:51:40 crc kubenswrapper[4813]: I1201 08:51:40.455673 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="d7e5aa34-61f6-476a-a08e-39ce74c1f82a" containerName="pull" Dec 01 08:51:40 crc kubenswrapper[4813]: E1201 08:51:40.455690 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d7e5aa34-61f6-476a-a08e-39ce74c1f82a" containerName="extract" Dec 01 08:51:40 crc kubenswrapper[4813]: I1201 08:51:40.455698 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="d7e5aa34-61f6-476a-a08e-39ce74c1f82a" containerName="extract" Dec 01 08:51:40 crc kubenswrapper[4813]: I1201 08:51:40.455849 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="d7e5aa34-61f6-476a-a08e-39ce74c1f82a" containerName="extract" Dec 01 08:51:40 crc kubenswrapper[4813]: I1201 08:51:40.456368 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-754cc864cb-4d2tv" Dec 01 08:51:40 crc kubenswrapper[4813]: I1201 08:51:40.461978 4813 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-controller-manager-service-cert" Dec 01 08:51:40 crc kubenswrapper[4813]: I1201 08:51:40.462234 4813 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-cert" Dec 01 08:51:40 crc kubenswrapper[4813]: I1201 08:51:40.462380 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"openshift-service-ca.crt" Dec 01 08:51:40 crc kubenswrapper[4813]: I1201 08:51:40.462600 4813 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"manager-account-dockercfg-568q5" Dec 01 08:51:40 crc kubenswrapper[4813]: I1201 08:51:40.463666 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"kube-root-ca.crt" Dec 01 08:51:40 crc kubenswrapper[4813]: I1201 08:51:40.483277 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-754cc864cb-4d2tv"] Dec 01 08:51:40 crc kubenswrapper[4813]: I1201 08:51:40.527816 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/4017da37-9dca-4e2d-bab0-a7896bdf3cd5-apiservice-cert\") pod \"metallb-operator-controller-manager-754cc864cb-4d2tv\" (UID: \"4017da37-9dca-4e2d-bab0-a7896bdf3cd5\") " pod="metallb-system/metallb-operator-controller-manager-754cc864cb-4d2tv" Dec 01 08:51:40 crc kubenswrapper[4813]: I1201 08:51:40.527902 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/4017da37-9dca-4e2d-bab0-a7896bdf3cd5-webhook-cert\") pod \"metallb-operator-controller-manager-754cc864cb-4d2tv\" (UID: \"4017da37-9dca-4e2d-bab0-a7896bdf3cd5\") " pod="metallb-system/metallb-operator-controller-manager-754cc864cb-4d2tv" Dec 01 08:51:40 crc kubenswrapper[4813]: I1201 08:51:40.527938 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4wpm8\" (UniqueName: \"kubernetes.io/projected/4017da37-9dca-4e2d-bab0-a7896bdf3cd5-kube-api-access-4wpm8\") pod \"metallb-operator-controller-manager-754cc864cb-4d2tv\" (UID: \"4017da37-9dca-4e2d-bab0-a7896bdf3cd5\") " pod="metallb-system/metallb-operator-controller-manager-754cc864cb-4d2tv" Dec 01 08:51:40 crc kubenswrapper[4813]: I1201 08:51:40.628619 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/4017da37-9dca-4e2d-bab0-a7896bdf3cd5-webhook-cert\") pod \"metallb-operator-controller-manager-754cc864cb-4d2tv\" (UID: \"4017da37-9dca-4e2d-bab0-a7896bdf3cd5\") " pod="metallb-system/metallb-operator-controller-manager-754cc864cb-4d2tv" Dec 01 08:51:40 crc kubenswrapper[4813]: I1201 08:51:40.628690 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4wpm8\" (UniqueName: \"kubernetes.io/projected/4017da37-9dca-4e2d-bab0-a7896bdf3cd5-kube-api-access-4wpm8\") pod \"metallb-operator-controller-manager-754cc864cb-4d2tv\" (UID: \"4017da37-9dca-4e2d-bab0-a7896bdf3cd5\") " pod="metallb-system/metallb-operator-controller-manager-754cc864cb-4d2tv" Dec 01 08:51:40 crc kubenswrapper[4813]: I1201 08:51:40.628735 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/4017da37-9dca-4e2d-bab0-a7896bdf3cd5-apiservice-cert\") pod \"metallb-operator-controller-manager-754cc864cb-4d2tv\" (UID: \"4017da37-9dca-4e2d-bab0-a7896bdf3cd5\") " pod="metallb-system/metallb-operator-controller-manager-754cc864cb-4d2tv" Dec 01 08:51:40 crc kubenswrapper[4813]: I1201 08:51:40.635754 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/4017da37-9dca-4e2d-bab0-a7896bdf3cd5-webhook-cert\") pod \"metallb-operator-controller-manager-754cc864cb-4d2tv\" (UID: \"4017da37-9dca-4e2d-bab0-a7896bdf3cd5\") " pod="metallb-system/metallb-operator-controller-manager-754cc864cb-4d2tv" Dec 01 08:51:40 crc kubenswrapper[4813]: I1201 08:51:40.646510 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4wpm8\" (UniqueName: \"kubernetes.io/projected/4017da37-9dca-4e2d-bab0-a7896bdf3cd5-kube-api-access-4wpm8\") pod \"metallb-operator-controller-manager-754cc864cb-4d2tv\" (UID: \"4017da37-9dca-4e2d-bab0-a7896bdf3cd5\") " pod="metallb-system/metallb-operator-controller-manager-754cc864cb-4d2tv" Dec 01 08:51:40 crc kubenswrapper[4813]: I1201 08:51:40.653354 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/4017da37-9dca-4e2d-bab0-a7896bdf3cd5-apiservice-cert\") pod \"metallb-operator-controller-manager-754cc864cb-4d2tv\" (UID: \"4017da37-9dca-4e2d-bab0-a7896bdf3cd5\") " pod="metallb-system/metallb-operator-controller-manager-754cc864cb-4d2tv" Dec 01 08:51:40 crc kubenswrapper[4813]: I1201 08:51:40.757087 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-webhook-server-79d58979df-jnh5k"] Dec 01 08:51:40 crc kubenswrapper[4813]: I1201 08:51:40.757738 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-79d58979df-jnh5k" Dec 01 08:51:40 crc kubenswrapper[4813]: I1201 08:51:40.760789 4813 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-service-cert" Dec 01 08:51:40 crc kubenswrapper[4813]: I1201 08:51:40.760933 4813 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Dec 01 08:51:40 crc kubenswrapper[4813]: I1201 08:51:40.761697 4813 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-dockercfg-qw7t5" Dec 01 08:51:40 crc kubenswrapper[4813]: I1201 08:51:40.775496 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-754cc864cb-4d2tv" Dec 01 08:51:40 crc kubenswrapper[4813]: I1201 08:51:40.849320 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-79d58979df-jnh5k"] Dec 01 08:51:40 crc kubenswrapper[4813]: I1201 08:51:40.932275 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ed86505b-1475-4ca2-9c22-73ce93b80ab9-webhook-cert\") pod \"metallb-operator-webhook-server-79d58979df-jnh5k\" (UID: \"ed86505b-1475-4ca2-9c22-73ce93b80ab9\") " pod="metallb-system/metallb-operator-webhook-server-79d58979df-jnh5k" Dec 01 08:51:40 crc kubenswrapper[4813]: I1201 08:51:40.932375 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-npqdl\" (UniqueName: \"kubernetes.io/projected/ed86505b-1475-4ca2-9c22-73ce93b80ab9-kube-api-access-npqdl\") pod \"metallb-operator-webhook-server-79d58979df-jnh5k\" (UID: \"ed86505b-1475-4ca2-9c22-73ce93b80ab9\") " pod="metallb-system/metallb-operator-webhook-server-79d58979df-jnh5k" Dec 01 08:51:40 crc kubenswrapper[4813]: I1201 08:51:40.932408 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/ed86505b-1475-4ca2-9c22-73ce93b80ab9-apiservice-cert\") pod \"metallb-operator-webhook-server-79d58979df-jnh5k\" (UID: \"ed86505b-1475-4ca2-9c22-73ce93b80ab9\") " pod="metallb-system/metallb-operator-webhook-server-79d58979df-jnh5k" Dec 01 08:51:41 crc kubenswrapper[4813]: I1201 08:51:41.033761 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-npqdl\" (UniqueName: \"kubernetes.io/projected/ed86505b-1475-4ca2-9c22-73ce93b80ab9-kube-api-access-npqdl\") pod \"metallb-operator-webhook-server-79d58979df-jnh5k\" (UID: \"ed86505b-1475-4ca2-9c22-73ce93b80ab9\") " pod="metallb-system/metallb-operator-webhook-server-79d58979df-jnh5k" Dec 01 08:51:41 crc kubenswrapper[4813]: I1201 08:51:41.034045 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/ed86505b-1475-4ca2-9c22-73ce93b80ab9-apiservice-cert\") pod \"metallb-operator-webhook-server-79d58979df-jnh5k\" (UID: \"ed86505b-1475-4ca2-9c22-73ce93b80ab9\") " pod="metallb-system/metallb-operator-webhook-server-79d58979df-jnh5k" Dec 01 08:51:41 crc kubenswrapper[4813]: I1201 08:51:41.034154 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ed86505b-1475-4ca2-9c22-73ce93b80ab9-webhook-cert\") pod \"metallb-operator-webhook-server-79d58979df-jnh5k\" (UID: \"ed86505b-1475-4ca2-9c22-73ce93b80ab9\") " pod="metallb-system/metallb-operator-webhook-server-79d58979df-jnh5k" Dec 01 08:51:41 crc kubenswrapper[4813]: I1201 08:51:41.040847 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ed86505b-1475-4ca2-9c22-73ce93b80ab9-webhook-cert\") pod \"metallb-operator-webhook-server-79d58979df-jnh5k\" (UID: \"ed86505b-1475-4ca2-9c22-73ce93b80ab9\") " pod="metallb-system/metallb-operator-webhook-server-79d58979df-jnh5k" Dec 01 08:51:41 crc kubenswrapper[4813]: I1201 08:51:41.043560 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/ed86505b-1475-4ca2-9c22-73ce93b80ab9-apiservice-cert\") pod \"metallb-operator-webhook-server-79d58979df-jnh5k\" (UID: \"ed86505b-1475-4ca2-9c22-73ce93b80ab9\") " pod="metallb-system/metallb-operator-webhook-server-79d58979df-jnh5k" Dec 01 08:51:41 crc kubenswrapper[4813]: I1201 08:51:41.061616 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-npqdl\" (UniqueName: \"kubernetes.io/projected/ed86505b-1475-4ca2-9c22-73ce93b80ab9-kube-api-access-npqdl\") pod \"metallb-operator-webhook-server-79d58979df-jnh5k\" (UID: \"ed86505b-1475-4ca2-9c22-73ce93b80ab9\") " pod="metallb-system/metallb-operator-webhook-server-79d58979df-jnh5k" Dec 01 08:51:41 crc kubenswrapper[4813]: I1201 08:51:41.073440 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-79d58979df-jnh5k" Dec 01 08:51:41 crc kubenswrapper[4813]: I1201 08:51:41.116232 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-754cc864cb-4d2tv"] Dec 01 08:51:41 crc kubenswrapper[4813]: W1201 08:51:41.128237 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4017da37_9dca_4e2d_bab0_a7896bdf3cd5.slice/crio-1d26a6fc8cd9721d944052dbecc1853a1f286ad9ed8a06eecfc2aa31337e03b7 WatchSource:0}: Error finding container 1d26a6fc8cd9721d944052dbecc1853a1f286ad9ed8a06eecfc2aa31337e03b7: Status 404 returned error can't find the container with id 1d26a6fc8cd9721d944052dbecc1853a1f286ad9ed8a06eecfc2aa31337e03b7 Dec 01 08:51:41 crc kubenswrapper[4813]: I1201 08:51:41.278240 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-79d58979df-jnh5k"] Dec 01 08:51:41 crc kubenswrapper[4813]: W1201 08:51:41.281412 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poded86505b_1475_4ca2_9c22_73ce93b80ab9.slice/crio-0a8fadf5ae878146615b5c97753feb918327ece1170d5262710229a6cb91699e WatchSource:0}: Error finding container 0a8fadf5ae878146615b5c97753feb918327ece1170d5262710229a6cb91699e: Status 404 returned error can't find the container with id 0a8fadf5ae878146615b5c97753feb918327ece1170d5262710229a6cb91699e Dec 01 08:51:41 crc kubenswrapper[4813]: I1201 08:51:41.908326 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-754cc864cb-4d2tv" event={"ID":"4017da37-9dca-4e2d-bab0-a7896bdf3cd5","Type":"ContainerStarted","Data":"1d26a6fc8cd9721d944052dbecc1853a1f286ad9ed8a06eecfc2aa31337e03b7"} Dec 01 08:51:41 crc kubenswrapper[4813]: I1201 08:51:41.909937 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-79d58979df-jnh5k" event={"ID":"ed86505b-1475-4ca2-9c22-73ce93b80ab9","Type":"ContainerStarted","Data":"0a8fadf5ae878146615b5c97753feb918327ece1170d5262710229a6cb91699e"} Dec 01 08:51:47 crc kubenswrapper[4813]: I1201 08:51:47.061744 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-79d58979df-jnh5k" event={"ID":"ed86505b-1475-4ca2-9c22-73ce93b80ab9","Type":"ContainerStarted","Data":"6c42e21912c05841ac6d9a6f0846cf2d9cdb2222062593f965d5237f76e228c4"} Dec 01 08:51:47 crc kubenswrapper[4813]: I1201 08:51:47.062710 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-79d58979df-jnh5k" Dec 01 08:51:47 crc kubenswrapper[4813]: I1201 08:51:47.064180 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-754cc864cb-4d2tv" event={"ID":"4017da37-9dca-4e2d-bab0-a7896bdf3cd5","Type":"ContainerStarted","Data":"8f8e07f54490f71dd05a1c4429a4556e281ec11a641887c940688867119d5d82"} Dec 01 08:51:47 crc kubenswrapper[4813]: I1201 08:51:47.064477 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-754cc864cb-4d2tv" Dec 01 08:51:47 crc kubenswrapper[4813]: I1201 08:51:47.093337 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-webhook-server-79d58979df-jnh5k" podStartSLOduration=2.534052451 podStartE2EDuration="7.093282418s" podCreationTimestamp="2025-12-01 08:51:40 +0000 UTC" firstStartedPulling="2025-12-01 08:51:41.28969648 +0000 UTC m=+685.912618076" lastFinishedPulling="2025-12-01 08:51:45.848926457 +0000 UTC m=+690.471848043" observedRunningTime="2025-12-01 08:51:47.090595319 +0000 UTC m=+691.713516995" watchObservedRunningTime="2025-12-01 08:51:47.093282418 +0000 UTC m=+691.716204014" Dec 01 08:51:47 crc kubenswrapper[4813]: I1201 08:51:47.115647 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-controller-manager-754cc864cb-4d2tv" podStartSLOduration=2.406684438 podStartE2EDuration="7.11562764s" podCreationTimestamp="2025-12-01 08:51:40 +0000 UTC" firstStartedPulling="2025-12-01 08:51:41.132701608 +0000 UTC m=+685.755623184" lastFinishedPulling="2025-12-01 08:51:45.8416448 +0000 UTC m=+690.464566386" observedRunningTime="2025-12-01 08:51:47.112184472 +0000 UTC m=+691.735106058" watchObservedRunningTime="2025-12-01 08:51:47.11562764 +0000 UTC m=+691.738549226" Dec 01 08:51:47 crc kubenswrapper[4813]: I1201 08:51:47.222044 4813 patch_prober.go:28] interesting pod/machine-config-daemon-9wn4q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 08:51:47 crc kubenswrapper[4813]: I1201 08:51:47.222195 4813 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9wn4q" podUID="3d423402-e08f-42e9-b60c-6eb9bf2f6d48" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 08:52:01 crc kubenswrapper[4813]: I1201 08:52:01.081989 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-79d58979df-jnh5k" Dec 01 08:52:13 crc kubenswrapper[4813]: I1201 08:52:13.338329 4813 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Dec 01 08:52:17 crc kubenswrapper[4813]: I1201 08:52:17.221372 4813 patch_prober.go:28] interesting pod/machine-config-daemon-9wn4q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 08:52:17 crc kubenswrapper[4813]: I1201 08:52:17.221669 4813 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9wn4q" podUID="3d423402-e08f-42e9-b60c-6eb9bf2f6d48" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 08:52:20 crc kubenswrapper[4813]: I1201 08:52:20.779097 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-754cc864cb-4d2tv" Dec 01 08:52:21 crc kubenswrapper[4813]: I1201 08:52:21.558756 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-zplkx"] Dec 01 08:52:21 crc kubenswrapper[4813]: I1201 08:52:21.561476 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-zplkx" Dec 01 08:52:21 crc kubenswrapper[4813]: I1201 08:52:21.562290 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-webhook-server-7fcb986d4-rgj8f"] Dec 01 08:52:21 crc kubenswrapper[4813]: I1201 08:52:21.562954 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-rgj8f" Dec 01 08:52:21 crc kubenswrapper[4813]: W1201 08:52:21.566936 4813 reflector.go:561] object-"metallb-system"/"frr-k8s-certs-secret": failed to list *v1.Secret: secrets "frr-k8s-certs-secret" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "metallb-system": no relationship found between node 'crc' and this object Dec 01 08:52:21 crc kubenswrapper[4813]: E1201 08:52:21.567081 4813 reflector.go:158] "Unhandled Error" err="object-\"metallb-system\"/\"frr-k8s-certs-secret\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"frr-k8s-certs-secret\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"metallb-system\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 01 08:52:21 crc kubenswrapper[4813]: W1201 08:52:21.567181 4813 reflector.go:561] object-"metallb-system"/"frr-k8s-daemon-dockercfg-dfhlr": failed to list *v1.Secret: secrets "frr-k8s-daemon-dockercfg-dfhlr" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "metallb-system": no relationship found between node 'crc' and this object Dec 01 08:52:21 crc kubenswrapper[4813]: E1201 08:52:21.567204 4813 reflector.go:158] "Unhandled Error" err="object-\"metallb-system\"/\"frr-k8s-daemon-dockercfg-dfhlr\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"frr-k8s-daemon-dockercfg-dfhlr\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"metallb-system\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 01 08:52:21 crc kubenswrapper[4813]: W1201 08:52:21.567228 4813 reflector.go:561] object-"metallb-system"/"frr-k8s-webhook-server-cert": failed to list *v1.Secret: secrets "frr-k8s-webhook-server-cert" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "metallb-system": no relationship found between node 'crc' and this object Dec 01 08:52:21 crc kubenswrapper[4813]: E1201 08:52:21.567289 4813 reflector.go:158] "Unhandled Error" err="object-\"metallb-system\"/\"frr-k8s-webhook-server-cert\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"frr-k8s-webhook-server-cert\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"metallb-system\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 01 08:52:21 crc kubenswrapper[4813]: W1201 08:52:21.567237 4813 reflector.go:561] object-"metallb-system"/"frr-startup": failed to list *v1.ConfigMap: configmaps "frr-startup" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "metallb-system": no relationship found between node 'crc' and this object Dec 01 08:52:21 crc kubenswrapper[4813]: E1201 08:52:21.567323 4813 reflector.go:158] "Unhandled Error" err="object-\"metallb-system\"/\"frr-startup\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"frr-startup\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"metallb-system\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 01 08:52:21 crc kubenswrapper[4813]: I1201 08:52:21.601073 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-7fcb986d4-rgj8f"] Dec 01 08:52:21 crc kubenswrapper[4813]: I1201 08:52:21.658599 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/speaker-227b6"] Dec 01 08:52:21 crc kubenswrapper[4813]: I1201 08:52:21.659689 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-227b6" Dec 01 08:52:21 crc kubenswrapper[4813]: I1201 08:52:21.661357 4813 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret" Dec 01 08:52:21 crc kubenswrapper[4813]: I1201 08:52:21.661673 4813 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-5kkx4" Dec 01 08:52:21 crc kubenswrapper[4813]: I1201 08:52:21.661797 4813 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist" Dec 01 08:52:21 crc kubenswrapper[4813]: I1201 08:52:21.666837 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2" Dec 01 08:52:21 crc kubenswrapper[4813]: I1201 08:52:21.683375 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/controller-f8648f98b-m47sh"] Dec 01 08:52:21 crc kubenswrapper[4813]: I1201 08:52:21.684382 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-f8648f98b-m47sh" Dec 01 08:52:21 crc kubenswrapper[4813]: I1201 08:52:21.685742 4813 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret" Dec 01 08:52:21 crc kubenswrapper[4813]: I1201 08:52:21.698413 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-f8648f98b-m47sh"] Dec 01 08:52:21 crc kubenswrapper[4813]: I1201 08:52:21.735427 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6lg8w\" (UniqueName: \"kubernetes.io/projected/f1cda8d0-a6b1-478a-9f8e-83d1fa070bf5-kube-api-access-6lg8w\") pod \"frr-k8s-zplkx\" (UID: \"f1cda8d0-a6b1-478a-9f8e-83d1fa070bf5\") " pod="metallb-system/frr-k8s-zplkx" Dec 01 08:52:21 crc kubenswrapper[4813]: I1201 08:52:21.735623 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/97457a58-aba0-4e0a-b812-ef4fd0912116-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-rgj8f\" (UID: \"97457a58-aba0-4e0a-b812-ef4fd0912116\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-rgj8f" Dec 01 08:52:21 crc kubenswrapper[4813]: I1201 08:52:21.735661 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/f1cda8d0-a6b1-478a-9f8e-83d1fa070bf5-reloader\") pod \"frr-k8s-zplkx\" (UID: \"f1cda8d0-a6b1-478a-9f8e-83d1fa070bf5\") " pod="metallb-system/frr-k8s-zplkx" Dec 01 08:52:21 crc kubenswrapper[4813]: I1201 08:52:21.735685 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/f1cda8d0-a6b1-478a-9f8e-83d1fa070bf5-frr-conf\") pod \"frr-k8s-zplkx\" (UID: \"f1cda8d0-a6b1-478a-9f8e-83d1fa070bf5\") " pod="metallb-system/frr-k8s-zplkx" Dec 01 08:52:21 crc kubenswrapper[4813]: I1201 08:52:21.735703 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/f1cda8d0-a6b1-478a-9f8e-83d1fa070bf5-metrics-certs\") pod \"frr-k8s-zplkx\" (UID: \"f1cda8d0-a6b1-478a-9f8e-83d1fa070bf5\") " pod="metallb-system/frr-k8s-zplkx" Dec 01 08:52:21 crc kubenswrapper[4813]: I1201 08:52:21.735718 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/f1cda8d0-a6b1-478a-9f8e-83d1fa070bf5-frr-startup\") pod \"frr-k8s-zplkx\" (UID: \"f1cda8d0-a6b1-478a-9f8e-83d1fa070bf5\") " pod="metallb-system/frr-k8s-zplkx" Dec 01 08:52:21 crc kubenswrapper[4813]: I1201 08:52:21.735734 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cvfsd\" (UniqueName: \"kubernetes.io/projected/97457a58-aba0-4e0a-b812-ef4fd0912116-kube-api-access-cvfsd\") pod \"frr-k8s-webhook-server-7fcb986d4-rgj8f\" (UID: \"97457a58-aba0-4e0a-b812-ef4fd0912116\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-rgj8f" Dec 01 08:52:21 crc kubenswrapper[4813]: I1201 08:52:21.735880 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/f1cda8d0-a6b1-478a-9f8e-83d1fa070bf5-frr-sockets\") pod \"frr-k8s-zplkx\" (UID: \"f1cda8d0-a6b1-478a-9f8e-83d1fa070bf5\") " pod="metallb-system/frr-k8s-zplkx" Dec 01 08:52:21 crc kubenswrapper[4813]: I1201 08:52:21.736016 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/f1cda8d0-a6b1-478a-9f8e-83d1fa070bf5-metrics\") pod \"frr-k8s-zplkx\" (UID: \"f1cda8d0-a6b1-478a-9f8e-83d1fa070bf5\") " pod="metallb-system/frr-k8s-zplkx" Dec 01 08:52:21 crc kubenswrapper[4813]: I1201 08:52:21.836768 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ce2e13a7-5901-440b-8a7c-6516b77aea04-metrics-certs\") pod \"controller-f8648f98b-m47sh\" (UID: \"ce2e13a7-5901-440b-8a7c-6516b77aea04\") " pod="metallb-system/controller-f8648f98b-m47sh" Dec 01 08:52:21 crc kubenswrapper[4813]: I1201 08:52:21.836847 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6lg8w\" (UniqueName: \"kubernetes.io/projected/f1cda8d0-a6b1-478a-9f8e-83d1fa070bf5-kube-api-access-6lg8w\") pod \"frr-k8s-zplkx\" (UID: \"f1cda8d0-a6b1-478a-9f8e-83d1fa070bf5\") " pod="metallb-system/frr-k8s-zplkx" Dec 01 08:52:21 crc kubenswrapper[4813]: I1201 08:52:21.836870 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/ce2e13a7-5901-440b-8a7c-6516b77aea04-cert\") pod \"controller-f8648f98b-m47sh\" (UID: \"ce2e13a7-5901-440b-8a7c-6516b77aea04\") " pod="metallb-system/controller-f8648f98b-m47sh" Dec 01 08:52:21 crc kubenswrapper[4813]: I1201 08:52:21.836895 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/97457a58-aba0-4e0a-b812-ef4fd0912116-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-rgj8f\" (UID: \"97457a58-aba0-4e0a-b812-ef4fd0912116\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-rgj8f" Dec 01 08:52:21 crc kubenswrapper[4813]: I1201 08:52:21.836948 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/1bf9a820-0d04-43e4-a59b-e6168a6e11fe-metrics-certs\") pod \"speaker-227b6\" (UID: \"1bf9a820-0d04-43e4-a59b-e6168a6e11fe\") " pod="metallb-system/speaker-227b6" Dec 01 08:52:21 crc kubenswrapper[4813]: I1201 08:52:21.836974 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jlstg\" (UniqueName: \"kubernetes.io/projected/1bf9a820-0d04-43e4-a59b-e6168a6e11fe-kube-api-access-jlstg\") pod \"speaker-227b6\" (UID: \"1bf9a820-0d04-43e4-a59b-e6168a6e11fe\") " pod="metallb-system/speaker-227b6" Dec 01 08:52:21 crc kubenswrapper[4813]: I1201 08:52:21.837034 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/f1cda8d0-a6b1-478a-9f8e-83d1fa070bf5-reloader\") pod \"frr-k8s-zplkx\" (UID: \"f1cda8d0-a6b1-478a-9f8e-83d1fa070bf5\") " pod="metallb-system/frr-k8s-zplkx" Dec 01 08:52:21 crc kubenswrapper[4813]: I1201 08:52:21.837061 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/1bf9a820-0d04-43e4-a59b-e6168a6e11fe-memberlist\") pod \"speaker-227b6\" (UID: \"1bf9a820-0d04-43e4-a59b-e6168a6e11fe\") " pod="metallb-system/speaker-227b6" Dec 01 08:52:21 crc kubenswrapper[4813]: I1201 08:52:21.837138 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/1bf9a820-0d04-43e4-a59b-e6168a6e11fe-metallb-excludel2\") pod \"speaker-227b6\" (UID: \"1bf9a820-0d04-43e4-a59b-e6168a6e11fe\") " pod="metallb-system/speaker-227b6" Dec 01 08:52:21 crc kubenswrapper[4813]: I1201 08:52:21.837171 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rjq2s\" (UniqueName: \"kubernetes.io/projected/ce2e13a7-5901-440b-8a7c-6516b77aea04-kube-api-access-rjq2s\") pod \"controller-f8648f98b-m47sh\" (UID: \"ce2e13a7-5901-440b-8a7c-6516b77aea04\") " pod="metallb-system/controller-f8648f98b-m47sh" Dec 01 08:52:21 crc kubenswrapper[4813]: I1201 08:52:21.837193 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/f1cda8d0-a6b1-478a-9f8e-83d1fa070bf5-frr-conf\") pod \"frr-k8s-zplkx\" (UID: \"f1cda8d0-a6b1-478a-9f8e-83d1fa070bf5\") " pod="metallb-system/frr-k8s-zplkx" Dec 01 08:52:21 crc kubenswrapper[4813]: I1201 08:52:21.837223 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/f1cda8d0-a6b1-478a-9f8e-83d1fa070bf5-metrics-certs\") pod \"frr-k8s-zplkx\" (UID: \"f1cda8d0-a6b1-478a-9f8e-83d1fa070bf5\") " pod="metallb-system/frr-k8s-zplkx" Dec 01 08:52:21 crc kubenswrapper[4813]: I1201 08:52:21.837240 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/f1cda8d0-a6b1-478a-9f8e-83d1fa070bf5-frr-startup\") pod \"frr-k8s-zplkx\" (UID: \"f1cda8d0-a6b1-478a-9f8e-83d1fa070bf5\") " pod="metallb-system/frr-k8s-zplkx" Dec 01 08:52:21 crc kubenswrapper[4813]: I1201 08:52:21.837258 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cvfsd\" (UniqueName: \"kubernetes.io/projected/97457a58-aba0-4e0a-b812-ef4fd0912116-kube-api-access-cvfsd\") pod \"frr-k8s-webhook-server-7fcb986d4-rgj8f\" (UID: \"97457a58-aba0-4e0a-b812-ef4fd0912116\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-rgj8f" Dec 01 08:52:21 crc kubenswrapper[4813]: I1201 08:52:21.837294 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/f1cda8d0-a6b1-478a-9f8e-83d1fa070bf5-frr-sockets\") pod \"frr-k8s-zplkx\" (UID: \"f1cda8d0-a6b1-478a-9f8e-83d1fa070bf5\") " pod="metallb-system/frr-k8s-zplkx" Dec 01 08:52:21 crc kubenswrapper[4813]: I1201 08:52:21.837315 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/f1cda8d0-a6b1-478a-9f8e-83d1fa070bf5-metrics\") pod \"frr-k8s-zplkx\" (UID: \"f1cda8d0-a6b1-478a-9f8e-83d1fa070bf5\") " pod="metallb-system/frr-k8s-zplkx" Dec 01 08:52:21 crc kubenswrapper[4813]: I1201 08:52:21.837746 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/f1cda8d0-a6b1-478a-9f8e-83d1fa070bf5-metrics\") pod \"frr-k8s-zplkx\" (UID: \"f1cda8d0-a6b1-478a-9f8e-83d1fa070bf5\") " pod="metallb-system/frr-k8s-zplkx" Dec 01 08:52:21 crc kubenswrapper[4813]: I1201 08:52:21.837745 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/f1cda8d0-a6b1-478a-9f8e-83d1fa070bf5-reloader\") pod \"frr-k8s-zplkx\" (UID: \"f1cda8d0-a6b1-478a-9f8e-83d1fa070bf5\") " pod="metallb-system/frr-k8s-zplkx" Dec 01 08:52:21 crc kubenswrapper[4813]: I1201 08:52:21.838110 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/f1cda8d0-a6b1-478a-9f8e-83d1fa070bf5-frr-sockets\") pod \"frr-k8s-zplkx\" (UID: \"f1cda8d0-a6b1-478a-9f8e-83d1fa070bf5\") " pod="metallb-system/frr-k8s-zplkx" Dec 01 08:52:21 crc kubenswrapper[4813]: I1201 08:52:21.838319 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/f1cda8d0-a6b1-478a-9f8e-83d1fa070bf5-frr-conf\") pod \"frr-k8s-zplkx\" (UID: \"f1cda8d0-a6b1-478a-9f8e-83d1fa070bf5\") " pod="metallb-system/frr-k8s-zplkx" Dec 01 08:52:21 crc kubenswrapper[4813]: I1201 08:52:21.862157 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6lg8w\" (UniqueName: \"kubernetes.io/projected/f1cda8d0-a6b1-478a-9f8e-83d1fa070bf5-kube-api-access-6lg8w\") pod \"frr-k8s-zplkx\" (UID: \"f1cda8d0-a6b1-478a-9f8e-83d1fa070bf5\") " pod="metallb-system/frr-k8s-zplkx" Dec 01 08:52:21 crc kubenswrapper[4813]: I1201 08:52:21.864442 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cvfsd\" (UniqueName: \"kubernetes.io/projected/97457a58-aba0-4e0a-b812-ef4fd0912116-kube-api-access-cvfsd\") pod \"frr-k8s-webhook-server-7fcb986d4-rgj8f\" (UID: \"97457a58-aba0-4e0a-b812-ef4fd0912116\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-rgj8f" Dec 01 08:52:21 crc kubenswrapper[4813]: I1201 08:52:21.938440 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ce2e13a7-5901-440b-8a7c-6516b77aea04-metrics-certs\") pod \"controller-f8648f98b-m47sh\" (UID: \"ce2e13a7-5901-440b-8a7c-6516b77aea04\") " pod="metallb-system/controller-f8648f98b-m47sh" Dec 01 08:52:21 crc kubenswrapper[4813]: I1201 08:52:21.938536 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/ce2e13a7-5901-440b-8a7c-6516b77aea04-cert\") pod \"controller-f8648f98b-m47sh\" (UID: \"ce2e13a7-5901-440b-8a7c-6516b77aea04\") " pod="metallb-system/controller-f8648f98b-m47sh" Dec 01 08:52:21 crc kubenswrapper[4813]: I1201 08:52:21.938607 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/1bf9a820-0d04-43e4-a59b-e6168a6e11fe-metrics-certs\") pod \"speaker-227b6\" (UID: \"1bf9a820-0d04-43e4-a59b-e6168a6e11fe\") " pod="metallb-system/speaker-227b6" Dec 01 08:52:21 crc kubenswrapper[4813]: I1201 08:52:21.938630 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jlstg\" (UniqueName: \"kubernetes.io/projected/1bf9a820-0d04-43e4-a59b-e6168a6e11fe-kube-api-access-jlstg\") pod \"speaker-227b6\" (UID: \"1bf9a820-0d04-43e4-a59b-e6168a6e11fe\") " pod="metallb-system/speaker-227b6" Dec 01 08:52:21 crc kubenswrapper[4813]: I1201 08:52:21.938674 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/1bf9a820-0d04-43e4-a59b-e6168a6e11fe-memberlist\") pod \"speaker-227b6\" (UID: \"1bf9a820-0d04-43e4-a59b-e6168a6e11fe\") " pod="metallb-system/speaker-227b6" Dec 01 08:52:21 crc kubenswrapper[4813]: I1201 08:52:21.938692 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/1bf9a820-0d04-43e4-a59b-e6168a6e11fe-metallb-excludel2\") pod \"speaker-227b6\" (UID: \"1bf9a820-0d04-43e4-a59b-e6168a6e11fe\") " pod="metallb-system/speaker-227b6" Dec 01 08:52:21 crc kubenswrapper[4813]: I1201 08:52:21.938723 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rjq2s\" (UniqueName: \"kubernetes.io/projected/ce2e13a7-5901-440b-8a7c-6516b77aea04-kube-api-access-rjq2s\") pod \"controller-f8648f98b-m47sh\" (UID: \"ce2e13a7-5901-440b-8a7c-6516b77aea04\") " pod="metallb-system/controller-f8648f98b-m47sh" Dec 01 08:52:21 crc kubenswrapper[4813]: E1201 08:52:21.939224 4813 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Dec 01 08:52:21 crc kubenswrapper[4813]: E1201 08:52:21.939225 4813 secret.go:188] Couldn't get secret metallb-system/speaker-certs-secret: secret "speaker-certs-secret" not found Dec 01 08:52:21 crc kubenswrapper[4813]: E1201 08:52:21.939349 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/1bf9a820-0d04-43e4-a59b-e6168a6e11fe-memberlist podName:1bf9a820-0d04-43e4-a59b-e6168a6e11fe nodeName:}" failed. No retries permitted until 2025-12-01 08:52:22.439306843 +0000 UTC m=+727.062228429 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/1bf9a820-0d04-43e4-a59b-e6168a6e11fe-memberlist") pod "speaker-227b6" (UID: "1bf9a820-0d04-43e4-a59b-e6168a6e11fe") : secret "metallb-memberlist" not found Dec 01 08:52:21 crc kubenswrapper[4813]: E1201 08:52:21.939436 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/1bf9a820-0d04-43e4-a59b-e6168a6e11fe-metrics-certs podName:1bf9a820-0d04-43e4-a59b-e6168a6e11fe nodeName:}" failed. No retries permitted until 2025-12-01 08:52:22.439415706 +0000 UTC m=+727.062337312 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/1bf9a820-0d04-43e4-a59b-e6168a6e11fe-metrics-certs") pod "speaker-227b6" (UID: "1bf9a820-0d04-43e4-a59b-e6168a6e11fe") : secret "speaker-certs-secret" not found Dec 01 08:52:21 crc kubenswrapper[4813]: I1201 08:52:21.939997 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/1bf9a820-0d04-43e4-a59b-e6168a6e11fe-metallb-excludel2\") pod \"speaker-227b6\" (UID: \"1bf9a820-0d04-43e4-a59b-e6168a6e11fe\") " pod="metallb-system/speaker-227b6" Dec 01 08:52:21 crc kubenswrapper[4813]: I1201 08:52:21.942679 4813 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Dec 01 08:52:21 crc kubenswrapper[4813]: I1201 08:52:21.943126 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ce2e13a7-5901-440b-8a7c-6516b77aea04-metrics-certs\") pod \"controller-f8648f98b-m47sh\" (UID: \"ce2e13a7-5901-440b-8a7c-6516b77aea04\") " pod="metallb-system/controller-f8648f98b-m47sh" Dec 01 08:52:21 crc kubenswrapper[4813]: I1201 08:52:21.951970 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/ce2e13a7-5901-440b-8a7c-6516b77aea04-cert\") pod \"controller-f8648f98b-m47sh\" (UID: \"ce2e13a7-5901-440b-8a7c-6516b77aea04\") " pod="metallb-system/controller-f8648f98b-m47sh" Dec 01 08:52:21 crc kubenswrapper[4813]: I1201 08:52:21.956062 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jlstg\" (UniqueName: \"kubernetes.io/projected/1bf9a820-0d04-43e4-a59b-e6168a6e11fe-kube-api-access-jlstg\") pod \"speaker-227b6\" (UID: \"1bf9a820-0d04-43e4-a59b-e6168a6e11fe\") " pod="metallb-system/speaker-227b6" Dec 01 08:52:21 crc kubenswrapper[4813]: I1201 08:52:21.957458 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rjq2s\" (UniqueName: \"kubernetes.io/projected/ce2e13a7-5901-440b-8a7c-6516b77aea04-kube-api-access-rjq2s\") pod \"controller-f8648f98b-m47sh\" (UID: \"ce2e13a7-5901-440b-8a7c-6516b77aea04\") " pod="metallb-system/controller-f8648f98b-m47sh" Dec 01 08:52:21 crc kubenswrapper[4813]: I1201 08:52:21.997486 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-f8648f98b-m47sh" Dec 01 08:52:22 crc kubenswrapper[4813]: I1201 08:52:22.378802 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup" Dec 01 08:52:22 crc kubenswrapper[4813]: I1201 08:52:22.388517 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/f1cda8d0-a6b1-478a-9f8e-83d1fa070bf5-frr-startup\") pod \"frr-k8s-zplkx\" (UID: \"f1cda8d0-a6b1-478a-9f8e-83d1fa070bf5\") " pod="metallb-system/frr-k8s-zplkx" Dec 01 08:52:22 crc kubenswrapper[4813]: I1201 08:52:22.412457 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-f8648f98b-m47sh"] Dec 01 08:52:22 crc kubenswrapper[4813]: I1201 08:52:22.446100 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/1bf9a820-0d04-43e4-a59b-e6168a6e11fe-memberlist\") pod \"speaker-227b6\" (UID: \"1bf9a820-0d04-43e4-a59b-e6168a6e11fe\") " pod="metallb-system/speaker-227b6" Dec 01 08:52:22 crc kubenswrapper[4813]: I1201 08:52:22.446271 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/1bf9a820-0d04-43e4-a59b-e6168a6e11fe-metrics-certs\") pod \"speaker-227b6\" (UID: \"1bf9a820-0d04-43e4-a59b-e6168a6e11fe\") " pod="metallb-system/speaker-227b6" Dec 01 08:52:22 crc kubenswrapper[4813]: E1201 08:52:22.446650 4813 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Dec 01 08:52:22 crc kubenswrapper[4813]: E1201 08:52:22.446788 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/1bf9a820-0d04-43e4-a59b-e6168a6e11fe-memberlist podName:1bf9a820-0d04-43e4-a59b-e6168a6e11fe nodeName:}" failed. No retries permitted until 2025-12-01 08:52:23.446749258 +0000 UTC m=+728.069670904 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/1bf9a820-0d04-43e4-a59b-e6168a6e11fe-memberlist") pod "speaker-227b6" (UID: "1bf9a820-0d04-43e4-a59b-e6168a6e11fe") : secret "metallb-memberlist" not found Dec 01 08:52:22 crc kubenswrapper[4813]: I1201 08:52:22.454353 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/1bf9a820-0d04-43e4-a59b-e6168a6e11fe-metrics-certs\") pod \"speaker-227b6\" (UID: \"1bf9a820-0d04-43e4-a59b-e6168a6e11fe\") " pod="metallb-system/speaker-227b6" Dec 01 08:52:22 crc kubenswrapper[4813]: I1201 08:52:22.662811 4813 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret" Dec 01 08:52:22 crc kubenswrapper[4813]: I1201 08:52:22.671265 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/f1cda8d0-a6b1-478a-9f8e-83d1fa070bf5-metrics-certs\") pod \"frr-k8s-zplkx\" (UID: \"f1cda8d0-a6b1-478a-9f8e-83d1fa070bf5\") " pod="metallb-system/frr-k8s-zplkx" Dec 01 08:52:22 crc kubenswrapper[4813]: I1201 08:52:22.694302 4813 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-daemon-dockercfg-dfhlr" Dec 01 08:52:22 crc kubenswrapper[4813]: I1201 08:52:22.778186 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-zplkx" Dec 01 08:52:22 crc kubenswrapper[4813]: E1201 08:52:22.837764 4813 secret.go:188] Couldn't get secret metallb-system/frr-k8s-webhook-server-cert: failed to sync secret cache: timed out waiting for the condition Dec 01 08:52:22 crc kubenswrapper[4813]: E1201 08:52:22.838004 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/97457a58-aba0-4e0a-b812-ef4fd0912116-cert podName:97457a58-aba0-4e0a-b812-ef4fd0912116 nodeName:}" failed. No retries permitted until 2025-12-01 08:52:23.337949075 +0000 UTC m=+727.960870681 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/97457a58-aba0-4e0a-b812-ef4fd0912116-cert") pod "frr-k8s-webhook-server-7fcb986d4-rgj8f" (UID: "97457a58-aba0-4e0a-b812-ef4fd0912116") : failed to sync secret cache: timed out waiting for the condition Dec 01 08:52:22 crc kubenswrapper[4813]: I1201 08:52:22.850948 4813 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert" Dec 01 08:52:23 crc kubenswrapper[4813]: I1201 08:52:23.350816 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-zplkx" event={"ID":"f1cda8d0-a6b1-478a-9f8e-83d1fa070bf5","Type":"ContainerStarted","Data":"79f799850ed980b3692602061c312cd31a25ee402de881cad22166e32bbc1cf1"} Dec 01 08:52:23 crc kubenswrapper[4813]: I1201 08:52:23.352808 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-f8648f98b-m47sh" event={"ID":"ce2e13a7-5901-440b-8a7c-6516b77aea04","Type":"ContainerStarted","Data":"556b91155635494e5a3a18a773d06bf26a43977d2a134388e32eefe779ffb6b8"} Dec 01 08:52:23 crc kubenswrapper[4813]: I1201 08:52:23.352840 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-f8648f98b-m47sh" event={"ID":"ce2e13a7-5901-440b-8a7c-6516b77aea04","Type":"ContainerStarted","Data":"53e830beb112f0e30903a4998a2cfa109a7c56da15a5865d187dd7e771f507ec"} Dec 01 08:52:23 crc kubenswrapper[4813]: I1201 08:52:23.363579 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/97457a58-aba0-4e0a-b812-ef4fd0912116-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-rgj8f\" (UID: \"97457a58-aba0-4e0a-b812-ef4fd0912116\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-rgj8f" Dec 01 08:52:23 crc kubenswrapper[4813]: I1201 08:52:23.369163 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/97457a58-aba0-4e0a-b812-ef4fd0912116-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-rgj8f\" (UID: \"97457a58-aba0-4e0a-b812-ef4fd0912116\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-rgj8f" Dec 01 08:52:23 crc kubenswrapper[4813]: I1201 08:52:23.386387 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-rgj8f" Dec 01 08:52:23 crc kubenswrapper[4813]: I1201 08:52:23.468199 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/1bf9a820-0d04-43e4-a59b-e6168a6e11fe-memberlist\") pod \"speaker-227b6\" (UID: \"1bf9a820-0d04-43e4-a59b-e6168a6e11fe\") " pod="metallb-system/speaker-227b6" Dec 01 08:52:23 crc kubenswrapper[4813]: I1201 08:52:23.483277 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/1bf9a820-0d04-43e4-a59b-e6168a6e11fe-memberlist\") pod \"speaker-227b6\" (UID: \"1bf9a820-0d04-43e4-a59b-e6168a6e11fe\") " pod="metallb-system/speaker-227b6" Dec 01 08:52:23 crc kubenswrapper[4813]: I1201 08:52:23.616238 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-7fcb986d4-rgj8f"] Dec 01 08:52:23 crc kubenswrapper[4813]: W1201 08:52:23.623629 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod97457a58_aba0_4e0a_b812_ef4fd0912116.slice/crio-a4ac1502177784646aa0d2f2d66a52bcadf8aa82a1d036647c25f75179860612 WatchSource:0}: Error finding container a4ac1502177784646aa0d2f2d66a52bcadf8aa82a1d036647c25f75179860612: Status 404 returned error can't find the container with id a4ac1502177784646aa0d2f2d66a52bcadf8aa82a1d036647c25f75179860612 Dec 01 08:52:23 crc kubenswrapper[4813]: I1201 08:52:23.772466 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-227b6" Dec 01 08:52:23 crc kubenswrapper[4813]: W1201 08:52:23.814947 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1bf9a820_0d04_43e4_a59b_e6168a6e11fe.slice/crio-a37a07923a765dfb4e3f1f7f06aa62c8066942d213d2c90ebbb372f78fc05c64 WatchSource:0}: Error finding container a37a07923a765dfb4e3f1f7f06aa62c8066942d213d2c90ebbb372f78fc05c64: Status 404 returned error can't find the container with id a37a07923a765dfb4e3f1f7f06aa62c8066942d213d2c90ebbb372f78fc05c64 Dec 01 08:52:24 crc kubenswrapper[4813]: I1201 08:52:24.363332 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-227b6" event={"ID":"1bf9a820-0d04-43e4-a59b-e6168a6e11fe","Type":"ContainerStarted","Data":"61b172e06b891d355ddd9e1cc2ff1ec1d2e64cb20b50b6a69f68d61026fb9521"} Dec 01 08:52:24 crc kubenswrapper[4813]: I1201 08:52:24.363408 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-227b6" event={"ID":"1bf9a820-0d04-43e4-a59b-e6168a6e11fe","Type":"ContainerStarted","Data":"a37a07923a765dfb4e3f1f7f06aa62c8066942d213d2c90ebbb372f78fc05c64"} Dec 01 08:52:24 crc kubenswrapper[4813]: I1201 08:52:24.364205 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-rgj8f" event={"ID":"97457a58-aba0-4e0a-b812-ef4fd0912116","Type":"ContainerStarted","Data":"a4ac1502177784646aa0d2f2d66a52bcadf8aa82a1d036647c25f75179860612"} Dec 01 08:52:27 crc kubenswrapper[4813]: I1201 08:52:27.407579 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-f8648f98b-m47sh" event={"ID":"ce2e13a7-5901-440b-8a7c-6516b77aea04","Type":"ContainerStarted","Data":"6ca869d709dc67e70e3335dc551f90dbd48e0a5bb13c6ccad34e160aaea51c78"} Dec 01 08:52:27 crc kubenswrapper[4813]: I1201 08:52:27.408433 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-f8648f98b-m47sh" Dec 01 08:52:27 crc kubenswrapper[4813]: I1201 08:52:27.409782 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-227b6" event={"ID":"1bf9a820-0d04-43e4-a59b-e6168a6e11fe","Type":"ContainerStarted","Data":"2ba887da16f2201c274044d95eb8bb3c141e37e130de6523490b0e9c0e74df4b"} Dec 01 08:52:27 crc kubenswrapper[4813]: I1201 08:52:27.409974 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-227b6" Dec 01 08:52:27 crc kubenswrapper[4813]: I1201 08:52:27.425290 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/controller-f8648f98b-m47sh" podStartSLOduration=2.201112494 podStartE2EDuration="6.425235387s" podCreationTimestamp="2025-12-01 08:52:21 +0000 UTC" firstStartedPulling="2025-12-01 08:52:22.591730194 +0000 UTC m=+727.214651770" lastFinishedPulling="2025-12-01 08:52:26.815853077 +0000 UTC m=+731.438774663" observedRunningTime="2025-12-01 08:52:27.424359643 +0000 UTC m=+732.047281229" watchObservedRunningTime="2025-12-01 08:52:27.425235387 +0000 UTC m=+732.048156993" Dec 01 08:52:27 crc kubenswrapper[4813]: I1201 08:52:27.441526 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/speaker-227b6" podStartSLOduration=3.841523983 podStartE2EDuration="6.441505652s" podCreationTimestamp="2025-12-01 08:52:21 +0000 UTC" firstStartedPulling="2025-12-01 08:52:24.217399419 +0000 UTC m=+728.840320995" lastFinishedPulling="2025-12-01 08:52:26.817381078 +0000 UTC m=+731.440302664" observedRunningTime="2025-12-01 08:52:27.440654229 +0000 UTC m=+732.063575815" watchObservedRunningTime="2025-12-01 08:52:27.441505652 +0000 UTC m=+732.064427228" Dec 01 08:52:30 crc kubenswrapper[4813]: I1201 08:52:30.438282 4813 generic.go:334] "Generic (PLEG): container finished" podID="f1cda8d0-a6b1-478a-9f8e-83d1fa070bf5" containerID="244d6435bc7acccc5fac44b9c859e80783a5af5e986f16d82e0d535a1edf6862" exitCode=0 Dec 01 08:52:30 crc kubenswrapper[4813]: I1201 08:52:30.438440 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-zplkx" event={"ID":"f1cda8d0-a6b1-478a-9f8e-83d1fa070bf5","Type":"ContainerDied","Data":"244d6435bc7acccc5fac44b9c859e80783a5af5e986f16d82e0d535a1edf6862"} Dec 01 08:52:30 crc kubenswrapper[4813]: I1201 08:52:30.440563 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-rgj8f" event={"ID":"97457a58-aba0-4e0a-b812-ef4fd0912116","Type":"ContainerStarted","Data":"5552408d273eed969e8c61afc8d3542c3bdc23b9798b7fd46d051d0d382e6822"} Dec 01 08:52:30 crc kubenswrapper[4813]: I1201 08:52:30.441274 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-rgj8f" Dec 01 08:52:30 crc kubenswrapper[4813]: I1201 08:52:30.497759 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-rgj8f" podStartSLOduration=3.1335483379999998 podStartE2EDuration="9.497735266s" podCreationTimestamp="2025-12-01 08:52:21 +0000 UTC" firstStartedPulling="2025-12-01 08:52:23.626308859 +0000 UTC m=+728.249230435" lastFinishedPulling="2025-12-01 08:52:29.990495777 +0000 UTC m=+734.613417363" observedRunningTime="2025-12-01 08:52:30.494717126 +0000 UTC m=+735.117638732" watchObservedRunningTime="2025-12-01 08:52:30.497735266 +0000 UTC m=+735.120656852" Dec 01 08:52:31 crc kubenswrapper[4813]: I1201 08:52:31.472800 4813 generic.go:334] "Generic (PLEG): container finished" podID="f1cda8d0-a6b1-478a-9f8e-83d1fa070bf5" containerID="f79faaa8083568931ffa686cd66ce0e0d0afe97dc47a6d93e04bff9812eeb2ce" exitCode=0 Dec 01 08:52:31 crc kubenswrapper[4813]: I1201 08:52:31.472918 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-zplkx" event={"ID":"f1cda8d0-a6b1-478a-9f8e-83d1fa070bf5","Type":"ContainerDied","Data":"f79faaa8083568931ffa686cd66ce0e0d0afe97dc47a6d93e04bff9812eeb2ce"} Dec 01 08:52:32 crc kubenswrapper[4813]: I1201 08:52:32.481882 4813 generic.go:334] "Generic (PLEG): container finished" podID="f1cda8d0-a6b1-478a-9f8e-83d1fa070bf5" containerID="476398796bdf7f2590a42b461daa4ad618b0cb90214faff47549c86079c19d12" exitCode=0 Dec 01 08:52:32 crc kubenswrapper[4813]: I1201 08:52:32.481944 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-zplkx" event={"ID":"f1cda8d0-a6b1-478a-9f8e-83d1fa070bf5","Type":"ContainerDied","Data":"476398796bdf7f2590a42b461daa4ad618b0cb90214faff47549c86079c19d12"} Dec 01 08:52:33 crc kubenswrapper[4813]: I1201 08:52:33.491557 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-zplkx" event={"ID":"f1cda8d0-a6b1-478a-9f8e-83d1fa070bf5","Type":"ContainerStarted","Data":"327a358ed8dad921cf5b3f49a9729715c2fe454ac3272779b85a2f7539bd1ae5"} Dec 01 08:52:33 crc kubenswrapper[4813]: I1201 08:52:33.492111 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-zplkx" event={"ID":"f1cda8d0-a6b1-478a-9f8e-83d1fa070bf5","Type":"ContainerStarted","Data":"592df4cc69f7937f5dca60f462a73f799c175f7080ff9dff1f5f84618a014c82"} Dec 01 08:52:33 crc kubenswrapper[4813]: I1201 08:52:33.492124 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-zplkx" event={"ID":"f1cda8d0-a6b1-478a-9f8e-83d1fa070bf5","Type":"ContainerStarted","Data":"fc7d3e38807ac0dfb457503ccb81b126e5ec26901cadcfc8784f070bf969c2d0"} Dec 01 08:52:33 crc kubenswrapper[4813]: I1201 08:52:33.492134 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-zplkx" event={"ID":"f1cda8d0-a6b1-478a-9f8e-83d1fa070bf5","Type":"ContainerStarted","Data":"b69f4ba9d4140f800932e8fb7a4637e3d76e8df3ab541ae5736eb08da6e24273"} Dec 01 08:52:33 crc kubenswrapper[4813]: I1201 08:52:33.492144 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-zplkx" event={"ID":"f1cda8d0-a6b1-478a-9f8e-83d1fa070bf5","Type":"ContainerStarted","Data":"d7c3e38966fc5067d885b24ddc5482c321c89e39e3eb6a0231bef3eac1b84c59"} Dec 01 08:52:34 crc kubenswrapper[4813]: I1201 08:52:34.503160 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-zplkx" event={"ID":"f1cda8d0-a6b1-478a-9f8e-83d1fa070bf5","Type":"ContainerStarted","Data":"cacdb5a0a0ea940aa5f611e154e31d280a56ddc33db787daf95a87a9c685db20"} Dec 01 08:52:34 crc kubenswrapper[4813]: I1201 08:52:34.503801 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-zplkx" Dec 01 08:52:34 crc kubenswrapper[4813]: I1201 08:52:34.542643 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-zplkx" podStartSLOduration=6.4759532029999995 podStartE2EDuration="13.542578518s" podCreationTimestamp="2025-12-01 08:52:21 +0000 UTC" firstStartedPulling="2025-12-01 08:52:22.901116694 +0000 UTC m=+727.524038280" lastFinishedPulling="2025-12-01 08:52:29.967742009 +0000 UTC m=+734.590663595" observedRunningTime="2025-12-01 08:52:34.537981995 +0000 UTC m=+739.160903611" watchObservedRunningTime="2025-12-01 08:52:34.542578518 +0000 UTC m=+739.165500124" Dec 01 08:52:37 crc kubenswrapper[4813]: I1201 08:52:37.778441 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="metallb-system/frr-k8s-zplkx" Dec 01 08:52:37 crc kubenswrapper[4813]: I1201 08:52:37.839514 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="metallb-system/frr-k8s-zplkx" Dec 01 08:52:42 crc kubenswrapper[4813]: I1201 08:52:42.001492 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/controller-f8648f98b-m47sh" Dec 01 08:52:42 crc kubenswrapper[4813]: I1201 08:52:42.788585 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-zplkx" Dec 01 08:52:43 crc kubenswrapper[4813]: I1201 08:52:43.394066 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-rgj8f" Dec 01 08:52:43 crc kubenswrapper[4813]: I1201 08:52:43.776866 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/speaker-227b6" Dec 01 08:52:47 crc kubenswrapper[4813]: I1201 08:52:47.222142 4813 patch_prober.go:28] interesting pod/machine-config-daemon-9wn4q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 08:52:47 crc kubenswrapper[4813]: I1201 08:52:47.222635 4813 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9wn4q" podUID="3d423402-e08f-42e9-b60c-6eb9bf2f6d48" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 08:52:47 crc kubenswrapper[4813]: I1201 08:52:47.222757 4813 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-9wn4q" Dec 01 08:52:47 crc kubenswrapper[4813]: I1201 08:52:47.223797 4813 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"c9bc69c1312504b715819da3ef57a1ae8b88ab55017f5060cf77c80a5fa0c1e8"} pod="openshift-machine-config-operator/machine-config-daemon-9wn4q" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 01 08:52:47 crc kubenswrapper[4813]: I1201 08:52:47.224019 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-9wn4q" podUID="3d423402-e08f-42e9-b60c-6eb9bf2f6d48" containerName="machine-config-daemon" containerID="cri-o://c9bc69c1312504b715819da3ef57a1ae8b88ab55017f5060cf77c80a5fa0c1e8" gracePeriod=600 Dec 01 08:52:47 crc kubenswrapper[4813]: I1201 08:52:47.686837 4813 generic.go:334] "Generic (PLEG): container finished" podID="3d423402-e08f-42e9-b60c-6eb9bf2f6d48" containerID="c9bc69c1312504b715819da3ef57a1ae8b88ab55017f5060cf77c80a5fa0c1e8" exitCode=0 Dec 01 08:52:47 crc kubenswrapper[4813]: I1201 08:52:47.686888 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-9wn4q" event={"ID":"3d423402-e08f-42e9-b60c-6eb9bf2f6d48","Type":"ContainerDied","Data":"c9bc69c1312504b715819da3ef57a1ae8b88ab55017f5060cf77c80a5fa0c1e8"} Dec 01 08:52:47 crc kubenswrapper[4813]: I1201 08:52:47.686945 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-9wn4q" event={"ID":"3d423402-e08f-42e9-b60c-6eb9bf2f6d48","Type":"ContainerStarted","Data":"bc619037491f7f2216e483229bcdbf7a96de1d588814d7688551705019230c50"} Dec 01 08:52:47 crc kubenswrapper[4813]: I1201 08:52:47.687048 4813 scope.go:117] "RemoveContainer" containerID="e6f5f635caf35fdd2a0d88216d0cd5f28601a09788d14975d2ea2dca78233ac3" Dec 01 08:52:53 crc kubenswrapper[4813]: I1201 08:52:53.501514 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-index-z8nm9"] Dec 01 08:52:53 crc kubenswrapper[4813]: I1201 08:52:53.506046 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-index-z8nm9" Dec 01 08:52:53 crc kubenswrapper[4813]: I1201 08:52:53.509433 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-index-dockercfg-kjmkp" Dec 01 08:52:53 crc kubenswrapper[4813]: I1201 08:52:53.511733 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Dec 01 08:52:53 crc kubenswrapper[4813]: I1201 08:52:53.513926 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Dec 01 08:52:53 crc kubenswrapper[4813]: I1201 08:52:53.518791 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-index-z8nm9"] Dec 01 08:52:53 crc kubenswrapper[4813]: I1201 08:52:53.563080 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wlhbf\" (UniqueName: \"kubernetes.io/projected/c4d16fd1-e8d3-4bb5-af1f-204e484c8530-kube-api-access-wlhbf\") pod \"mariadb-operator-index-z8nm9\" (UID: \"c4d16fd1-e8d3-4bb5-af1f-204e484c8530\") " pod="openstack-operators/mariadb-operator-index-z8nm9" Dec 01 08:52:53 crc kubenswrapper[4813]: I1201 08:52:53.663885 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wlhbf\" (UniqueName: \"kubernetes.io/projected/c4d16fd1-e8d3-4bb5-af1f-204e484c8530-kube-api-access-wlhbf\") pod \"mariadb-operator-index-z8nm9\" (UID: \"c4d16fd1-e8d3-4bb5-af1f-204e484c8530\") " pod="openstack-operators/mariadb-operator-index-z8nm9" Dec 01 08:52:53 crc kubenswrapper[4813]: I1201 08:52:53.682866 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wlhbf\" (UniqueName: \"kubernetes.io/projected/c4d16fd1-e8d3-4bb5-af1f-204e484c8530-kube-api-access-wlhbf\") pod \"mariadb-operator-index-z8nm9\" (UID: \"c4d16fd1-e8d3-4bb5-af1f-204e484c8530\") " pod="openstack-operators/mariadb-operator-index-z8nm9" Dec 01 08:52:53 crc kubenswrapper[4813]: I1201 08:52:53.865223 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-index-z8nm9" Dec 01 08:52:54 crc kubenswrapper[4813]: I1201 08:52:54.368134 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-index-z8nm9"] Dec 01 08:52:54 crc kubenswrapper[4813]: I1201 08:52:54.746722 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-index-z8nm9" event={"ID":"c4d16fd1-e8d3-4bb5-af1f-204e484c8530","Type":"ContainerStarted","Data":"d96636e0a20ffd8ed70847e315e9ff215bb3c73fc947fae546d00aee27a50e41"} Dec 01 08:53:02 crc kubenswrapper[4813]: I1201 08:53:02.856199 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-index-z8nm9" event={"ID":"c4d16fd1-e8d3-4bb5-af1f-204e484c8530","Type":"ContainerStarted","Data":"c7d761c15931bdb12ada0b86c51712ab26b21bf2c891adcc8553e58602301b2f"} Dec 01 08:53:02 crc kubenswrapper[4813]: I1201 08:53:02.878092 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-index-z8nm9" podStartSLOduration=2.178535481 podStartE2EDuration="9.878042754s" podCreationTimestamp="2025-12-01 08:52:53 +0000 UTC" firstStartedPulling="2025-12-01 08:52:54.391190274 +0000 UTC m=+759.014111870" lastFinishedPulling="2025-12-01 08:53:02.090697557 +0000 UTC m=+766.713619143" observedRunningTime="2025-12-01 08:53:02.872793823 +0000 UTC m=+767.495715429" watchObservedRunningTime="2025-12-01 08:53:02.878042754 +0000 UTC m=+767.500964350" Dec 01 08:53:03 crc kubenswrapper[4813]: I1201 08:53:03.865627 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-index-z8nm9" Dec 01 08:53:03 crc kubenswrapper[4813]: I1201 08:53:03.865701 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/mariadb-operator-index-z8nm9" Dec 01 08:53:03 crc kubenswrapper[4813]: I1201 08:53:03.911599 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/mariadb-operator-index-z8nm9" Dec 01 08:53:13 crc kubenswrapper[4813]: I1201 08:53:13.925936 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-index-z8nm9" Dec 01 08:53:22 crc kubenswrapper[4813]: I1201 08:53:22.971003 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534fxznfm"] Dec 01 08:53:22 crc kubenswrapper[4813]: I1201 08:53:22.974129 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534fxznfm" Dec 01 08:53:22 crc kubenswrapper[4813]: I1201 08:53:22.977210 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-fml9r" Dec 01 08:53:23 crc kubenswrapper[4813]: I1201 08:53:23.015654 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534fxznfm"] Dec 01 08:53:23 crc kubenswrapper[4813]: I1201 08:53:23.129694 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/348e5b65-5859-4870-aaa1-85100418c446-util\") pod \"27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534fxznfm\" (UID: \"348e5b65-5859-4870-aaa1-85100418c446\") " pod="openstack-operators/27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534fxznfm" Dec 01 08:53:23 crc kubenswrapper[4813]: I1201 08:53:23.130098 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/348e5b65-5859-4870-aaa1-85100418c446-bundle\") pod \"27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534fxznfm\" (UID: \"348e5b65-5859-4870-aaa1-85100418c446\") " pod="openstack-operators/27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534fxznfm" Dec 01 08:53:23 crc kubenswrapper[4813]: I1201 08:53:23.130242 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6lx77\" (UniqueName: \"kubernetes.io/projected/348e5b65-5859-4870-aaa1-85100418c446-kube-api-access-6lx77\") pod \"27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534fxznfm\" (UID: \"348e5b65-5859-4870-aaa1-85100418c446\") " pod="openstack-operators/27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534fxznfm" Dec 01 08:53:23 crc kubenswrapper[4813]: I1201 08:53:23.231177 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/348e5b65-5859-4870-aaa1-85100418c446-util\") pod \"27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534fxznfm\" (UID: \"348e5b65-5859-4870-aaa1-85100418c446\") " pod="openstack-operators/27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534fxznfm" Dec 01 08:53:23 crc kubenswrapper[4813]: I1201 08:53:23.231301 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/348e5b65-5859-4870-aaa1-85100418c446-bundle\") pod \"27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534fxznfm\" (UID: \"348e5b65-5859-4870-aaa1-85100418c446\") " pod="openstack-operators/27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534fxznfm" Dec 01 08:53:23 crc kubenswrapper[4813]: I1201 08:53:23.231350 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6lx77\" (UniqueName: \"kubernetes.io/projected/348e5b65-5859-4870-aaa1-85100418c446-kube-api-access-6lx77\") pod \"27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534fxznfm\" (UID: \"348e5b65-5859-4870-aaa1-85100418c446\") " pod="openstack-operators/27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534fxznfm" Dec 01 08:53:23 crc kubenswrapper[4813]: I1201 08:53:23.231991 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/348e5b65-5859-4870-aaa1-85100418c446-util\") pod \"27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534fxznfm\" (UID: \"348e5b65-5859-4870-aaa1-85100418c446\") " pod="openstack-operators/27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534fxznfm" Dec 01 08:53:23 crc kubenswrapper[4813]: I1201 08:53:23.232441 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/348e5b65-5859-4870-aaa1-85100418c446-bundle\") pod \"27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534fxznfm\" (UID: \"348e5b65-5859-4870-aaa1-85100418c446\") " pod="openstack-operators/27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534fxznfm" Dec 01 08:53:23 crc kubenswrapper[4813]: I1201 08:53:23.263947 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6lx77\" (UniqueName: \"kubernetes.io/projected/348e5b65-5859-4870-aaa1-85100418c446-kube-api-access-6lx77\") pod \"27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534fxznfm\" (UID: \"348e5b65-5859-4870-aaa1-85100418c446\") " pod="openstack-operators/27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534fxznfm" Dec 01 08:53:23 crc kubenswrapper[4813]: I1201 08:53:23.306292 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534fxznfm" Dec 01 08:53:23 crc kubenswrapper[4813]: I1201 08:53:23.546321 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534fxznfm"] Dec 01 08:53:24 crc kubenswrapper[4813]: I1201 08:53:24.002518 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534fxznfm" event={"ID":"348e5b65-5859-4870-aaa1-85100418c446","Type":"ContainerStarted","Data":"d256c046466a493e42913f1f5e3ff9dba71296e78f801ee1ba3d7f51edc4bbdf"} Dec 01 08:53:28 crc kubenswrapper[4813]: I1201 08:53:28.038663 4813 generic.go:334] "Generic (PLEG): container finished" podID="348e5b65-5859-4870-aaa1-85100418c446" containerID="6728f57e2f7744cf7ec174d5a9f8e49de494f20365bc3d286390a66b714260e0" exitCode=0 Dec 01 08:53:28 crc kubenswrapper[4813]: I1201 08:53:28.039068 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534fxznfm" event={"ID":"348e5b65-5859-4870-aaa1-85100418c446","Type":"ContainerDied","Data":"6728f57e2f7744cf7ec174d5a9f8e49de494f20365bc3d286390a66b714260e0"} Dec 01 08:53:30 crc kubenswrapper[4813]: I1201 08:53:30.057395 4813 generic.go:334] "Generic (PLEG): container finished" podID="348e5b65-5859-4870-aaa1-85100418c446" containerID="be43a5c128772057e5349a736cc9bad664c0f0c15e38b3b70ab423e136bca998" exitCode=0 Dec 01 08:53:30 crc kubenswrapper[4813]: I1201 08:53:30.057502 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534fxznfm" event={"ID":"348e5b65-5859-4870-aaa1-85100418c446","Type":"ContainerDied","Data":"be43a5c128772057e5349a736cc9bad664c0f0c15e38b3b70ab423e136bca998"} Dec 01 08:53:31 crc kubenswrapper[4813]: I1201 08:53:31.069783 4813 generic.go:334] "Generic (PLEG): container finished" podID="348e5b65-5859-4870-aaa1-85100418c446" containerID="b1134778a50ac07e3bd0a14620a57aaf591e9bb521245d9851ad3a1f57aea8b8" exitCode=0 Dec 01 08:53:31 crc kubenswrapper[4813]: I1201 08:53:31.069894 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534fxznfm" event={"ID":"348e5b65-5859-4870-aaa1-85100418c446","Type":"ContainerDied","Data":"b1134778a50ac07e3bd0a14620a57aaf591e9bb521245d9851ad3a1f57aea8b8"} Dec 01 08:53:32 crc kubenswrapper[4813]: I1201 08:53:32.421614 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534fxznfm" Dec 01 08:53:32 crc kubenswrapper[4813]: I1201 08:53:32.595187 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/348e5b65-5859-4870-aaa1-85100418c446-util\") pod \"348e5b65-5859-4870-aaa1-85100418c446\" (UID: \"348e5b65-5859-4870-aaa1-85100418c446\") " Dec 01 08:53:32 crc kubenswrapper[4813]: I1201 08:53:32.595259 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/348e5b65-5859-4870-aaa1-85100418c446-bundle\") pod \"348e5b65-5859-4870-aaa1-85100418c446\" (UID: \"348e5b65-5859-4870-aaa1-85100418c446\") " Dec 01 08:53:32 crc kubenswrapper[4813]: I1201 08:53:32.595308 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6lx77\" (UniqueName: \"kubernetes.io/projected/348e5b65-5859-4870-aaa1-85100418c446-kube-api-access-6lx77\") pod \"348e5b65-5859-4870-aaa1-85100418c446\" (UID: \"348e5b65-5859-4870-aaa1-85100418c446\") " Dec 01 08:53:32 crc kubenswrapper[4813]: I1201 08:53:32.598060 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/348e5b65-5859-4870-aaa1-85100418c446-bundle" (OuterVolumeSpecName: "bundle") pod "348e5b65-5859-4870-aaa1-85100418c446" (UID: "348e5b65-5859-4870-aaa1-85100418c446"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 08:53:32 crc kubenswrapper[4813]: I1201 08:53:32.605412 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/348e5b65-5859-4870-aaa1-85100418c446-kube-api-access-6lx77" (OuterVolumeSpecName: "kube-api-access-6lx77") pod "348e5b65-5859-4870-aaa1-85100418c446" (UID: "348e5b65-5859-4870-aaa1-85100418c446"). InnerVolumeSpecName "kube-api-access-6lx77". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:53:32 crc kubenswrapper[4813]: I1201 08:53:32.697325 4813 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/348e5b65-5859-4870-aaa1-85100418c446-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 08:53:32 crc kubenswrapper[4813]: I1201 08:53:32.697384 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6lx77\" (UniqueName: \"kubernetes.io/projected/348e5b65-5859-4870-aaa1-85100418c446-kube-api-access-6lx77\") on node \"crc\" DevicePath \"\"" Dec 01 08:53:32 crc kubenswrapper[4813]: I1201 08:53:32.913193 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/348e5b65-5859-4870-aaa1-85100418c446-util" (OuterVolumeSpecName: "util") pod "348e5b65-5859-4870-aaa1-85100418c446" (UID: "348e5b65-5859-4870-aaa1-85100418c446"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 08:53:32 crc kubenswrapper[4813]: I1201 08:53:32.926196 4813 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/348e5b65-5859-4870-aaa1-85100418c446-util\") on node \"crc\" DevicePath \"\"" Dec 01 08:53:33 crc kubenswrapper[4813]: I1201 08:53:33.104382 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534fxznfm" event={"ID":"348e5b65-5859-4870-aaa1-85100418c446","Type":"ContainerDied","Data":"d256c046466a493e42913f1f5e3ff9dba71296e78f801ee1ba3d7f51edc4bbdf"} Dec 01 08:53:33 crc kubenswrapper[4813]: I1201 08:53:33.104674 4813 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d256c046466a493e42913f1f5e3ff9dba71296e78f801ee1ba3d7f51edc4bbdf" Dec 01 08:53:33 crc kubenswrapper[4813]: I1201 08:53:33.105217 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534fxznfm" Dec 01 08:53:41 crc kubenswrapper[4813]: I1201 08:53:41.318041 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-679c7f6ffd-mfx9v"] Dec 01 08:53:41 crc kubenswrapper[4813]: E1201 08:53:41.319270 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="348e5b65-5859-4870-aaa1-85100418c446" containerName="util" Dec 01 08:53:41 crc kubenswrapper[4813]: I1201 08:53:41.319326 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="348e5b65-5859-4870-aaa1-85100418c446" containerName="util" Dec 01 08:53:41 crc kubenswrapper[4813]: E1201 08:53:41.319399 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="348e5b65-5859-4870-aaa1-85100418c446" containerName="pull" Dec 01 08:53:41 crc kubenswrapper[4813]: I1201 08:53:41.319411 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="348e5b65-5859-4870-aaa1-85100418c446" containerName="pull" Dec 01 08:53:41 crc kubenswrapper[4813]: E1201 08:53:41.319432 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="348e5b65-5859-4870-aaa1-85100418c446" containerName="extract" Dec 01 08:53:41 crc kubenswrapper[4813]: I1201 08:53:41.319448 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="348e5b65-5859-4870-aaa1-85100418c446" containerName="extract" Dec 01 08:53:41 crc kubenswrapper[4813]: I1201 08:53:41.319751 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="348e5b65-5859-4870-aaa1-85100418c446" containerName="extract" Dec 01 08:53:41 crc kubenswrapper[4813]: I1201 08:53:41.320829 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-679c7f6ffd-mfx9v" Dec 01 08:53:41 crc kubenswrapper[4813]: I1201 08:53:41.323704 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-dockercfg-z4lxl" Dec 01 08:53:41 crc kubenswrapper[4813]: I1201 08:53:41.323808 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"webhook-server-cert" Dec 01 08:53:41 crc kubenswrapper[4813]: I1201 08:53:41.325226 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-service-cert" Dec 01 08:53:41 crc kubenswrapper[4813]: I1201 08:53:41.325619 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-679c7f6ffd-mfx9v"] Dec 01 08:53:41 crc kubenswrapper[4813]: I1201 08:53:41.474415 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hsfrr\" (UniqueName: \"kubernetes.io/projected/84350e15-9af9-4de1-bf0b-9e75c91b180a-kube-api-access-hsfrr\") pod \"mariadb-operator-controller-manager-679c7f6ffd-mfx9v\" (UID: \"84350e15-9af9-4de1-bf0b-9e75c91b180a\") " pod="openstack-operators/mariadb-operator-controller-manager-679c7f6ffd-mfx9v" Dec 01 08:53:41 crc kubenswrapper[4813]: I1201 08:53:41.474495 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/84350e15-9af9-4de1-bf0b-9e75c91b180a-apiservice-cert\") pod \"mariadb-operator-controller-manager-679c7f6ffd-mfx9v\" (UID: \"84350e15-9af9-4de1-bf0b-9e75c91b180a\") " pod="openstack-operators/mariadb-operator-controller-manager-679c7f6ffd-mfx9v" Dec 01 08:53:41 crc kubenswrapper[4813]: I1201 08:53:41.474526 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/84350e15-9af9-4de1-bf0b-9e75c91b180a-webhook-cert\") pod \"mariadb-operator-controller-manager-679c7f6ffd-mfx9v\" (UID: \"84350e15-9af9-4de1-bf0b-9e75c91b180a\") " pod="openstack-operators/mariadb-operator-controller-manager-679c7f6ffd-mfx9v" Dec 01 08:53:41 crc kubenswrapper[4813]: I1201 08:53:41.575323 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hsfrr\" (UniqueName: \"kubernetes.io/projected/84350e15-9af9-4de1-bf0b-9e75c91b180a-kube-api-access-hsfrr\") pod \"mariadb-operator-controller-manager-679c7f6ffd-mfx9v\" (UID: \"84350e15-9af9-4de1-bf0b-9e75c91b180a\") " pod="openstack-operators/mariadb-operator-controller-manager-679c7f6ffd-mfx9v" Dec 01 08:53:41 crc kubenswrapper[4813]: I1201 08:53:41.575395 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/84350e15-9af9-4de1-bf0b-9e75c91b180a-apiservice-cert\") pod \"mariadb-operator-controller-manager-679c7f6ffd-mfx9v\" (UID: \"84350e15-9af9-4de1-bf0b-9e75c91b180a\") " pod="openstack-operators/mariadb-operator-controller-manager-679c7f6ffd-mfx9v" Dec 01 08:53:41 crc kubenswrapper[4813]: I1201 08:53:41.575420 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/84350e15-9af9-4de1-bf0b-9e75c91b180a-webhook-cert\") pod \"mariadb-operator-controller-manager-679c7f6ffd-mfx9v\" (UID: \"84350e15-9af9-4de1-bf0b-9e75c91b180a\") " pod="openstack-operators/mariadb-operator-controller-manager-679c7f6ffd-mfx9v" Dec 01 08:53:41 crc kubenswrapper[4813]: I1201 08:53:41.581075 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/84350e15-9af9-4de1-bf0b-9e75c91b180a-webhook-cert\") pod \"mariadb-operator-controller-manager-679c7f6ffd-mfx9v\" (UID: \"84350e15-9af9-4de1-bf0b-9e75c91b180a\") " pod="openstack-operators/mariadb-operator-controller-manager-679c7f6ffd-mfx9v" Dec 01 08:53:41 crc kubenswrapper[4813]: I1201 08:53:41.587429 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/84350e15-9af9-4de1-bf0b-9e75c91b180a-apiservice-cert\") pod \"mariadb-operator-controller-manager-679c7f6ffd-mfx9v\" (UID: \"84350e15-9af9-4de1-bf0b-9e75c91b180a\") " pod="openstack-operators/mariadb-operator-controller-manager-679c7f6ffd-mfx9v" Dec 01 08:53:41 crc kubenswrapper[4813]: I1201 08:53:41.592280 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hsfrr\" (UniqueName: \"kubernetes.io/projected/84350e15-9af9-4de1-bf0b-9e75c91b180a-kube-api-access-hsfrr\") pod \"mariadb-operator-controller-manager-679c7f6ffd-mfx9v\" (UID: \"84350e15-9af9-4de1-bf0b-9e75c91b180a\") " pod="openstack-operators/mariadb-operator-controller-manager-679c7f6ffd-mfx9v" Dec 01 08:53:41 crc kubenswrapper[4813]: I1201 08:53:41.643668 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-679c7f6ffd-mfx9v" Dec 01 08:53:41 crc kubenswrapper[4813]: I1201 08:53:41.855644 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-679c7f6ffd-mfx9v"] Dec 01 08:53:42 crc kubenswrapper[4813]: I1201 08:53:42.185496 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-679c7f6ffd-mfx9v" event={"ID":"84350e15-9af9-4de1-bf0b-9e75c91b180a","Type":"ContainerStarted","Data":"a7271f59525fb78749cb368b151efb83eb5516aefb5053e5b02244a6bad89245"} Dec 01 08:53:46 crc kubenswrapper[4813]: I1201 08:53:46.213159 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-679c7f6ffd-mfx9v" event={"ID":"84350e15-9af9-4de1-bf0b-9e75c91b180a","Type":"ContainerStarted","Data":"4cbeee28f90ffdf966d5d0157b2f7373e15ff1820f16530504a35e976c1128d2"} Dec 01 08:53:46 crc kubenswrapper[4813]: I1201 08:53:46.213809 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-679c7f6ffd-mfx9v" Dec 01 08:53:46 crc kubenswrapper[4813]: I1201 08:53:46.297880 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-controller-manager-679c7f6ffd-mfx9v" podStartSLOduration=1.7811073309999998 podStartE2EDuration="5.297845293s" podCreationTimestamp="2025-12-01 08:53:41 +0000 UTC" firstStartedPulling="2025-12-01 08:53:41.866372358 +0000 UTC m=+806.489293944" lastFinishedPulling="2025-12-01 08:53:45.38311028 +0000 UTC m=+810.006031906" observedRunningTime="2025-12-01 08:53:46.294300295 +0000 UTC m=+810.917221881" watchObservedRunningTime="2025-12-01 08:53:46.297845293 +0000 UTC m=+810.920766879" Dec 01 08:53:51 crc kubenswrapper[4813]: I1201 08:53:51.652128 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-679c7f6ffd-mfx9v" Dec 01 08:53:53 crc kubenswrapper[4813]: I1201 08:53:53.369104 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-index-pvsjq"] Dec 01 08:53:53 crc kubenswrapper[4813]: I1201 08:53:53.369775 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-index-pvsjq" Dec 01 08:53:53 crc kubenswrapper[4813]: I1201 08:53:53.376719 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-index-dockercfg-cps9c" Dec 01 08:53:53 crc kubenswrapper[4813]: I1201 08:53:53.384899 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-index-pvsjq"] Dec 01 08:53:53 crc kubenswrapper[4813]: I1201 08:53:53.391930 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dr5wc\" (UniqueName: \"kubernetes.io/projected/b95142c9-199e-4ec0-ae62-93456918c4be-kube-api-access-dr5wc\") pod \"infra-operator-index-pvsjq\" (UID: \"b95142c9-199e-4ec0-ae62-93456918c4be\") " pod="openstack-operators/infra-operator-index-pvsjq" Dec 01 08:53:53 crc kubenswrapper[4813]: I1201 08:53:53.493145 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dr5wc\" (UniqueName: \"kubernetes.io/projected/b95142c9-199e-4ec0-ae62-93456918c4be-kube-api-access-dr5wc\") pod \"infra-operator-index-pvsjq\" (UID: \"b95142c9-199e-4ec0-ae62-93456918c4be\") " pod="openstack-operators/infra-operator-index-pvsjq" Dec 01 08:53:53 crc kubenswrapper[4813]: I1201 08:53:53.512144 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dr5wc\" (UniqueName: \"kubernetes.io/projected/b95142c9-199e-4ec0-ae62-93456918c4be-kube-api-access-dr5wc\") pod \"infra-operator-index-pvsjq\" (UID: \"b95142c9-199e-4ec0-ae62-93456918c4be\") " pod="openstack-operators/infra-operator-index-pvsjq" Dec 01 08:53:53 crc kubenswrapper[4813]: I1201 08:53:53.683910 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-index-pvsjq" Dec 01 08:53:54 crc kubenswrapper[4813]: I1201 08:53:54.150026 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-index-pvsjq"] Dec 01 08:53:54 crc kubenswrapper[4813]: W1201 08:53:54.160435 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb95142c9_199e_4ec0_ae62_93456918c4be.slice/crio-2c89725af9641bbe7b6946297b6f05e0fb8199d89f5e62926a14077e69e23ccc WatchSource:0}: Error finding container 2c89725af9641bbe7b6946297b6f05e0fb8199d89f5e62926a14077e69e23ccc: Status 404 returned error can't find the container with id 2c89725af9641bbe7b6946297b6f05e0fb8199d89f5e62926a14077e69e23ccc Dec 01 08:53:54 crc kubenswrapper[4813]: I1201 08:53:54.280313 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-index-pvsjq" event={"ID":"b95142c9-199e-4ec0-ae62-93456918c4be","Type":"ContainerStarted","Data":"2c89725af9641bbe7b6946297b6f05e0fb8199d89f5e62926a14077e69e23ccc"} Dec 01 08:53:55 crc kubenswrapper[4813]: I1201 08:53:55.298935 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-index-pvsjq" event={"ID":"b95142c9-199e-4ec0-ae62-93456918c4be","Type":"ContainerStarted","Data":"3e68c8273c05eb1a108fb6163c2cb5749fb2302fc583c191c542ec84b9cf231f"} Dec 01 08:53:55 crc kubenswrapper[4813]: I1201 08:53:55.319128 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-index-pvsjq" podStartSLOduration=1.381225087 podStartE2EDuration="2.319113098s" podCreationTimestamp="2025-12-01 08:53:53 +0000 UTC" firstStartedPulling="2025-12-01 08:53:54.163056031 +0000 UTC m=+818.785977627" lastFinishedPulling="2025-12-01 08:53:55.100944012 +0000 UTC m=+819.723865638" observedRunningTime="2025-12-01 08:53:55.317703819 +0000 UTC m=+819.940625405" watchObservedRunningTime="2025-12-01 08:53:55.319113098 +0000 UTC m=+819.942034684" Dec 01 08:53:56 crc kubenswrapper[4813]: I1201 08:53:56.565785 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/infra-operator-index-pvsjq"] Dec 01 08:53:57 crc kubenswrapper[4813]: I1201 08:53:57.173942 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-index-8b5ct"] Dec 01 08:53:57 crc kubenswrapper[4813]: I1201 08:53:57.175592 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-index-8b5ct" Dec 01 08:53:57 crc kubenswrapper[4813]: I1201 08:53:57.192527 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-index-8b5ct"] Dec 01 08:53:57 crc kubenswrapper[4813]: I1201 08:53:57.315378 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/infra-operator-index-pvsjq" podUID="b95142c9-199e-4ec0-ae62-93456918c4be" containerName="registry-server" containerID="cri-o://3e68c8273c05eb1a108fb6163c2cb5749fb2302fc583c191c542ec84b9cf231f" gracePeriod=2 Dec 01 08:53:57 crc kubenswrapper[4813]: I1201 08:53:57.352374 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m92t7\" (UniqueName: \"kubernetes.io/projected/35db6389-d26f-45f7-919a-734a16107934-kube-api-access-m92t7\") pod \"infra-operator-index-8b5ct\" (UID: \"35db6389-d26f-45f7-919a-734a16107934\") " pod="openstack-operators/infra-operator-index-8b5ct" Dec 01 08:53:57 crc kubenswrapper[4813]: I1201 08:53:57.453153 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m92t7\" (UniqueName: \"kubernetes.io/projected/35db6389-d26f-45f7-919a-734a16107934-kube-api-access-m92t7\") pod \"infra-operator-index-8b5ct\" (UID: \"35db6389-d26f-45f7-919a-734a16107934\") " pod="openstack-operators/infra-operator-index-8b5ct" Dec 01 08:53:57 crc kubenswrapper[4813]: I1201 08:53:57.478591 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m92t7\" (UniqueName: \"kubernetes.io/projected/35db6389-d26f-45f7-919a-734a16107934-kube-api-access-m92t7\") pod \"infra-operator-index-8b5ct\" (UID: \"35db6389-d26f-45f7-919a-734a16107934\") " pod="openstack-operators/infra-operator-index-8b5ct" Dec 01 08:53:57 crc kubenswrapper[4813]: I1201 08:53:57.512294 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-index-8b5ct" Dec 01 08:53:57 crc kubenswrapper[4813]: I1201 08:53:57.730394 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-index-pvsjq" Dec 01 08:53:57 crc kubenswrapper[4813]: I1201 08:53:57.748518 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-index-8b5ct"] Dec 01 08:53:57 crc kubenswrapper[4813]: W1201 08:53:57.765979 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod35db6389_d26f_45f7_919a_734a16107934.slice/crio-5e8a1fb5f1aa201194f382989c965d75e3eb21fece89babc4e0a7fcf5f58d2fd WatchSource:0}: Error finding container 5e8a1fb5f1aa201194f382989c965d75e3eb21fece89babc4e0a7fcf5f58d2fd: Status 404 returned error can't find the container with id 5e8a1fb5f1aa201194f382989c965d75e3eb21fece89babc4e0a7fcf5f58d2fd Dec 01 08:53:57 crc kubenswrapper[4813]: I1201 08:53:57.857363 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dr5wc\" (UniqueName: \"kubernetes.io/projected/b95142c9-199e-4ec0-ae62-93456918c4be-kube-api-access-dr5wc\") pod \"b95142c9-199e-4ec0-ae62-93456918c4be\" (UID: \"b95142c9-199e-4ec0-ae62-93456918c4be\") " Dec 01 08:53:57 crc kubenswrapper[4813]: I1201 08:53:57.861904 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b95142c9-199e-4ec0-ae62-93456918c4be-kube-api-access-dr5wc" (OuterVolumeSpecName: "kube-api-access-dr5wc") pod "b95142c9-199e-4ec0-ae62-93456918c4be" (UID: "b95142c9-199e-4ec0-ae62-93456918c4be"). InnerVolumeSpecName "kube-api-access-dr5wc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:53:57 crc kubenswrapper[4813]: I1201 08:53:57.958735 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dr5wc\" (UniqueName: \"kubernetes.io/projected/b95142c9-199e-4ec0-ae62-93456918c4be-kube-api-access-dr5wc\") on node \"crc\" DevicePath \"\"" Dec 01 08:53:58 crc kubenswrapper[4813]: I1201 08:53:58.322705 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-index-8b5ct" event={"ID":"35db6389-d26f-45f7-919a-734a16107934","Type":"ContainerStarted","Data":"5e8a1fb5f1aa201194f382989c965d75e3eb21fece89babc4e0a7fcf5f58d2fd"} Dec 01 08:53:58 crc kubenswrapper[4813]: I1201 08:53:58.324358 4813 generic.go:334] "Generic (PLEG): container finished" podID="b95142c9-199e-4ec0-ae62-93456918c4be" containerID="3e68c8273c05eb1a108fb6163c2cb5749fb2302fc583c191c542ec84b9cf231f" exitCode=0 Dec 01 08:53:58 crc kubenswrapper[4813]: I1201 08:53:58.324411 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-index-pvsjq" event={"ID":"b95142c9-199e-4ec0-ae62-93456918c4be","Type":"ContainerDied","Data":"3e68c8273c05eb1a108fb6163c2cb5749fb2302fc583c191c542ec84b9cf231f"} Dec 01 08:53:58 crc kubenswrapper[4813]: I1201 08:53:58.324452 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-index-pvsjq" event={"ID":"b95142c9-199e-4ec0-ae62-93456918c4be","Type":"ContainerDied","Data":"2c89725af9641bbe7b6946297b6f05e0fb8199d89f5e62926a14077e69e23ccc"} Dec 01 08:53:58 crc kubenswrapper[4813]: I1201 08:53:58.324467 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-index-pvsjq" Dec 01 08:53:58 crc kubenswrapper[4813]: I1201 08:53:58.324490 4813 scope.go:117] "RemoveContainer" containerID="3e68c8273c05eb1a108fb6163c2cb5749fb2302fc583c191c542ec84b9cf231f" Dec 01 08:53:58 crc kubenswrapper[4813]: I1201 08:53:58.347516 4813 scope.go:117] "RemoveContainer" containerID="3e68c8273c05eb1a108fb6163c2cb5749fb2302fc583c191c542ec84b9cf231f" Dec 01 08:53:58 crc kubenswrapper[4813]: E1201 08:53:58.348298 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3e68c8273c05eb1a108fb6163c2cb5749fb2302fc583c191c542ec84b9cf231f\": container with ID starting with 3e68c8273c05eb1a108fb6163c2cb5749fb2302fc583c191c542ec84b9cf231f not found: ID does not exist" containerID="3e68c8273c05eb1a108fb6163c2cb5749fb2302fc583c191c542ec84b9cf231f" Dec 01 08:53:58 crc kubenswrapper[4813]: I1201 08:53:58.348686 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3e68c8273c05eb1a108fb6163c2cb5749fb2302fc583c191c542ec84b9cf231f"} err="failed to get container status \"3e68c8273c05eb1a108fb6163c2cb5749fb2302fc583c191c542ec84b9cf231f\": rpc error: code = NotFound desc = could not find container \"3e68c8273c05eb1a108fb6163c2cb5749fb2302fc583c191c542ec84b9cf231f\": container with ID starting with 3e68c8273c05eb1a108fb6163c2cb5749fb2302fc583c191c542ec84b9cf231f not found: ID does not exist" Dec 01 08:53:58 crc kubenswrapper[4813]: I1201 08:53:58.363176 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/infra-operator-index-pvsjq"] Dec 01 08:53:58 crc kubenswrapper[4813]: I1201 08:53:58.371008 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/infra-operator-index-pvsjq"] Dec 01 08:53:58 crc kubenswrapper[4813]: I1201 08:53:58.401278 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b95142c9-199e-4ec0-ae62-93456918c4be" path="/var/lib/kubelet/pods/b95142c9-199e-4ec0-ae62-93456918c4be/volumes" Dec 01 08:53:59 crc kubenswrapper[4813]: I1201 08:53:59.340164 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-index-8b5ct" event={"ID":"35db6389-d26f-45f7-919a-734a16107934","Type":"ContainerStarted","Data":"3750d324ce7a125845adb78d7ce4957cc4f47a0c9eec5a9e81edba3398829947"} Dec 01 08:53:59 crc kubenswrapper[4813]: I1201 08:53:59.365419 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-index-8b5ct" podStartSLOduration=1.731305887 podStartE2EDuration="2.365390361s" podCreationTimestamp="2025-12-01 08:53:57 +0000 UTC" firstStartedPulling="2025-12-01 08:53:57.769568148 +0000 UTC m=+822.392489734" lastFinishedPulling="2025-12-01 08:53:58.403652622 +0000 UTC m=+823.026574208" observedRunningTime="2025-12-01 08:53:59.362281976 +0000 UTC m=+823.985203602" watchObservedRunningTime="2025-12-01 08:53:59.365390361 +0000 UTC m=+823.988311977" Dec 01 08:54:07 crc kubenswrapper[4813]: I1201 08:54:07.514265 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/infra-operator-index-8b5ct" Dec 01 08:54:07 crc kubenswrapper[4813]: I1201 08:54:07.514846 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-index-8b5ct" Dec 01 08:54:07 crc kubenswrapper[4813]: I1201 08:54:07.545058 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/infra-operator-index-8b5ct" Dec 01 08:54:08 crc kubenswrapper[4813]: I1201 08:54:08.465230 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-index-8b5ct" Dec 01 08:54:14 crc kubenswrapper[4813]: I1201 08:54:14.884181 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5d4rn6x"] Dec 01 08:54:14 crc kubenswrapper[4813]: E1201 08:54:14.886587 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b95142c9-199e-4ec0-ae62-93456918c4be" containerName="registry-server" Dec 01 08:54:14 crc kubenswrapper[4813]: I1201 08:54:14.886773 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="b95142c9-199e-4ec0-ae62-93456918c4be" containerName="registry-server" Dec 01 08:54:14 crc kubenswrapper[4813]: I1201 08:54:14.886947 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="b95142c9-199e-4ec0-ae62-93456918c4be" containerName="registry-server" Dec 01 08:54:14 crc kubenswrapper[4813]: I1201 08:54:14.888340 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5d4rn6x" Dec 01 08:54:14 crc kubenswrapper[4813]: I1201 08:54:14.890099 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5d4rn6x"] Dec 01 08:54:14 crc kubenswrapper[4813]: I1201 08:54:14.891453 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-fml9r" Dec 01 08:54:14 crc kubenswrapper[4813]: I1201 08:54:14.935494 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/db90917b-21c8-43d8-99c5-eb77590da02f-util\") pod \"5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5d4rn6x\" (UID: \"db90917b-21c8-43d8-99c5-eb77590da02f\") " pod="openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5d4rn6x" Dec 01 08:54:14 crc kubenswrapper[4813]: I1201 08:54:14.935578 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4vjtf\" (UniqueName: \"kubernetes.io/projected/db90917b-21c8-43d8-99c5-eb77590da02f-kube-api-access-4vjtf\") pod \"5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5d4rn6x\" (UID: \"db90917b-21c8-43d8-99c5-eb77590da02f\") " pod="openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5d4rn6x" Dec 01 08:54:14 crc kubenswrapper[4813]: I1201 08:54:14.935827 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/db90917b-21c8-43d8-99c5-eb77590da02f-bundle\") pod \"5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5d4rn6x\" (UID: \"db90917b-21c8-43d8-99c5-eb77590da02f\") " pod="openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5d4rn6x" Dec 01 08:54:15 crc kubenswrapper[4813]: I1201 08:54:15.037218 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/db90917b-21c8-43d8-99c5-eb77590da02f-bundle\") pod \"5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5d4rn6x\" (UID: \"db90917b-21c8-43d8-99c5-eb77590da02f\") " pod="openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5d4rn6x" Dec 01 08:54:15 crc kubenswrapper[4813]: I1201 08:54:15.037346 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/db90917b-21c8-43d8-99c5-eb77590da02f-util\") pod \"5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5d4rn6x\" (UID: \"db90917b-21c8-43d8-99c5-eb77590da02f\") " pod="openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5d4rn6x" Dec 01 08:54:15 crc kubenswrapper[4813]: I1201 08:54:15.037374 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4vjtf\" (UniqueName: \"kubernetes.io/projected/db90917b-21c8-43d8-99c5-eb77590da02f-kube-api-access-4vjtf\") pod \"5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5d4rn6x\" (UID: \"db90917b-21c8-43d8-99c5-eb77590da02f\") " pod="openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5d4rn6x" Dec 01 08:54:15 crc kubenswrapper[4813]: I1201 08:54:15.037932 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/db90917b-21c8-43d8-99c5-eb77590da02f-bundle\") pod \"5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5d4rn6x\" (UID: \"db90917b-21c8-43d8-99c5-eb77590da02f\") " pod="openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5d4rn6x" Dec 01 08:54:15 crc kubenswrapper[4813]: I1201 08:54:15.037956 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/db90917b-21c8-43d8-99c5-eb77590da02f-util\") pod \"5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5d4rn6x\" (UID: \"db90917b-21c8-43d8-99c5-eb77590da02f\") " pod="openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5d4rn6x" Dec 01 08:54:15 crc kubenswrapper[4813]: I1201 08:54:15.076371 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4vjtf\" (UniqueName: \"kubernetes.io/projected/db90917b-21c8-43d8-99c5-eb77590da02f-kube-api-access-4vjtf\") pod \"5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5d4rn6x\" (UID: \"db90917b-21c8-43d8-99c5-eb77590da02f\") " pod="openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5d4rn6x" Dec 01 08:54:15 crc kubenswrapper[4813]: I1201 08:54:15.219401 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5d4rn6x" Dec 01 08:54:15 crc kubenswrapper[4813]: I1201 08:54:15.461444 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5d4rn6x"] Dec 01 08:54:16 crc kubenswrapper[4813]: I1201 08:54:16.488096 4813 generic.go:334] "Generic (PLEG): container finished" podID="db90917b-21c8-43d8-99c5-eb77590da02f" containerID="cc8ed80f25c94a3910eb49f3e867bbe55dc9a7d59f8c1d3d1c1505e51538c778" exitCode=0 Dec 01 08:54:16 crc kubenswrapper[4813]: I1201 08:54:16.488224 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5d4rn6x" event={"ID":"db90917b-21c8-43d8-99c5-eb77590da02f","Type":"ContainerDied","Data":"cc8ed80f25c94a3910eb49f3e867bbe55dc9a7d59f8c1d3d1c1505e51538c778"} Dec 01 08:54:16 crc kubenswrapper[4813]: I1201 08:54:16.488553 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5d4rn6x" event={"ID":"db90917b-21c8-43d8-99c5-eb77590da02f","Type":"ContainerStarted","Data":"b34c8829c37003c504f2f0c7dc3759db65b7d36245ee2a4d02509d03559b1e6e"} Dec 01 08:54:17 crc kubenswrapper[4813]: I1201 08:54:17.497615 4813 generic.go:334] "Generic (PLEG): container finished" podID="db90917b-21c8-43d8-99c5-eb77590da02f" containerID="c05742026abe1865b48f1118ca08065c5b800894aeeaacc582124c7505e4a7ef" exitCode=0 Dec 01 08:54:17 crc kubenswrapper[4813]: I1201 08:54:17.497668 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5d4rn6x" event={"ID":"db90917b-21c8-43d8-99c5-eb77590da02f","Type":"ContainerDied","Data":"c05742026abe1865b48f1118ca08065c5b800894aeeaacc582124c7505e4a7ef"} Dec 01 08:54:18 crc kubenswrapper[4813]: I1201 08:54:18.510401 4813 generic.go:334] "Generic (PLEG): container finished" podID="db90917b-21c8-43d8-99c5-eb77590da02f" containerID="2776f225b4fec97833b5b31c52c31da4f313fc6edf4b20c00fc9f06c1e82ad5b" exitCode=0 Dec 01 08:54:18 crc kubenswrapper[4813]: I1201 08:54:18.510504 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5d4rn6x" event={"ID":"db90917b-21c8-43d8-99c5-eb77590da02f","Type":"ContainerDied","Data":"2776f225b4fec97833b5b31c52c31da4f313fc6edf4b20c00fc9f06c1e82ad5b"} Dec 01 08:54:19 crc kubenswrapper[4813]: I1201 08:54:19.882281 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5d4rn6x" Dec 01 08:54:20 crc kubenswrapper[4813]: I1201 08:54:20.029126 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4vjtf\" (UniqueName: \"kubernetes.io/projected/db90917b-21c8-43d8-99c5-eb77590da02f-kube-api-access-4vjtf\") pod \"db90917b-21c8-43d8-99c5-eb77590da02f\" (UID: \"db90917b-21c8-43d8-99c5-eb77590da02f\") " Dec 01 08:54:20 crc kubenswrapper[4813]: I1201 08:54:20.029244 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/db90917b-21c8-43d8-99c5-eb77590da02f-util\") pod \"db90917b-21c8-43d8-99c5-eb77590da02f\" (UID: \"db90917b-21c8-43d8-99c5-eb77590da02f\") " Dec 01 08:54:20 crc kubenswrapper[4813]: I1201 08:54:20.029357 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/db90917b-21c8-43d8-99c5-eb77590da02f-bundle\") pod \"db90917b-21c8-43d8-99c5-eb77590da02f\" (UID: \"db90917b-21c8-43d8-99c5-eb77590da02f\") " Dec 01 08:54:20 crc kubenswrapper[4813]: I1201 08:54:20.030546 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/db90917b-21c8-43d8-99c5-eb77590da02f-bundle" (OuterVolumeSpecName: "bundle") pod "db90917b-21c8-43d8-99c5-eb77590da02f" (UID: "db90917b-21c8-43d8-99c5-eb77590da02f"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 08:54:20 crc kubenswrapper[4813]: I1201 08:54:20.033847 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/db90917b-21c8-43d8-99c5-eb77590da02f-kube-api-access-4vjtf" (OuterVolumeSpecName: "kube-api-access-4vjtf") pod "db90917b-21c8-43d8-99c5-eb77590da02f" (UID: "db90917b-21c8-43d8-99c5-eb77590da02f"). InnerVolumeSpecName "kube-api-access-4vjtf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:54:20 crc kubenswrapper[4813]: I1201 08:54:20.058769 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/db90917b-21c8-43d8-99c5-eb77590da02f-util" (OuterVolumeSpecName: "util") pod "db90917b-21c8-43d8-99c5-eb77590da02f" (UID: "db90917b-21c8-43d8-99c5-eb77590da02f"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 08:54:20 crc kubenswrapper[4813]: I1201 08:54:20.130106 4813 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/db90917b-21c8-43d8-99c5-eb77590da02f-util\") on node \"crc\" DevicePath \"\"" Dec 01 08:54:20 crc kubenswrapper[4813]: I1201 08:54:20.130143 4813 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/db90917b-21c8-43d8-99c5-eb77590da02f-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 08:54:20 crc kubenswrapper[4813]: I1201 08:54:20.130152 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4vjtf\" (UniqueName: \"kubernetes.io/projected/db90917b-21c8-43d8-99c5-eb77590da02f-kube-api-access-4vjtf\") on node \"crc\" DevicePath \"\"" Dec 01 08:54:20 crc kubenswrapper[4813]: I1201 08:54:20.526724 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5d4rn6x" event={"ID":"db90917b-21c8-43d8-99c5-eb77590da02f","Type":"ContainerDied","Data":"b34c8829c37003c504f2f0c7dc3759db65b7d36245ee2a4d02509d03559b1e6e"} Dec 01 08:54:20 crc kubenswrapper[4813]: I1201 08:54:20.526788 4813 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b34c8829c37003c504f2f0c7dc3759db65b7d36245ee2a4d02509d03559b1e6e" Dec 01 08:54:20 crc kubenswrapper[4813]: I1201 08:54:20.526828 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5d4rn6x" Dec 01 08:54:24 crc kubenswrapper[4813]: I1201 08:54:24.156935 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-controller-manager-84d8d7c7c5-bcbsv"] Dec 01 08:54:24 crc kubenswrapper[4813]: E1201 08:54:24.158131 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="db90917b-21c8-43d8-99c5-eb77590da02f" containerName="pull" Dec 01 08:54:24 crc kubenswrapper[4813]: I1201 08:54:24.158163 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="db90917b-21c8-43d8-99c5-eb77590da02f" containerName="pull" Dec 01 08:54:24 crc kubenswrapper[4813]: E1201 08:54:24.158185 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="db90917b-21c8-43d8-99c5-eb77590da02f" containerName="extract" Dec 01 08:54:24 crc kubenswrapper[4813]: I1201 08:54:24.158194 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="db90917b-21c8-43d8-99c5-eb77590da02f" containerName="extract" Dec 01 08:54:24 crc kubenswrapper[4813]: E1201 08:54:24.158213 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="db90917b-21c8-43d8-99c5-eb77590da02f" containerName="util" Dec 01 08:54:24 crc kubenswrapper[4813]: I1201 08:54:24.158221 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="db90917b-21c8-43d8-99c5-eb77590da02f" containerName="util" Dec 01 08:54:24 crc kubenswrapper[4813]: I1201 08:54:24.158406 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="db90917b-21c8-43d8-99c5-eb77590da02f" containerName="extract" Dec 01 08:54:24 crc kubenswrapper[4813]: I1201 08:54:24.159472 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-84d8d7c7c5-bcbsv" Dec 01 08:54:24 crc kubenswrapper[4813]: I1201 08:54:24.161880 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-service-cert" Dec 01 08:54:24 crc kubenswrapper[4813]: I1201 08:54:24.161879 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-pcpl8" Dec 01 08:54:24 crc kubenswrapper[4813]: I1201 08:54:24.202461 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-84d8d7c7c5-bcbsv"] Dec 01 08:54:24 crc kubenswrapper[4813]: I1201 08:54:24.331281 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/42e1cee7-cd3a-4194-af37-e29d5f04389a-apiservice-cert\") pod \"infra-operator-controller-manager-84d8d7c7c5-bcbsv\" (UID: \"42e1cee7-cd3a-4194-af37-e29d5f04389a\") " pod="openstack-operators/infra-operator-controller-manager-84d8d7c7c5-bcbsv" Dec 01 08:54:24 crc kubenswrapper[4813]: I1201 08:54:24.331373 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-csgbg\" (UniqueName: \"kubernetes.io/projected/42e1cee7-cd3a-4194-af37-e29d5f04389a-kube-api-access-csgbg\") pod \"infra-operator-controller-manager-84d8d7c7c5-bcbsv\" (UID: \"42e1cee7-cd3a-4194-af37-e29d5f04389a\") " pod="openstack-operators/infra-operator-controller-manager-84d8d7c7c5-bcbsv" Dec 01 08:54:24 crc kubenswrapper[4813]: I1201 08:54:24.331400 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/42e1cee7-cd3a-4194-af37-e29d5f04389a-webhook-cert\") pod \"infra-operator-controller-manager-84d8d7c7c5-bcbsv\" (UID: \"42e1cee7-cd3a-4194-af37-e29d5f04389a\") " pod="openstack-operators/infra-operator-controller-manager-84d8d7c7c5-bcbsv" Dec 01 08:54:24 crc kubenswrapper[4813]: I1201 08:54:24.459793 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-csgbg\" (UniqueName: \"kubernetes.io/projected/42e1cee7-cd3a-4194-af37-e29d5f04389a-kube-api-access-csgbg\") pod \"infra-operator-controller-manager-84d8d7c7c5-bcbsv\" (UID: \"42e1cee7-cd3a-4194-af37-e29d5f04389a\") " pod="openstack-operators/infra-operator-controller-manager-84d8d7c7c5-bcbsv" Dec 01 08:54:24 crc kubenswrapper[4813]: I1201 08:54:24.459842 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/42e1cee7-cd3a-4194-af37-e29d5f04389a-webhook-cert\") pod \"infra-operator-controller-manager-84d8d7c7c5-bcbsv\" (UID: \"42e1cee7-cd3a-4194-af37-e29d5f04389a\") " pod="openstack-operators/infra-operator-controller-manager-84d8d7c7c5-bcbsv" Dec 01 08:54:24 crc kubenswrapper[4813]: I1201 08:54:24.459905 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/42e1cee7-cd3a-4194-af37-e29d5f04389a-apiservice-cert\") pod \"infra-operator-controller-manager-84d8d7c7c5-bcbsv\" (UID: \"42e1cee7-cd3a-4194-af37-e29d5f04389a\") " pod="openstack-operators/infra-operator-controller-manager-84d8d7c7c5-bcbsv" Dec 01 08:54:24 crc kubenswrapper[4813]: I1201 08:54:24.466814 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/42e1cee7-cd3a-4194-af37-e29d5f04389a-apiservice-cert\") pod \"infra-operator-controller-manager-84d8d7c7c5-bcbsv\" (UID: \"42e1cee7-cd3a-4194-af37-e29d5f04389a\") " pod="openstack-operators/infra-operator-controller-manager-84d8d7c7c5-bcbsv" Dec 01 08:54:24 crc kubenswrapper[4813]: I1201 08:54:24.480067 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-csgbg\" (UniqueName: \"kubernetes.io/projected/42e1cee7-cd3a-4194-af37-e29d5f04389a-kube-api-access-csgbg\") pod \"infra-operator-controller-manager-84d8d7c7c5-bcbsv\" (UID: \"42e1cee7-cd3a-4194-af37-e29d5f04389a\") " pod="openstack-operators/infra-operator-controller-manager-84d8d7c7c5-bcbsv" Dec 01 08:54:24 crc kubenswrapper[4813]: I1201 08:54:24.482613 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/42e1cee7-cd3a-4194-af37-e29d5f04389a-webhook-cert\") pod \"infra-operator-controller-manager-84d8d7c7c5-bcbsv\" (UID: \"42e1cee7-cd3a-4194-af37-e29d5f04389a\") " pod="openstack-operators/infra-operator-controller-manager-84d8d7c7c5-bcbsv" Dec 01 08:54:24 crc kubenswrapper[4813]: I1201 08:54:24.777329 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-84d8d7c7c5-bcbsv" Dec 01 08:54:25 crc kubenswrapper[4813]: I1201 08:54:25.253638 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-84d8d7c7c5-bcbsv"] Dec 01 08:54:25 crc kubenswrapper[4813]: W1201 08:54:25.262062 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod42e1cee7_cd3a_4194_af37_e29d5f04389a.slice/crio-ac474539ad6a4f7cb7237f27e668c11f699497cde5edcd134b113b93c7dae48e WatchSource:0}: Error finding container ac474539ad6a4f7cb7237f27e668c11f699497cde5edcd134b113b93c7dae48e: Status 404 returned error can't find the container with id ac474539ad6a4f7cb7237f27e668c11f699497cde5edcd134b113b93c7dae48e Dec 01 08:54:25 crc kubenswrapper[4813]: I1201 08:54:25.568685 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-84d8d7c7c5-bcbsv" event={"ID":"42e1cee7-cd3a-4194-af37-e29d5f04389a","Type":"ContainerStarted","Data":"ac474539ad6a4f7cb7237f27e668c11f699497cde5edcd134b113b93c7dae48e"} Dec 01 08:54:27 crc kubenswrapper[4813]: I1201 08:54:27.216200 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/openstack-galera-0"] Dec 01 08:54:27 crc kubenswrapper[4813]: I1201 08:54:27.258170 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/openstack-galera-0" Dec 01 08:54:27 crc kubenswrapper[4813]: I1201 08:54:27.267050 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/openstack-galera-1"] Dec 01 08:54:27 crc kubenswrapper[4813]: I1201 08:54:27.269460 4813 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"galera-openstack-dockercfg-gqbg9" Dec 01 08:54:27 crc kubenswrapper[4813]: I1201 08:54:27.270307 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"openstack-config-data" Dec 01 08:54:27 crc kubenswrapper[4813]: I1201 08:54:27.270401 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"openshift-service-ca.crt" Dec 01 08:54:27 crc kubenswrapper[4813]: I1201 08:54:27.271014 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"openstack-scripts" Dec 01 08:54:27 crc kubenswrapper[4813]: I1201 08:54:27.271074 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"kube-root-ca.crt" Dec 01 08:54:27 crc kubenswrapper[4813]: I1201 08:54:27.279473 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/openstack-galera-1" Dec 01 08:54:27 crc kubenswrapper[4813]: I1201 08:54:27.294497 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/openstack-galera-2"] Dec 01 08:54:27 crc kubenswrapper[4813]: I1201 08:54:27.296658 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/openstack-galera-2" Dec 01 08:54:27 crc kubenswrapper[4813]: I1201 08:54:27.302057 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/openstack-galera-0"] Dec 01 08:54:27 crc kubenswrapper[4813]: I1201 08:54:27.311513 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/openstack-galera-1"] Dec 01 08:54:27 crc kubenswrapper[4813]: I1201 08:54:27.322320 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/openstack-galera-2"] Dec 01 08:54:27 crc kubenswrapper[4813]: I1201 08:54:27.399643 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/73988763-722d-4652-8fac-053ba8217547-config-data-generated\") pod \"openstack-galera-2\" (UID: \"73988763-722d-4652-8fac-053ba8217547\") " pod="glance-kuttl-tests/openstack-galera-2" Dec 01 08:54:27 crc kubenswrapper[4813]: I1201 08:54:27.399685 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage15-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage15-crc\") pod \"openstack-galera-2\" (UID: \"73988763-722d-4652-8fac-053ba8217547\") " pod="glance-kuttl-tests/openstack-galera-2" Dec 01 08:54:27 crc kubenswrapper[4813]: I1201 08:54:27.399708 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-str7c\" (UniqueName: \"kubernetes.io/projected/73988763-722d-4652-8fac-053ba8217547-kube-api-access-str7c\") pod \"openstack-galera-2\" (UID: \"73988763-722d-4652-8fac-053ba8217547\") " pod="glance-kuttl-tests/openstack-galera-2" Dec 01 08:54:27 crc kubenswrapper[4813]: I1201 08:54:27.399726 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d2wgq\" (UniqueName: \"kubernetes.io/projected/43c6449e-e9d4-4a5b-9231-d719d2a9e488-kube-api-access-d2wgq\") pod \"openstack-galera-0\" (UID: \"43c6449e-e9d4-4a5b-9231-d719d2a9e488\") " pod="glance-kuttl-tests/openstack-galera-0" Dec 01 08:54:27 crc kubenswrapper[4813]: I1201 08:54:27.399745 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/73988763-722d-4652-8fac-053ba8217547-config-data-default\") pod \"openstack-galera-2\" (UID: \"73988763-722d-4652-8fac-053ba8217547\") " pod="glance-kuttl-tests/openstack-galera-2" Dec 01 08:54:27 crc kubenswrapper[4813]: I1201 08:54:27.399772 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/41360e5e-2620-4912-8035-0e1a9e0e715d-config-data-default\") pod \"openstack-galera-1\" (UID: \"41360e5e-2620-4912-8035-0e1a9e0e715d\") " pod="glance-kuttl-tests/openstack-galera-1" Dec 01 08:54:27 crc kubenswrapper[4813]: I1201 08:54:27.399788 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/73988763-722d-4652-8fac-053ba8217547-kolla-config\") pod \"openstack-galera-2\" (UID: \"73988763-722d-4652-8fac-053ba8217547\") " pod="glance-kuttl-tests/openstack-galera-2" Dec 01 08:54:27 crc kubenswrapper[4813]: I1201 08:54:27.399818 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/43c6449e-e9d4-4a5b-9231-d719d2a9e488-config-data-generated\") pod \"openstack-galera-0\" (UID: \"43c6449e-e9d4-4a5b-9231-d719d2a9e488\") " pod="glance-kuttl-tests/openstack-galera-0" Dec 01 08:54:27 crc kubenswrapper[4813]: I1201 08:54:27.399842 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/41360e5e-2620-4912-8035-0e1a9e0e715d-kolla-config\") pod \"openstack-galera-1\" (UID: \"41360e5e-2620-4912-8035-0e1a9e0e715d\") " pod="glance-kuttl-tests/openstack-galera-1" Dec 01 08:54:27 crc kubenswrapper[4813]: I1201 08:54:27.399858 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/41360e5e-2620-4912-8035-0e1a9e0e715d-operator-scripts\") pod \"openstack-galera-1\" (UID: \"41360e5e-2620-4912-8035-0e1a9e0e715d\") " pod="glance-kuttl-tests/openstack-galera-1" Dec 01 08:54:27 crc kubenswrapper[4813]: I1201 08:54:27.399876 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/43c6449e-e9d4-4a5b-9231-d719d2a9e488-config-data-default\") pod \"openstack-galera-0\" (UID: \"43c6449e-e9d4-4a5b-9231-d719d2a9e488\") " pod="glance-kuttl-tests/openstack-galera-0" Dec 01 08:54:27 crc kubenswrapper[4813]: I1201 08:54:27.399900 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"openstack-galera-0\" (UID: \"43c6449e-e9d4-4a5b-9231-d719d2a9e488\") " pod="glance-kuttl-tests/openstack-galera-0" Dec 01 08:54:27 crc kubenswrapper[4813]: I1201 08:54:27.399923 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/43c6449e-e9d4-4a5b-9231-d719d2a9e488-operator-scripts\") pod \"openstack-galera-0\" (UID: \"43c6449e-e9d4-4a5b-9231-d719d2a9e488\") " pod="glance-kuttl-tests/openstack-galera-0" Dec 01 08:54:27 crc kubenswrapper[4813]: I1201 08:54:27.399944 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/41360e5e-2620-4912-8035-0e1a9e0e715d-config-data-generated\") pod \"openstack-galera-1\" (UID: \"41360e5e-2620-4912-8035-0e1a9e0e715d\") " pod="glance-kuttl-tests/openstack-galera-1" Dec 01 08:54:27 crc kubenswrapper[4813]: I1201 08:54:27.399980 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"openstack-galera-1\" (UID: \"41360e5e-2620-4912-8035-0e1a9e0e715d\") " pod="glance-kuttl-tests/openstack-galera-1" Dec 01 08:54:27 crc kubenswrapper[4813]: I1201 08:54:27.399999 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/73988763-722d-4652-8fac-053ba8217547-operator-scripts\") pod \"openstack-galera-2\" (UID: \"73988763-722d-4652-8fac-053ba8217547\") " pod="glance-kuttl-tests/openstack-galera-2" Dec 01 08:54:27 crc kubenswrapper[4813]: I1201 08:54:27.400019 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ssdv7\" (UniqueName: \"kubernetes.io/projected/41360e5e-2620-4912-8035-0e1a9e0e715d-kube-api-access-ssdv7\") pod \"openstack-galera-1\" (UID: \"41360e5e-2620-4912-8035-0e1a9e0e715d\") " pod="glance-kuttl-tests/openstack-galera-1" Dec 01 08:54:27 crc kubenswrapper[4813]: I1201 08:54:27.400033 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/43c6449e-e9d4-4a5b-9231-d719d2a9e488-kolla-config\") pod \"openstack-galera-0\" (UID: \"43c6449e-e9d4-4a5b-9231-d719d2a9e488\") " pod="glance-kuttl-tests/openstack-galera-0" Dec 01 08:54:27 crc kubenswrapper[4813]: I1201 08:54:27.501862 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ssdv7\" (UniqueName: \"kubernetes.io/projected/41360e5e-2620-4912-8035-0e1a9e0e715d-kube-api-access-ssdv7\") pod \"openstack-galera-1\" (UID: \"41360e5e-2620-4912-8035-0e1a9e0e715d\") " pod="glance-kuttl-tests/openstack-galera-1" Dec 01 08:54:27 crc kubenswrapper[4813]: I1201 08:54:27.501928 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/43c6449e-e9d4-4a5b-9231-d719d2a9e488-kolla-config\") pod \"openstack-galera-0\" (UID: \"43c6449e-e9d4-4a5b-9231-d719d2a9e488\") " pod="glance-kuttl-tests/openstack-galera-0" Dec 01 08:54:27 crc kubenswrapper[4813]: I1201 08:54:27.501993 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/73988763-722d-4652-8fac-053ba8217547-config-data-generated\") pod \"openstack-galera-2\" (UID: \"73988763-722d-4652-8fac-053ba8217547\") " pod="glance-kuttl-tests/openstack-galera-2" Dec 01 08:54:27 crc kubenswrapper[4813]: I1201 08:54:27.502028 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage15-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage15-crc\") pod \"openstack-galera-2\" (UID: \"73988763-722d-4652-8fac-053ba8217547\") " pod="glance-kuttl-tests/openstack-galera-2" Dec 01 08:54:27 crc kubenswrapper[4813]: I1201 08:54:27.502064 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-str7c\" (UniqueName: \"kubernetes.io/projected/73988763-722d-4652-8fac-053ba8217547-kube-api-access-str7c\") pod \"openstack-galera-2\" (UID: \"73988763-722d-4652-8fac-053ba8217547\") " pod="glance-kuttl-tests/openstack-galera-2" Dec 01 08:54:27 crc kubenswrapper[4813]: I1201 08:54:27.502117 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d2wgq\" (UniqueName: \"kubernetes.io/projected/43c6449e-e9d4-4a5b-9231-d719d2a9e488-kube-api-access-d2wgq\") pod \"openstack-galera-0\" (UID: \"43c6449e-e9d4-4a5b-9231-d719d2a9e488\") " pod="glance-kuttl-tests/openstack-galera-0" Dec 01 08:54:27 crc kubenswrapper[4813]: I1201 08:54:27.502147 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/73988763-722d-4652-8fac-053ba8217547-config-data-default\") pod \"openstack-galera-2\" (UID: \"73988763-722d-4652-8fac-053ba8217547\") " pod="glance-kuttl-tests/openstack-galera-2" Dec 01 08:54:27 crc kubenswrapper[4813]: I1201 08:54:27.502181 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/41360e5e-2620-4912-8035-0e1a9e0e715d-config-data-default\") pod \"openstack-galera-1\" (UID: \"41360e5e-2620-4912-8035-0e1a9e0e715d\") " pod="glance-kuttl-tests/openstack-galera-1" Dec 01 08:54:27 crc kubenswrapper[4813]: I1201 08:54:27.502214 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/73988763-722d-4652-8fac-053ba8217547-kolla-config\") pod \"openstack-galera-2\" (UID: \"73988763-722d-4652-8fac-053ba8217547\") " pod="glance-kuttl-tests/openstack-galera-2" Dec 01 08:54:27 crc kubenswrapper[4813]: I1201 08:54:27.502260 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/43c6449e-e9d4-4a5b-9231-d719d2a9e488-config-data-generated\") pod \"openstack-galera-0\" (UID: \"43c6449e-e9d4-4a5b-9231-d719d2a9e488\") " pod="glance-kuttl-tests/openstack-galera-0" Dec 01 08:54:27 crc kubenswrapper[4813]: I1201 08:54:27.502322 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/41360e5e-2620-4912-8035-0e1a9e0e715d-kolla-config\") pod \"openstack-galera-1\" (UID: \"41360e5e-2620-4912-8035-0e1a9e0e715d\") " pod="glance-kuttl-tests/openstack-galera-1" Dec 01 08:54:27 crc kubenswrapper[4813]: I1201 08:54:27.502362 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/41360e5e-2620-4912-8035-0e1a9e0e715d-operator-scripts\") pod \"openstack-galera-1\" (UID: \"41360e5e-2620-4912-8035-0e1a9e0e715d\") " pod="glance-kuttl-tests/openstack-galera-1" Dec 01 08:54:27 crc kubenswrapper[4813]: I1201 08:54:27.502394 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/43c6449e-e9d4-4a5b-9231-d719d2a9e488-config-data-default\") pod \"openstack-galera-0\" (UID: \"43c6449e-e9d4-4a5b-9231-d719d2a9e488\") " pod="glance-kuttl-tests/openstack-galera-0" Dec 01 08:54:27 crc kubenswrapper[4813]: I1201 08:54:27.502447 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"openstack-galera-0\" (UID: \"43c6449e-e9d4-4a5b-9231-d719d2a9e488\") " pod="glance-kuttl-tests/openstack-galera-0" Dec 01 08:54:27 crc kubenswrapper[4813]: I1201 08:54:27.502460 4813 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage15-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage15-crc\") pod \"openstack-galera-2\" (UID: \"73988763-722d-4652-8fac-053ba8217547\") device mount path \"/mnt/openstack/pv15\"" pod="glance-kuttl-tests/openstack-galera-2" Dec 01 08:54:27 crc kubenswrapper[4813]: I1201 08:54:27.502473 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/43c6449e-e9d4-4a5b-9231-d719d2a9e488-operator-scripts\") pod \"openstack-galera-0\" (UID: \"43c6449e-e9d4-4a5b-9231-d719d2a9e488\") " pod="glance-kuttl-tests/openstack-galera-0" Dec 01 08:54:27 crc kubenswrapper[4813]: I1201 08:54:27.502500 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/41360e5e-2620-4912-8035-0e1a9e0e715d-config-data-generated\") pod \"openstack-galera-1\" (UID: \"41360e5e-2620-4912-8035-0e1a9e0e715d\") " pod="glance-kuttl-tests/openstack-galera-1" Dec 01 08:54:27 crc kubenswrapper[4813]: I1201 08:54:27.502532 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"openstack-galera-1\" (UID: \"41360e5e-2620-4912-8035-0e1a9e0e715d\") " pod="glance-kuttl-tests/openstack-galera-1" Dec 01 08:54:27 crc kubenswrapper[4813]: I1201 08:54:27.502547 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/73988763-722d-4652-8fac-053ba8217547-config-data-generated\") pod \"openstack-galera-2\" (UID: \"73988763-722d-4652-8fac-053ba8217547\") " pod="glance-kuttl-tests/openstack-galera-2" Dec 01 08:54:27 crc kubenswrapper[4813]: I1201 08:54:27.502557 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/73988763-722d-4652-8fac-053ba8217547-operator-scripts\") pod \"openstack-galera-2\" (UID: \"73988763-722d-4652-8fac-053ba8217547\") " pod="glance-kuttl-tests/openstack-galera-2" Dec 01 08:54:27 crc kubenswrapper[4813]: I1201 08:54:27.502843 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/43c6449e-e9d4-4a5b-9231-d719d2a9e488-config-data-generated\") pod \"openstack-galera-0\" (UID: \"43c6449e-e9d4-4a5b-9231-d719d2a9e488\") " pod="glance-kuttl-tests/openstack-galera-0" Dec 01 08:54:27 crc kubenswrapper[4813]: I1201 08:54:27.503191 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/73988763-722d-4652-8fac-053ba8217547-config-data-default\") pod \"openstack-galera-2\" (UID: \"73988763-722d-4652-8fac-053ba8217547\") " pod="glance-kuttl-tests/openstack-galera-2" Dec 01 08:54:27 crc kubenswrapper[4813]: I1201 08:54:27.503190 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/43c6449e-e9d4-4a5b-9231-d719d2a9e488-kolla-config\") pod \"openstack-galera-0\" (UID: \"43c6449e-e9d4-4a5b-9231-d719d2a9e488\") " pod="glance-kuttl-tests/openstack-galera-0" Dec 01 08:54:27 crc kubenswrapper[4813]: I1201 08:54:27.503369 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/73988763-722d-4652-8fac-053ba8217547-kolla-config\") pod \"openstack-galera-2\" (UID: \"73988763-722d-4652-8fac-053ba8217547\") " pod="glance-kuttl-tests/openstack-galera-2" Dec 01 08:54:27 crc kubenswrapper[4813]: I1201 08:54:27.503395 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/41360e5e-2620-4912-8035-0e1a9e0e715d-kolla-config\") pod \"openstack-galera-1\" (UID: \"41360e5e-2620-4912-8035-0e1a9e0e715d\") " pod="glance-kuttl-tests/openstack-galera-1" Dec 01 08:54:27 crc kubenswrapper[4813]: I1201 08:54:27.503454 4813 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"openstack-galera-0\" (UID: \"43c6449e-e9d4-4a5b-9231-d719d2a9e488\") device mount path \"/mnt/openstack/pv06\"" pod="glance-kuttl-tests/openstack-galera-0" Dec 01 08:54:27 crc kubenswrapper[4813]: I1201 08:54:27.503659 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/41360e5e-2620-4912-8035-0e1a9e0e715d-config-data-generated\") pod \"openstack-galera-1\" (UID: \"41360e5e-2620-4912-8035-0e1a9e0e715d\") " pod="glance-kuttl-tests/openstack-galera-1" Dec 01 08:54:27 crc kubenswrapper[4813]: I1201 08:54:27.503728 4813 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"openstack-galera-1\" (UID: \"41360e5e-2620-4912-8035-0e1a9e0e715d\") device mount path \"/mnt/openstack/pv09\"" pod="glance-kuttl-tests/openstack-galera-1" Dec 01 08:54:27 crc kubenswrapper[4813]: I1201 08:54:27.504708 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/41360e5e-2620-4912-8035-0e1a9e0e715d-operator-scripts\") pod \"openstack-galera-1\" (UID: \"41360e5e-2620-4912-8035-0e1a9e0e715d\") " pod="glance-kuttl-tests/openstack-galera-1" Dec 01 08:54:27 crc kubenswrapper[4813]: I1201 08:54:27.504782 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/41360e5e-2620-4912-8035-0e1a9e0e715d-config-data-default\") pod \"openstack-galera-1\" (UID: \"41360e5e-2620-4912-8035-0e1a9e0e715d\") " pod="glance-kuttl-tests/openstack-galera-1" Dec 01 08:54:27 crc kubenswrapper[4813]: I1201 08:54:27.504956 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/73988763-722d-4652-8fac-053ba8217547-operator-scripts\") pod \"openstack-galera-2\" (UID: \"73988763-722d-4652-8fac-053ba8217547\") " pod="glance-kuttl-tests/openstack-galera-2" Dec 01 08:54:27 crc kubenswrapper[4813]: I1201 08:54:27.505422 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/43c6449e-e9d4-4a5b-9231-d719d2a9e488-operator-scripts\") pod \"openstack-galera-0\" (UID: \"43c6449e-e9d4-4a5b-9231-d719d2a9e488\") " pod="glance-kuttl-tests/openstack-galera-0" Dec 01 08:54:27 crc kubenswrapper[4813]: I1201 08:54:27.505626 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/43c6449e-e9d4-4a5b-9231-d719d2a9e488-config-data-default\") pod \"openstack-galera-0\" (UID: \"43c6449e-e9d4-4a5b-9231-d719d2a9e488\") " pod="glance-kuttl-tests/openstack-galera-0" Dec 01 08:54:27 crc kubenswrapper[4813]: I1201 08:54:27.522382 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"openstack-galera-1\" (UID: \"41360e5e-2620-4912-8035-0e1a9e0e715d\") " pod="glance-kuttl-tests/openstack-galera-1" Dec 01 08:54:27 crc kubenswrapper[4813]: I1201 08:54:27.524171 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-str7c\" (UniqueName: \"kubernetes.io/projected/73988763-722d-4652-8fac-053ba8217547-kube-api-access-str7c\") pod \"openstack-galera-2\" (UID: \"73988763-722d-4652-8fac-053ba8217547\") " pod="glance-kuttl-tests/openstack-galera-2" Dec 01 08:54:27 crc kubenswrapper[4813]: I1201 08:54:27.527244 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ssdv7\" (UniqueName: \"kubernetes.io/projected/41360e5e-2620-4912-8035-0e1a9e0e715d-kube-api-access-ssdv7\") pod \"openstack-galera-1\" (UID: \"41360e5e-2620-4912-8035-0e1a9e0e715d\") " pod="glance-kuttl-tests/openstack-galera-1" Dec 01 08:54:27 crc kubenswrapper[4813]: I1201 08:54:27.527598 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d2wgq\" (UniqueName: \"kubernetes.io/projected/43c6449e-e9d4-4a5b-9231-d719d2a9e488-kube-api-access-d2wgq\") pod \"openstack-galera-0\" (UID: \"43c6449e-e9d4-4a5b-9231-d719d2a9e488\") " pod="glance-kuttl-tests/openstack-galera-0" Dec 01 08:54:27 crc kubenswrapper[4813]: I1201 08:54:27.529920 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage15-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage15-crc\") pod \"openstack-galera-2\" (UID: \"73988763-722d-4652-8fac-053ba8217547\") " pod="glance-kuttl-tests/openstack-galera-2" Dec 01 08:54:27 crc kubenswrapper[4813]: I1201 08:54:27.544931 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"openstack-galera-0\" (UID: \"43c6449e-e9d4-4a5b-9231-d719d2a9e488\") " pod="glance-kuttl-tests/openstack-galera-0" Dec 01 08:54:27 crc kubenswrapper[4813]: I1201 08:54:27.582820 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-84d8d7c7c5-bcbsv" event={"ID":"42e1cee7-cd3a-4194-af37-e29d5f04389a","Type":"ContainerStarted","Data":"01e4000fcaa813ab832d04bb6b4deffd5c3f3fcc49674f0acf6f665517a9cdd0"} Dec 01 08:54:27 crc kubenswrapper[4813]: I1201 08:54:27.619489 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/openstack-galera-0" Dec 01 08:54:27 crc kubenswrapper[4813]: I1201 08:54:27.629594 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/openstack-galera-1" Dec 01 08:54:27 crc kubenswrapper[4813]: I1201 08:54:27.638300 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/openstack-galera-2" Dec 01 08:54:27 crc kubenswrapper[4813]: I1201 08:54:27.844949 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/openstack-galera-0"] Dec 01 08:54:28 crc kubenswrapper[4813]: I1201 08:54:28.121146 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/openstack-galera-2"] Dec 01 08:54:28 crc kubenswrapper[4813]: I1201 08:54:28.128805 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/openstack-galera-1"] Dec 01 08:54:28 crc kubenswrapper[4813]: W1201 08:54:28.173089 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod41360e5e_2620_4912_8035_0e1a9e0e715d.slice/crio-44694b5fe25334469ed41d3b7c562e53f7524f7f68c06165c5d770d0aa2bd941 WatchSource:0}: Error finding container 44694b5fe25334469ed41d3b7c562e53f7524f7f68c06165c5d770d0aa2bd941: Status 404 returned error can't find the container with id 44694b5fe25334469ed41d3b7c562e53f7524f7f68c06165c5d770d0aa2bd941 Dec 01 08:54:28 crc kubenswrapper[4813]: I1201 08:54:28.592367 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstack-galera-0" event={"ID":"43c6449e-e9d4-4a5b-9231-d719d2a9e488","Type":"ContainerStarted","Data":"e8a88efbe8dab4460c62a1c604a1e043a164a3ce2d479526c91ba789344981e3"} Dec 01 08:54:28 crc kubenswrapper[4813]: I1201 08:54:28.593594 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstack-galera-1" event={"ID":"41360e5e-2620-4912-8035-0e1a9e0e715d","Type":"ContainerStarted","Data":"44694b5fe25334469ed41d3b7c562e53f7524f7f68c06165c5d770d0aa2bd941"} Dec 01 08:54:28 crc kubenswrapper[4813]: I1201 08:54:28.595382 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstack-galera-2" event={"ID":"73988763-722d-4652-8fac-053ba8217547","Type":"ContainerStarted","Data":"ab843b8177030ce9338918bf46de7538032e92a628fb669df6123fd1ee964d6c"} Dec 01 08:54:30 crc kubenswrapper[4813]: I1201 08:54:30.812846 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-84d8d7c7c5-bcbsv" event={"ID":"42e1cee7-cd3a-4194-af37-e29d5f04389a","Type":"ContainerStarted","Data":"ce31b4e09a2acc90072d380007f5b9a42af04d5db29e67e6451cfcba75bdf6e1"} Dec 01 08:54:30 crc kubenswrapper[4813]: I1201 08:54:30.817946 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-84d8d7c7c5-bcbsv" Dec 01 08:54:30 crc kubenswrapper[4813]: I1201 08:54:30.841489 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-controller-manager-84d8d7c7c5-bcbsv" podStartSLOduration=1.4787173089999999 podStartE2EDuration="6.841425769s" podCreationTimestamp="2025-12-01 08:54:24 +0000 UTC" firstStartedPulling="2025-12-01 08:54:25.269958524 +0000 UTC m=+849.892880110" lastFinishedPulling="2025-12-01 08:54:30.632666984 +0000 UTC m=+855.255588570" observedRunningTime="2025-12-01 08:54:30.839691001 +0000 UTC m=+855.462612587" watchObservedRunningTime="2025-12-01 08:54:30.841425769 +0000 UTC m=+855.464347355" Dec 01 08:54:32 crc kubenswrapper[4813]: I1201 08:54:32.829824 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-84d8d7c7c5-bcbsv" Dec 01 08:54:35 crc kubenswrapper[4813]: E1201 08:54:35.176386 4813 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = reading blob sha256:9edb696970a5d944aa0a013096ed565e4092e9a751f383a80fdde57bd71155e3: Get \"https://cdn01.quay.io/quayio-production-s3/sha256/9e/9edb696970a5d944aa0a013096ed565e4092e9a751f383a80fdde57bd71155e3?X-Amz-Algorithm=AWS4-HMAC-SHA256&X-Amz-Credential=AKIATAAF2YHTGR23ZTE6%2F20251201%2Fus-east-1%2Fs3%2Faws4_request&X-Amz-Date=20251201T085429Z&X-Amz-Expires=600&X-Amz-SignedHeaders=host&X-Amz-Signature=1dd09fcad766c4176985301f4dcce042f380dd2f4ec361757b1385aec10b6dab®ion=us-east-1&namespace=podified-antelope-centos9&username=openshift-release-dev+ocm_access_1b89217552bc42d1be3fb06a1aed001a&repo_name=openstack-mariadb&akamai_signature=exp=1764580169~hmac=bc5534b95b9438cd25cb97712a0ed6eba93ed19bceca1603056bd07e14723ce3\": EOF" image="quay.io/podified-antelope-centos9/openstack-mariadb@sha256:10452e2144368e2f128c8fb8ef9e54880b06ef1d71d9f084a0217dcb099c51ce" Dec 01 08:54:35 crc kubenswrapper[4813]: E1201 08:54:35.177479 4813 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:mysql-bootstrap,Image:quay.io/podified-antelope-centos9/openstack-mariadb@sha256:10452e2144368e2f128c8fb8ef9e54880b06ef1d71d9f084a0217dcb099c51ce,Command:[bash /var/lib/operator-scripts/mysql_bootstrap.sh],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:True,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:mysql-db,ReadOnly:false,MountPath:/var/lib/mysql,SubPath:mysql,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data-default,ReadOnly:true,MountPath:/var/lib/config-data/default,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data-generated,ReadOnly:false,MountPath:/var/lib/config-data/generated,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:operator-scripts,ReadOnly:true,MountPath:/var/lib/operator-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kolla-config,ReadOnly:true,MountPath:/var/lib/kolla/config_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-ssdv7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod openstack-galera-1_glance-kuttl-tests(41360e5e-2620-4912-8035-0e1a9e0e715d): ErrImagePull: reading blob sha256:9edb696970a5d944aa0a013096ed565e4092e9a751f383a80fdde57bd71155e3: Get \"https://cdn01.quay.io/quayio-production-s3/sha256/9e/9edb696970a5d944aa0a013096ed565e4092e9a751f383a80fdde57bd71155e3?X-Amz-Algorithm=AWS4-HMAC-SHA256&X-Amz-Credential=AKIATAAF2YHTGR23ZTE6%2F20251201%2Fus-east-1%2Fs3%2Faws4_request&X-Amz-Date=20251201T085429Z&X-Amz-Expires=600&X-Amz-SignedHeaders=host&X-Amz-Signature=1dd09fcad766c4176985301f4dcce042f380dd2f4ec361757b1385aec10b6dab®ion=us-east-1&namespace=podified-antelope-centos9&username=openshift-release-dev+ocm_access_1b89217552bc42d1be3fb06a1aed001a&repo_name=openstack-mariadb&akamai_signature=exp=1764580169~hmac=bc5534b95b9438cd25cb97712a0ed6eba93ed19bceca1603056bd07e14723ce3\": EOF" logger="UnhandledError" Dec 01 08:54:35 crc kubenswrapper[4813]: E1201 08:54:35.178876 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"mysql-bootstrap\" with ErrImagePull: \"reading blob sha256:9edb696970a5d944aa0a013096ed565e4092e9a751f383a80fdde57bd71155e3: Get \\\"https://cdn01.quay.io/quayio-production-s3/sha256/9e/9edb696970a5d944aa0a013096ed565e4092e9a751f383a80fdde57bd71155e3?X-Amz-Algorithm=AWS4-HMAC-SHA256&X-Amz-Credential=AKIATAAF2YHTGR23ZTE6%2F20251201%2Fus-east-1%2Fs3%2Faws4_request&X-Amz-Date=20251201T085429Z&X-Amz-Expires=600&X-Amz-SignedHeaders=host&X-Amz-Signature=1dd09fcad766c4176985301f4dcce042f380dd2f4ec361757b1385aec10b6dab®ion=us-east-1&namespace=podified-antelope-centos9&username=openshift-release-dev+ocm_access_1b89217552bc42d1be3fb06a1aed001a&repo_name=openstack-mariadb&akamai_signature=exp=1764580169~hmac=bc5534b95b9438cd25cb97712a0ed6eba93ed19bceca1603056bd07e14723ce3\\\": EOF\"" pod="glance-kuttl-tests/openstack-galera-1" podUID="41360e5e-2620-4912-8035-0e1a9e0e715d" Dec 01 08:54:35 crc kubenswrapper[4813]: E1201 08:54:35.847358 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"mysql-bootstrap\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-mariadb@sha256:10452e2144368e2f128c8fb8ef9e54880b06ef1d71d9f084a0217dcb099c51ce\\\"\"" pod="glance-kuttl-tests/openstack-galera-1" podUID="41360e5e-2620-4912-8035-0e1a9e0e715d" Dec 01 08:54:38 crc kubenswrapper[4813]: I1201 08:54:38.878835 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstack-galera-2" event={"ID":"73988763-722d-4652-8fac-053ba8217547","Type":"ContainerStarted","Data":"3642faeb2d7a5ddd953c03fd847808337be3eca66a804d81a464654fcd35ae82"} Dec 01 08:54:38 crc kubenswrapper[4813]: I1201 08:54:38.880871 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstack-galera-0" event={"ID":"43c6449e-e9d4-4a5b-9231-d719d2a9e488","Type":"ContainerStarted","Data":"6731c56ca57e019f8d83a6b260a34487a0f5035a833742a7aaafaaa1b9463302"} Dec 01 08:54:41 crc kubenswrapper[4813]: I1201 08:54:41.164748 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-index-ddpc8"] Dec 01 08:54:41 crc kubenswrapper[4813]: I1201 08:54:41.166195 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-index-ddpc8" Dec 01 08:54:41 crc kubenswrapper[4813]: I1201 08:54:41.169396 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-index-dockercfg-zsrbg" Dec 01 08:54:41 crc kubenswrapper[4813]: I1201 08:54:41.174006 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-index-ddpc8"] Dec 01 08:54:41 crc kubenswrapper[4813]: I1201 08:54:41.231699 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z7t4d\" (UniqueName: \"kubernetes.io/projected/3f8cb860-7fc8-490c-b5d9-18eed8d8db79-kube-api-access-z7t4d\") pod \"rabbitmq-cluster-operator-index-ddpc8\" (UID: \"3f8cb860-7fc8-490c-b5d9-18eed8d8db79\") " pod="openstack-operators/rabbitmq-cluster-operator-index-ddpc8" Dec 01 08:54:41 crc kubenswrapper[4813]: I1201 08:54:41.333112 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z7t4d\" (UniqueName: \"kubernetes.io/projected/3f8cb860-7fc8-490c-b5d9-18eed8d8db79-kube-api-access-z7t4d\") pod \"rabbitmq-cluster-operator-index-ddpc8\" (UID: \"3f8cb860-7fc8-490c-b5d9-18eed8d8db79\") " pod="openstack-operators/rabbitmq-cluster-operator-index-ddpc8" Dec 01 08:54:41 crc kubenswrapper[4813]: I1201 08:54:41.369418 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z7t4d\" (UniqueName: \"kubernetes.io/projected/3f8cb860-7fc8-490c-b5d9-18eed8d8db79-kube-api-access-z7t4d\") pod \"rabbitmq-cluster-operator-index-ddpc8\" (UID: \"3f8cb860-7fc8-490c-b5d9-18eed8d8db79\") " pod="openstack-operators/rabbitmq-cluster-operator-index-ddpc8" Dec 01 08:54:41 crc kubenswrapper[4813]: I1201 08:54:41.485045 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-index-ddpc8" Dec 01 08:54:41 crc kubenswrapper[4813]: I1201 08:54:41.926635 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-index-ddpc8"] Dec 01 08:54:41 crc kubenswrapper[4813]: W1201 08:54:41.932926 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3f8cb860_7fc8_490c_b5d9_18eed8d8db79.slice/crio-f07d2f83e779093422c243c704ab50773bbc6fdcc0958ee65c257ccb2bfe3934 WatchSource:0}: Error finding container f07d2f83e779093422c243c704ab50773bbc6fdcc0958ee65c257ccb2bfe3934: Status 404 returned error can't find the container with id f07d2f83e779093422c243c704ab50773bbc6fdcc0958ee65c257ccb2bfe3934 Dec 01 08:54:42 crc kubenswrapper[4813]: I1201 08:54:42.908848 4813 generic.go:334] "Generic (PLEG): container finished" podID="73988763-722d-4652-8fac-053ba8217547" containerID="3642faeb2d7a5ddd953c03fd847808337be3eca66a804d81a464654fcd35ae82" exitCode=0 Dec 01 08:54:42 crc kubenswrapper[4813]: I1201 08:54:42.909127 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstack-galera-2" event={"ID":"73988763-722d-4652-8fac-053ba8217547","Type":"ContainerDied","Data":"3642faeb2d7a5ddd953c03fd847808337be3eca66a804d81a464654fcd35ae82"} Dec 01 08:54:42 crc kubenswrapper[4813]: I1201 08:54:42.912988 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstack-galera-0" event={"ID":"43c6449e-e9d4-4a5b-9231-d719d2a9e488","Type":"ContainerDied","Data":"6731c56ca57e019f8d83a6b260a34487a0f5035a833742a7aaafaaa1b9463302"} Dec 01 08:54:42 crc kubenswrapper[4813]: I1201 08:54:42.912991 4813 generic.go:334] "Generic (PLEG): container finished" podID="43c6449e-e9d4-4a5b-9231-d719d2a9e488" containerID="6731c56ca57e019f8d83a6b260a34487a0f5035a833742a7aaafaaa1b9463302" exitCode=0 Dec 01 08:54:42 crc kubenswrapper[4813]: I1201 08:54:42.915200 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-index-ddpc8" event={"ID":"3f8cb860-7fc8-490c-b5d9-18eed8d8db79","Type":"ContainerStarted","Data":"f07d2f83e779093422c243c704ab50773bbc6fdcc0958ee65c257ccb2bfe3934"} Dec 01 08:54:43 crc kubenswrapper[4813]: I1201 08:54:43.923395 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstack-galera-2" event={"ID":"73988763-722d-4652-8fac-053ba8217547","Type":"ContainerStarted","Data":"f7024039c493e9165151be3914c80e6c5d5142a4eaf9d9c184364cc64a3f8f0e"} Dec 01 08:54:43 crc kubenswrapper[4813]: I1201 08:54:43.925430 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstack-galera-0" event={"ID":"43c6449e-e9d4-4a5b-9231-d719d2a9e488","Type":"ContainerStarted","Data":"d5ede1c2b7da645352abfeb35f403960e3939f0491ad66bc5289579290d4c59c"} Dec 01 08:54:43 crc kubenswrapper[4813]: I1201 08:54:43.943576 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/openstack-galera-2" podStartSLOduration=7.953847098 podStartE2EDuration="17.943557879s" podCreationTimestamp="2025-12-01 08:54:26 +0000 UTC" firstStartedPulling="2025-12-01 08:54:28.239502211 +0000 UTC m=+852.862423797" lastFinishedPulling="2025-12-01 08:54:38.229212992 +0000 UTC m=+862.852134578" observedRunningTime="2025-12-01 08:54:43.943376124 +0000 UTC m=+868.566297710" watchObservedRunningTime="2025-12-01 08:54:43.943557879 +0000 UTC m=+868.566479465" Dec 01 08:54:43 crc kubenswrapper[4813]: I1201 08:54:43.968657 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/openstack-galera-0" podStartSLOduration=7.575354229 podStartE2EDuration="17.968638183s" podCreationTimestamp="2025-12-01 08:54:26 +0000 UTC" firstStartedPulling="2025-12-01 08:54:27.858949195 +0000 UTC m=+852.481870781" lastFinishedPulling="2025-12-01 08:54:38.252233139 +0000 UTC m=+862.875154735" observedRunningTime="2025-12-01 08:54:43.966476053 +0000 UTC m=+868.589397659" watchObservedRunningTime="2025-12-01 08:54:43.968638183 +0000 UTC m=+868.591559769" Dec 01 08:54:45 crc kubenswrapper[4813]: I1201 08:54:45.941208 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-index-ddpc8" event={"ID":"3f8cb860-7fc8-490c-b5d9-18eed8d8db79","Type":"ContainerStarted","Data":"51afe760f18b00bc5ad4ad6d1aa2d5d888620e8de006174ea8a3610a0f990352"} Dec 01 08:54:46 crc kubenswrapper[4813]: I1201 08:54:46.004942 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-index-ddpc8" podStartSLOduration=1.5616683550000001 podStartE2EDuration="5.004894634s" podCreationTimestamp="2025-12-01 08:54:41 +0000 UTC" firstStartedPulling="2025-12-01 08:54:41.934958581 +0000 UTC m=+866.557880167" lastFinishedPulling="2025-12-01 08:54:45.37818483 +0000 UTC m=+870.001106446" observedRunningTime="2025-12-01 08:54:45.999609328 +0000 UTC m=+870.622530914" watchObservedRunningTime="2025-12-01 08:54:46.004894634 +0000 UTC m=+870.627816220" Dec 01 08:54:47 crc kubenswrapper[4813]: I1201 08:54:47.221663 4813 patch_prober.go:28] interesting pod/machine-config-daemon-9wn4q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 08:54:47 crc kubenswrapper[4813]: I1201 08:54:47.221765 4813 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9wn4q" podUID="3d423402-e08f-42e9-b60c-6eb9bf2f6d48" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 08:54:47 crc kubenswrapper[4813]: I1201 08:54:47.620251 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/openstack-galera-0" Dec 01 08:54:47 crc kubenswrapper[4813]: I1201 08:54:47.620569 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/openstack-galera-0" Dec 01 08:54:47 crc kubenswrapper[4813]: I1201 08:54:47.639580 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/openstack-galera-2" Dec 01 08:54:47 crc kubenswrapper[4813]: I1201 08:54:47.639639 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/openstack-galera-2" Dec 01 08:54:48 crc kubenswrapper[4813]: I1201 08:54:48.537231 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/memcached-0"] Dec 01 08:54:48 crc kubenswrapper[4813]: I1201 08:54:48.538601 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/memcached-0" Dec 01 08:54:48 crc kubenswrapper[4813]: I1201 08:54:48.541215 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"memcached-config-data" Dec 01 08:54:48 crc kubenswrapper[4813]: I1201 08:54:48.541318 4813 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"memcached-memcached-dockercfg-5b5fz" Dec 01 08:54:48 crc kubenswrapper[4813]: I1201 08:54:48.548202 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/memcached-0"] Dec 01 08:54:48 crc kubenswrapper[4813]: I1201 08:54:48.661315 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j5c68\" (UniqueName: \"kubernetes.io/projected/88d8ade8-c257-464c-9dab-0052bed99037-kube-api-access-j5c68\") pod \"memcached-0\" (UID: \"88d8ade8-c257-464c-9dab-0052bed99037\") " pod="glance-kuttl-tests/memcached-0" Dec 01 08:54:48 crc kubenswrapper[4813]: I1201 08:54:48.661366 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/88d8ade8-c257-464c-9dab-0052bed99037-config-data\") pod \"memcached-0\" (UID: \"88d8ade8-c257-464c-9dab-0052bed99037\") " pod="glance-kuttl-tests/memcached-0" Dec 01 08:54:48 crc kubenswrapper[4813]: I1201 08:54:48.661404 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/88d8ade8-c257-464c-9dab-0052bed99037-kolla-config\") pod \"memcached-0\" (UID: \"88d8ade8-c257-464c-9dab-0052bed99037\") " pod="glance-kuttl-tests/memcached-0" Dec 01 08:54:48 crc kubenswrapper[4813]: I1201 08:54:48.763546 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/88d8ade8-c257-464c-9dab-0052bed99037-config-data\") pod \"memcached-0\" (UID: \"88d8ade8-c257-464c-9dab-0052bed99037\") " pod="glance-kuttl-tests/memcached-0" Dec 01 08:54:48 crc kubenswrapper[4813]: I1201 08:54:48.764024 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/88d8ade8-c257-464c-9dab-0052bed99037-kolla-config\") pod \"memcached-0\" (UID: \"88d8ade8-c257-464c-9dab-0052bed99037\") " pod="glance-kuttl-tests/memcached-0" Dec 01 08:54:48 crc kubenswrapper[4813]: I1201 08:54:48.764339 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j5c68\" (UniqueName: \"kubernetes.io/projected/88d8ade8-c257-464c-9dab-0052bed99037-kube-api-access-j5c68\") pod \"memcached-0\" (UID: \"88d8ade8-c257-464c-9dab-0052bed99037\") " pod="glance-kuttl-tests/memcached-0" Dec 01 08:54:48 crc kubenswrapper[4813]: I1201 08:54:48.764403 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/88d8ade8-c257-464c-9dab-0052bed99037-config-data\") pod \"memcached-0\" (UID: \"88d8ade8-c257-464c-9dab-0052bed99037\") " pod="glance-kuttl-tests/memcached-0" Dec 01 08:54:48 crc kubenswrapper[4813]: I1201 08:54:48.764757 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/88d8ade8-c257-464c-9dab-0052bed99037-kolla-config\") pod \"memcached-0\" (UID: \"88d8ade8-c257-464c-9dab-0052bed99037\") " pod="glance-kuttl-tests/memcached-0" Dec 01 08:54:48 crc kubenswrapper[4813]: I1201 08:54:48.791535 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j5c68\" (UniqueName: \"kubernetes.io/projected/88d8ade8-c257-464c-9dab-0052bed99037-kube-api-access-j5c68\") pod \"memcached-0\" (UID: \"88d8ade8-c257-464c-9dab-0052bed99037\") " pod="glance-kuttl-tests/memcached-0" Dec 01 08:54:48 crc kubenswrapper[4813]: I1201 08:54:48.901835 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/memcached-0" Dec 01 08:54:49 crc kubenswrapper[4813]: I1201 08:54:49.395036 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/memcached-0"] Dec 01 08:54:49 crc kubenswrapper[4813]: I1201 08:54:49.965214 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/memcached-0" event={"ID":"88d8ade8-c257-464c-9dab-0052bed99037","Type":"ContainerStarted","Data":"782914df518ac8b07705916b0c20a4afa91c8fa9818a88f110730f48a8000674"} Dec 01 08:54:50 crc kubenswrapper[4813]: I1201 08:54:50.974128 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstack-galera-1" event={"ID":"41360e5e-2620-4912-8035-0e1a9e0e715d","Type":"ContainerStarted","Data":"3a5a49e275f1242a56cc581a872d2ad308bdbc43109cfce09b8356e5e684bc3a"} Dec 01 08:54:51 crc kubenswrapper[4813]: I1201 08:54:51.486143 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/rabbitmq-cluster-operator-index-ddpc8" Dec 01 08:54:51 crc kubenswrapper[4813]: I1201 08:54:51.486472 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/rabbitmq-cluster-operator-index-ddpc8" Dec 01 08:54:51 crc kubenswrapper[4813]: I1201 08:54:51.517616 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/rabbitmq-cluster-operator-index-ddpc8" Dec 01 08:54:51 crc kubenswrapper[4813]: I1201 08:54:51.981110 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/memcached-0" event={"ID":"88d8ade8-c257-464c-9dab-0052bed99037","Type":"ContainerStarted","Data":"13647640697f73f4b7f03b0bcbd720969780ef3b11f627085ddd2eb78f70d473"} Dec 01 08:54:51 crc kubenswrapper[4813]: I1201 08:54:51.981291 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/memcached-0" Dec 01 08:54:51 crc kubenswrapper[4813]: I1201 08:54:51.996478 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/memcached-0" podStartSLOduration=1.676845468 podStartE2EDuration="3.996458398s" podCreationTimestamp="2025-12-01 08:54:48 +0000 UTC" firstStartedPulling="2025-12-01 08:54:49.398365856 +0000 UTC m=+874.021287432" lastFinishedPulling="2025-12-01 08:54:51.717978776 +0000 UTC m=+876.340900362" observedRunningTime="2025-12-01 08:54:51.994434722 +0000 UTC m=+876.617356308" watchObservedRunningTime="2025-12-01 08:54:51.996458398 +0000 UTC m=+876.619379984" Dec 01 08:54:52 crc kubenswrapper[4813]: I1201 08:54:52.006876 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/rabbitmq-cluster-operator-index-ddpc8" Dec 01 08:54:54 crc kubenswrapper[4813]: I1201 08:54:54.005227 4813 generic.go:334] "Generic (PLEG): container finished" podID="41360e5e-2620-4912-8035-0e1a9e0e715d" containerID="3a5a49e275f1242a56cc581a872d2ad308bdbc43109cfce09b8356e5e684bc3a" exitCode=0 Dec 01 08:54:54 crc kubenswrapper[4813]: I1201 08:54:54.005400 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstack-galera-1" event={"ID":"41360e5e-2620-4912-8035-0e1a9e0e715d","Type":"ContainerDied","Data":"3a5a49e275f1242a56cc581a872d2ad308bdbc43109cfce09b8356e5e684bc3a"} Dec 01 08:54:55 crc kubenswrapper[4813]: I1201 08:54:55.017561 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstack-galera-1" event={"ID":"41360e5e-2620-4912-8035-0e1a9e0e715d","Type":"ContainerStarted","Data":"f186dd2ac93b4f28668082627ae26d9d2e81b0ec2304484ef52e31414559378e"} Dec 01 08:54:55 crc kubenswrapper[4813]: I1201 08:54:55.050668 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/openstack-galera-1" podStartSLOduration=-9223372007.80417 podStartE2EDuration="29.050605215s" podCreationTimestamp="2025-12-01 08:54:26 +0000 UTC" firstStartedPulling="2025-12-01 08:54:28.175609373 +0000 UTC m=+852.798530959" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:54:55.047703035 +0000 UTC m=+879.670624661" watchObservedRunningTime="2025-12-01 08:54:55.050605215 +0000 UTC m=+879.673526811" Dec 01 08:54:57 crc kubenswrapper[4813]: I1201 08:54:57.629995 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/openstack-galera-1" Dec 01 08:54:57 crc kubenswrapper[4813]: I1201 08:54:57.630396 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/openstack-galera-1" Dec 01 08:54:58 crc kubenswrapper[4813]: I1201 08:54:58.904105 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/memcached-0" Dec 01 08:54:59 crc kubenswrapper[4813]: I1201 08:54:59.384448 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-grnss"] Dec 01 08:54:59 crc kubenswrapper[4813]: I1201 08:54:59.387457 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-grnss" Dec 01 08:54:59 crc kubenswrapper[4813]: I1201 08:54:59.401607 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-grnss"] Dec 01 08:54:59 crc kubenswrapper[4813]: I1201 08:54:59.424477 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qdwd5\" (UniqueName: \"kubernetes.io/projected/5a12a2ee-1fa9-4e27-b8b5-0b1c1f721c5b-kube-api-access-qdwd5\") pod \"redhat-operators-grnss\" (UID: \"5a12a2ee-1fa9-4e27-b8b5-0b1c1f721c5b\") " pod="openshift-marketplace/redhat-operators-grnss" Dec 01 08:54:59 crc kubenswrapper[4813]: I1201 08:54:59.424687 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5a12a2ee-1fa9-4e27-b8b5-0b1c1f721c5b-catalog-content\") pod \"redhat-operators-grnss\" (UID: \"5a12a2ee-1fa9-4e27-b8b5-0b1c1f721c5b\") " pod="openshift-marketplace/redhat-operators-grnss" Dec 01 08:54:59 crc kubenswrapper[4813]: I1201 08:54:59.424735 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5a12a2ee-1fa9-4e27-b8b5-0b1c1f721c5b-utilities\") pod \"redhat-operators-grnss\" (UID: \"5a12a2ee-1fa9-4e27-b8b5-0b1c1f721c5b\") " pod="openshift-marketplace/redhat-operators-grnss" Dec 01 08:54:59 crc kubenswrapper[4813]: I1201 08:54:59.525580 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qdwd5\" (UniqueName: \"kubernetes.io/projected/5a12a2ee-1fa9-4e27-b8b5-0b1c1f721c5b-kube-api-access-qdwd5\") pod \"redhat-operators-grnss\" (UID: \"5a12a2ee-1fa9-4e27-b8b5-0b1c1f721c5b\") " pod="openshift-marketplace/redhat-operators-grnss" Dec 01 08:54:59 crc kubenswrapper[4813]: I1201 08:54:59.525703 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5a12a2ee-1fa9-4e27-b8b5-0b1c1f721c5b-catalog-content\") pod \"redhat-operators-grnss\" (UID: \"5a12a2ee-1fa9-4e27-b8b5-0b1c1f721c5b\") " pod="openshift-marketplace/redhat-operators-grnss" Dec 01 08:54:59 crc kubenswrapper[4813]: I1201 08:54:59.525733 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5a12a2ee-1fa9-4e27-b8b5-0b1c1f721c5b-utilities\") pod \"redhat-operators-grnss\" (UID: \"5a12a2ee-1fa9-4e27-b8b5-0b1c1f721c5b\") " pod="openshift-marketplace/redhat-operators-grnss" Dec 01 08:54:59 crc kubenswrapper[4813]: I1201 08:54:59.526460 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5a12a2ee-1fa9-4e27-b8b5-0b1c1f721c5b-utilities\") pod \"redhat-operators-grnss\" (UID: \"5a12a2ee-1fa9-4e27-b8b5-0b1c1f721c5b\") " pod="openshift-marketplace/redhat-operators-grnss" Dec 01 08:54:59 crc kubenswrapper[4813]: I1201 08:54:59.526476 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5a12a2ee-1fa9-4e27-b8b5-0b1c1f721c5b-catalog-content\") pod \"redhat-operators-grnss\" (UID: \"5a12a2ee-1fa9-4e27-b8b5-0b1c1f721c5b\") " pod="openshift-marketplace/redhat-operators-grnss" Dec 01 08:54:59 crc kubenswrapper[4813]: I1201 08:54:59.552905 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qdwd5\" (UniqueName: \"kubernetes.io/projected/5a12a2ee-1fa9-4e27-b8b5-0b1c1f721c5b-kube-api-access-qdwd5\") pod \"redhat-operators-grnss\" (UID: \"5a12a2ee-1fa9-4e27-b8b5-0b1c1f721c5b\") " pod="openshift-marketplace/redhat-operators-grnss" Dec 01 08:54:59 crc kubenswrapper[4813]: I1201 08:54:59.708083 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-grnss" Dec 01 08:55:00 crc kubenswrapper[4813]: I1201 08:55:00.031957 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5908vlmf"] Dec 01 08:55:00 crc kubenswrapper[4813]: I1201 08:55:00.033378 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5908vlmf" Dec 01 08:55:00 crc kubenswrapper[4813]: I1201 08:55:00.035826 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-fml9r" Dec 01 08:55:00 crc kubenswrapper[4813]: I1201 08:55:00.045282 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5908vlmf"] Dec 01 08:55:00 crc kubenswrapper[4813]: I1201 08:55:00.139725 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/e1cf13b6-094e-4524-82e7-f165e3b24767-util\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5908vlmf\" (UID: \"e1cf13b6-094e-4524-82e7-f165e3b24767\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5908vlmf" Dec 01 08:55:00 crc kubenswrapper[4813]: I1201 08:55:00.139894 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/e1cf13b6-094e-4524-82e7-f165e3b24767-bundle\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5908vlmf\" (UID: \"e1cf13b6-094e-4524-82e7-f165e3b24767\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5908vlmf" Dec 01 08:55:00 crc kubenswrapper[4813]: I1201 08:55:00.139977 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nrtqw\" (UniqueName: \"kubernetes.io/projected/e1cf13b6-094e-4524-82e7-f165e3b24767-kube-api-access-nrtqw\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5908vlmf\" (UID: \"e1cf13b6-094e-4524-82e7-f165e3b24767\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5908vlmf" Dec 01 08:55:00 crc kubenswrapper[4813]: I1201 08:55:00.212120 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-grnss"] Dec 01 08:55:00 crc kubenswrapper[4813]: I1201 08:55:00.241247 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/e1cf13b6-094e-4524-82e7-f165e3b24767-util\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5908vlmf\" (UID: \"e1cf13b6-094e-4524-82e7-f165e3b24767\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5908vlmf" Dec 01 08:55:00 crc kubenswrapper[4813]: I1201 08:55:00.241368 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/e1cf13b6-094e-4524-82e7-f165e3b24767-bundle\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5908vlmf\" (UID: \"e1cf13b6-094e-4524-82e7-f165e3b24767\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5908vlmf" Dec 01 08:55:00 crc kubenswrapper[4813]: I1201 08:55:00.241397 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nrtqw\" (UniqueName: \"kubernetes.io/projected/e1cf13b6-094e-4524-82e7-f165e3b24767-kube-api-access-nrtqw\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5908vlmf\" (UID: \"e1cf13b6-094e-4524-82e7-f165e3b24767\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5908vlmf" Dec 01 08:55:00 crc kubenswrapper[4813]: I1201 08:55:00.241707 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/e1cf13b6-094e-4524-82e7-f165e3b24767-util\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5908vlmf\" (UID: \"e1cf13b6-094e-4524-82e7-f165e3b24767\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5908vlmf" Dec 01 08:55:00 crc kubenswrapper[4813]: I1201 08:55:00.241946 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/e1cf13b6-094e-4524-82e7-f165e3b24767-bundle\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5908vlmf\" (UID: \"e1cf13b6-094e-4524-82e7-f165e3b24767\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5908vlmf" Dec 01 08:55:00 crc kubenswrapper[4813]: I1201 08:55:00.264830 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nrtqw\" (UniqueName: \"kubernetes.io/projected/e1cf13b6-094e-4524-82e7-f165e3b24767-kube-api-access-nrtqw\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5908vlmf\" (UID: \"e1cf13b6-094e-4524-82e7-f165e3b24767\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5908vlmf" Dec 01 08:55:00 crc kubenswrapper[4813]: I1201 08:55:00.351065 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5908vlmf" Dec 01 08:55:00 crc kubenswrapper[4813]: I1201 08:55:00.497792 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/openstack-galera-2" Dec 01 08:55:00 crc kubenswrapper[4813]: I1201 08:55:00.587769 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/openstack-galera-2" Dec 01 08:55:00 crc kubenswrapper[4813]: I1201 08:55:00.638643 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5908vlmf"] Dec 01 08:55:00 crc kubenswrapper[4813]: W1201 08:55:00.651856 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode1cf13b6_094e_4524_82e7_f165e3b24767.slice/crio-7e8e14deee73e7fa940053ccd7cac77d3b855c88dfd6a7e5470e76c8b70ae340 WatchSource:0}: Error finding container 7e8e14deee73e7fa940053ccd7cac77d3b855c88dfd6a7e5470e76c8b70ae340: Status 404 returned error can't find the container with id 7e8e14deee73e7fa940053ccd7cac77d3b855c88dfd6a7e5470e76c8b70ae340 Dec 01 08:55:01 crc kubenswrapper[4813]: I1201 08:55:01.063069 4813 generic.go:334] "Generic (PLEG): container finished" podID="5a12a2ee-1fa9-4e27-b8b5-0b1c1f721c5b" containerID="96227608cd663c5dc1fe44bd6cac947107c43e4749e4a97e836e0e805d33960c" exitCode=0 Dec 01 08:55:01 crc kubenswrapper[4813]: I1201 08:55:01.063292 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-grnss" event={"ID":"5a12a2ee-1fa9-4e27-b8b5-0b1c1f721c5b","Type":"ContainerDied","Data":"96227608cd663c5dc1fe44bd6cac947107c43e4749e4a97e836e0e805d33960c"} Dec 01 08:55:01 crc kubenswrapper[4813]: I1201 08:55:01.063463 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-grnss" event={"ID":"5a12a2ee-1fa9-4e27-b8b5-0b1c1f721c5b","Type":"ContainerStarted","Data":"2867de8145a3863ccb0a345d7e772141359217d8621d36a9866643e887338d1b"} Dec 01 08:55:01 crc kubenswrapper[4813]: I1201 08:55:01.069708 4813 generic.go:334] "Generic (PLEG): container finished" podID="e1cf13b6-094e-4524-82e7-f165e3b24767" containerID="3087b1b199743282b6c6e5936792080167b630ca83056686c61c81117d0593c0" exitCode=0 Dec 01 08:55:01 crc kubenswrapper[4813]: I1201 08:55:01.069778 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5908vlmf" event={"ID":"e1cf13b6-094e-4524-82e7-f165e3b24767","Type":"ContainerDied","Data":"3087b1b199743282b6c6e5936792080167b630ca83056686c61c81117d0593c0"} Dec 01 08:55:01 crc kubenswrapper[4813]: I1201 08:55:01.069889 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5908vlmf" event={"ID":"e1cf13b6-094e-4524-82e7-f165e3b24767","Type":"ContainerStarted","Data":"7e8e14deee73e7fa940053ccd7cac77d3b855c88dfd6a7e5470e76c8b70ae340"} Dec 01 08:55:02 crc kubenswrapper[4813]: I1201 08:55:02.571364 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-d8vtz"] Dec 01 08:55:02 crc kubenswrapper[4813]: I1201 08:55:02.573451 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-d8vtz" Dec 01 08:55:02 crc kubenswrapper[4813]: I1201 08:55:02.588294 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-d8vtz"] Dec 01 08:55:02 crc kubenswrapper[4813]: I1201 08:55:02.668065 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b3995a58-451f-458c-ba0c-ed66510679bf-catalog-content\") pod \"certified-operators-d8vtz\" (UID: \"b3995a58-451f-458c-ba0c-ed66510679bf\") " pod="openshift-marketplace/certified-operators-d8vtz" Dec 01 08:55:02 crc kubenswrapper[4813]: I1201 08:55:02.668182 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-952nc\" (UniqueName: \"kubernetes.io/projected/b3995a58-451f-458c-ba0c-ed66510679bf-kube-api-access-952nc\") pod \"certified-operators-d8vtz\" (UID: \"b3995a58-451f-458c-ba0c-ed66510679bf\") " pod="openshift-marketplace/certified-operators-d8vtz" Dec 01 08:55:02 crc kubenswrapper[4813]: I1201 08:55:02.668281 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b3995a58-451f-458c-ba0c-ed66510679bf-utilities\") pod \"certified-operators-d8vtz\" (UID: \"b3995a58-451f-458c-ba0c-ed66510679bf\") " pod="openshift-marketplace/certified-operators-d8vtz" Dec 01 08:55:02 crc kubenswrapper[4813]: I1201 08:55:02.769484 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b3995a58-451f-458c-ba0c-ed66510679bf-catalog-content\") pod \"certified-operators-d8vtz\" (UID: \"b3995a58-451f-458c-ba0c-ed66510679bf\") " pod="openshift-marketplace/certified-operators-d8vtz" Dec 01 08:55:02 crc kubenswrapper[4813]: I1201 08:55:02.769580 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-952nc\" (UniqueName: \"kubernetes.io/projected/b3995a58-451f-458c-ba0c-ed66510679bf-kube-api-access-952nc\") pod \"certified-operators-d8vtz\" (UID: \"b3995a58-451f-458c-ba0c-ed66510679bf\") " pod="openshift-marketplace/certified-operators-d8vtz" Dec 01 08:55:02 crc kubenswrapper[4813]: I1201 08:55:02.769611 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b3995a58-451f-458c-ba0c-ed66510679bf-utilities\") pod \"certified-operators-d8vtz\" (UID: \"b3995a58-451f-458c-ba0c-ed66510679bf\") " pod="openshift-marketplace/certified-operators-d8vtz" Dec 01 08:55:02 crc kubenswrapper[4813]: I1201 08:55:02.769914 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b3995a58-451f-458c-ba0c-ed66510679bf-catalog-content\") pod \"certified-operators-d8vtz\" (UID: \"b3995a58-451f-458c-ba0c-ed66510679bf\") " pod="openshift-marketplace/certified-operators-d8vtz" Dec 01 08:55:02 crc kubenswrapper[4813]: I1201 08:55:02.770178 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b3995a58-451f-458c-ba0c-ed66510679bf-utilities\") pod \"certified-operators-d8vtz\" (UID: \"b3995a58-451f-458c-ba0c-ed66510679bf\") " pod="openshift-marketplace/certified-operators-d8vtz" Dec 01 08:55:02 crc kubenswrapper[4813]: I1201 08:55:02.792812 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-952nc\" (UniqueName: \"kubernetes.io/projected/b3995a58-451f-458c-ba0c-ed66510679bf-kube-api-access-952nc\") pod \"certified-operators-d8vtz\" (UID: \"b3995a58-451f-458c-ba0c-ed66510679bf\") " pod="openshift-marketplace/certified-operators-d8vtz" Dec 01 08:55:02 crc kubenswrapper[4813]: I1201 08:55:02.902008 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-d8vtz" Dec 01 08:55:03 crc kubenswrapper[4813]: I1201 08:55:03.090863 4813 generic.go:334] "Generic (PLEG): container finished" podID="e1cf13b6-094e-4524-82e7-f165e3b24767" containerID="ea57ebe323034f474f161dd5537a994724462fbe4ec60285cf2934c80c409ec0" exitCode=0 Dec 01 08:55:03 crc kubenswrapper[4813]: I1201 08:55:03.090928 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5908vlmf" event={"ID":"e1cf13b6-094e-4524-82e7-f165e3b24767","Type":"ContainerDied","Data":"ea57ebe323034f474f161dd5537a994724462fbe4ec60285cf2934c80c409ec0"} Dec 01 08:55:03 crc kubenswrapper[4813]: I1201 08:55:03.099340 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-grnss" event={"ID":"5a12a2ee-1fa9-4e27-b8b5-0b1c1f721c5b","Type":"ContainerStarted","Data":"fe936d6ad53b8083d4287dd8a478c15285de5e87babe01cdc5bbcbd8ab47a8cd"} Dec 01 08:55:03 crc kubenswrapper[4813]: I1201 08:55:03.434742 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-d8vtz"] Dec 01 08:55:04 crc kubenswrapper[4813]: I1201 08:55:04.112614 4813 generic.go:334] "Generic (PLEG): container finished" podID="e1cf13b6-094e-4524-82e7-f165e3b24767" containerID="06a6da92472edd0d069f498ec01bb66b4764863b11ad60a1218b0c6576ea89de" exitCode=0 Dec 01 08:55:04 crc kubenswrapper[4813]: I1201 08:55:04.112668 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5908vlmf" event={"ID":"e1cf13b6-094e-4524-82e7-f165e3b24767","Type":"ContainerDied","Data":"06a6da92472edd0d069f498ec01bb66b4764863b11ad60a1218b0c6576ea89de"} Dec 01 08:55:04 crc kubenswrapper[4813]: I1201 08:55:04.117336 4813 generic.go:334] "Generic (PLEG): container finished" podID="5a12a2ee-1fa9-4e27-b8b5-0b1c1f721c5b" containerID="fe936d6ad53b8083d4287dd8a478c15285de5e87babe01cdc5bbcbd8ab47a8cd" exitCode=0 Dec 01 08:55:04 crc kubenswrapper[4813]: I1201 08:55:04.117422 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-grnss" event={"ID":"5a12a2ee-1fa9-4e27-b8b5-0b1c1f721c5b","Type":"ContainerDied","Data":"fe936d6ad53b8083d4287dd8a478c15285de5e87babe01cdc5bbcbd8ab47a8cd"} Dec 01 08:55:04 crc kubenswrapper[4813]: I1201 08:55:04.120748 4813 generic.go:334] "Generic (PLEG): container finished" podID="b3995a58-451f-458c-ba0c-ed66510679bf" containerID="e9de420dd4f5c2cdb6c61d178e5f1dd2b9c6ad8227e7bd5026ebfad0c24f263e" exitCode=0 Dec 01 08:55:04 crc kubenswrapper[4813]: I1201 08:55:04.120816 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-d8vtz" event={"ID":"b3995a58-451f-458c-ba0c-ed66510679bf","Type":"ContainerDied","Data":"e9de420dd4f5c2cdb6c61d178e5f1dd2b9c6ad8227e7bd5026ebfad0c24f263e"} Dec 01 08:55:04 crc kubenswrapper[4813]: I1201 08:55:04.120856 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-d8vtz" event={"ID":"b3995a58-451f-458c-ba0c-ed66510679bf","Type":"ContainerStarted","Data":"8fa6779f65620513cac5a31cc67d6d03d2c99b0b303f2ffbee59154e4f982acc"} Dec 01 08:55:05 crc kubenswrapper[4813]: I1201 08:55:05.421217 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5908vlmf" Dec 01 08:55:05 crc kubenswrapper[4813]: I1201 08:55:05.527108 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/e1cf13b6-094e-4524-82e7-f165e3b24767-util\") pod \"e1cf13b6-094e-4524-82e7-f165e3b24767\" (UID: \"e1cf13b6-094e-4524-82e7-f165e3b24767\") " Dec 01 08:55:05 crc kubenswrapper[4813]: I1201 08:55:05.527348 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nrtqw\" (UniqueName: \"kubernetes.io/projected/e1cf13b6-094e-4524-82e7-f165e3b24767-kube-api-access-nrtqw\") pod \"e1cf13b6-094e-4524-82e7-f165e3b24767\" (UID: \"e1cf13b6-094e-4524-82e7-f165e3b24767\") " Dec 01 08:55:05 crc kubenswrapper[4813]: I1201 08:55:05.527906 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e1cf13b6-094e-4524-82e7-f165e3b24767-util" (OuterVolumeSpecName: "util") pod "e1cf13b6-094e-4524-82e7-f165e3b24767" (UID: "e1cf13b6-094e-4524-82e7-f165e3b24767"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 08:55:05 crc kubenswrapper[4813]: I1201 08:55:05.528127 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/e1cf13b6-094e-4524-82e7-f165e3b24767-bundle\") pod \"e1cf13b6-094e-4524-82e7-f165e3b24767\" (UID: \"e1cf13b6-094e-4524-82e7-f165e3b24767\") " Dec 01 08:55:05 crc kubenswrapper[4813]: I1201 08:55:05.528894 4813 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/e1cf13b6-094e-4524-82e7-f165e3b24767-util\") on node \"crc\" DevicePath \"\"" Dec 01 08:55:05 crc kubenswrapper[4813]: I1201 08:55:05.529181 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e1cf13b6-094e-4524-82e7-f165e3b24767-bundle" (OuterVolumeSpecName: "bundle") pod "e1cf13b6-094e-4524-82e7-f165e3b24767" (UID: "e1cf13b6-094e-4524-82e7-f165e3b24767"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 08:55:05 crc kubenswrapper[4813]: I1201 08:55:05.534388 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e1cf13b6-094e-4524-82e7-f165e3b24767-kube-api-access-nrtqw" (OuterVolumeSpecName: "kube-api-access-nrtqw") pod "e1cf13b6-094e-4524-82e7-f165e3b24767" (UID: "e1cf13b6-094e-4524-82e7-f165e3b24767"). InnerVolumeSpecName "kube-api-access-nrtqw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:55:05 crc kubenswrapper[4813]: I1201 08:55:05.630105 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nrtqw\" (UniqueName: \"kubernetes.io/projected/e1cf13b6-094e-4524-82e7-f165e3b24767-kube-api-access-nrtqw\") on node \"crc\" DevicePath \"\"" Dec 01 08:55:05 crc kubenswrapper[4813]: I1201 08:55:05.630151 4813 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/e1cf13b6-094e-4524-82e7-f165e3b24767-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 08:55:06 crc kubenswrapper[4813]: I1201 08:55:06.139181 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-grnss" event={"ID":"5a12a2ee-1fa9-4e27-b8b5-0b1c1f721c5b","Type":"ContainerStarted","Data":"8f54da59742c98d1ef046d47ca2d3d6ab5ed651541cbda700b84dbbe176420a1"} Dec 01 08:55:06 crc kubenswrapper[4813]: I1201 08:55:06.141114 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-d8vtz" event={"ID":"b3995a58-451f-458c-ba0c-ed66510679bf","Type":"ContainerStarted","Data":"3c9d57a69b19794bfa5cb175ec2e43a99049622e30fcd0c7688a30f89cc31a4f"} Dec 01 08:55:06 crc kubenswrapper[4813]: I1201 08:55:06.143695 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5908vlmf" Dec 01 08:55:06 crc kubenswrapper[4813]: I1201 08:55:06.143689 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5908vlmf" event={"ID":"e1cf13b6-094e-4524-82e7-f165e3b24767","Type":"ContainerDied","Data":"7e8e14deee73e7fa940053ccd7cac77d3b855c88dfd6a7e5470e76c8b70ae340"} Dec 01 08:55:06 crc kubenswrapper[4813]: I1201 08:55:06.143900 4813 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7e8e14deee73e7fa940053ccd7cac77d3b855c88dfd6a7e5470e76c8b70ae340" Dec 01 08:55:06 crc kubenswrapper[4813]: I1201 08:55:06.163198 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-grnss" podStartSLOduration=2.609294006 podStartE2EDuration="7.163149804s" podCreationTimestamp="2025-12-01 08:54:59 +0000 UTC" firstStartedPulling="2025-12-01 08:55:01.064837986 +0000 UTC m=+885.687759572" lastFinishedPulling="2025-12-01 08:55:05.618693784 +0000 UTC m=+890.241615370" observedRunningTime="2025-12-01 08:55:06.162784214 +0000 UTC m=+890.785705830" watchObservedRunningTime="2025-12-01 08:55:06.163149804 +0000 UTC m=+890.786071390" Dec 01 08:55:07 crc kubenswrapper[4813]: I1201 08:55:07.154084 4813 generic.go:334] "Generic (PLEG): container finished" podID="b3995a58-451f-458c-ba0c-ed66510679bf" containerID="3c9d57a69b19794bfa5cb175ec2e43a99049622e30fcd0c7688a30f89cc31a4f" exitCode=0 Dec 01 08:55:07 crc kubenswrapper[4813]: I1201 08:55:07.154215 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-d8vtz" event={"ID":"b3995a58-451f-458c-ba0c-ed66510679bf","Type":"ContainerDied","Data":"3c9d57a69b19794bfa5cb175ec2e43a99049622e30fcd0c7688a30f89cc31a4f"} Dec 01 08:55:07 crc kubenswrapper[4813]: I1201 08:55:07.368396 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-bb4hs"] Dec 01 08:55:07 crc kubenswrapper[4813]: E1201 08:55:07.383018 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e1cf13b6-094e-4524-82e7-f165e3b24767" containerName="extract" Dec 01 08:55:07 crc kubenswrapper[4813]: I1201 08:55:07.383093 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="e1cf13b6-094e-4524-82e7-f165e3b24767" containerName="extract" Dec 01 08:55:07 crc kubenswrapper[4813]: E1201 08:55:07.383134 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e1cf13b6-094e-4524-82e7-f165e3b24767" containerName="util" Dec 01 08:55:07 crc kubenswrapper[4813]: I1201 08:55:07.383151 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="e1cf13b6-094e-4524-82e7-f165e3b24767" containerName="util" Dec 01 08:55:07 crc kubenswrapper[4813]: E1201 08:55:07.383199 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e1cf13b6-094e-4524-82e7-f165e3b24767" containerName="pull" Dec 01 08:55:07 crc kubenswrapper[4813]: I1201 08:55:07.383212 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="e1cf13b6-094e-4524-82e7-f165e3b24767" containerName="pull" Dec 01 08:55:07 crc kubenswrapper[4813]: I1201 08:55:07.383772 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="e1cf13b6-094e-4524-82e7-f165e3b24767" containerName="extract" Dec 01 08:55:07 crc kubenswrapper[4813]: I1201 08:55:07.388217 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-bb4hs"] Dec 01 08:55:07 crc kubenswrapper[4813]: I1201 08:55:07.388345 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bb4hs" Dec 01 08:55:07 crc kubenswrapper[4813]: I1201 08:55:07.557095 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fb7b8647-2f8c-4508-8f11-555d31ad500b-catalog-content\") pod \"redhat-marketplace-bb4hs\" (UID: \"fb7b8647-2f8c-4508-8f11-555d31ad500b\") " pod="openshift-marketplace/redhat-marketplace-bb4hs" Dec 01 08:55:07 crc kubenswrapper[4813]: I1201 08:55:07.557654 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fz5pz\" (UniqueName: \"kubernetes.io/projected/fb7b8647-2f8c-4508-8f11-555d31ad500b-kube-api-access-fz5pz\") pod \"redhat-marketplace-bb4hs\" (UID: \"fb7b8647-2f8c-4508-8f11-555d31ad500b\") " pod="openshift-marketplace/redhat-marketplace-bb4hs" Dec 01 08:55:07 crc kubenswrapper[4813]: I1201 08:55:07.557745 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fb7b8647-2f8c-4508-8f11-555d31ad500b-utilities\") pod \"redhat-marketplace-bb4hs\" (UID: \"fb7b8647-2f8c-4508-8f11-555d31ad500b\") " pod="openshift-marketplace/redhat-marketplace-bb4hs" Dec 01 08:55:07 crc kubenswrapper[4813]: I1201 08:55:07.658565 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fz5pz\" (UniqueName: \"kubernetes.io/projected/fb7b8647-2f8c-4508-8f11-555d31ad500b-kube-api-access-fz5pz\") pod \"redhat-marketplace-bb4hs\" (UID: \"fb7b8647-2f8c-4508-8f11-555d31ad500b\") " pod="openshift-marketplace/redhat-marketplace-bb4hs" Dec 01 08:55:07 crc kubenswrapper[4813]: I1201 08:55:07.658653 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fb7b8647-2f8c-4508-8f11-555d31ad500b-utilities\") pod \"redhat-marketplace-bb4hs\" (UID: \"fb7b8647-2f8c-4508-8f11-555d31ad500b\") " pod="openshift-marketplace/redhat-marketplace-bb4hs" Dec 01 08:55:07 crc kubenswrapper[4813]: I1201 08:55:07.658692 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fb7b8647-2f8c-4508-8f11-555d31ad500b-catalog-content\") pod \"redhat-marketplace-bb4hs\" (UID: \"fb7b8647-2f8c-4508-8f11-555d31ad500b\") " pod="openshift-marketplace/redhat-marketplace-bb4hs" Dec 01 08:55:07 crc kubenswrapper[4813]: I1201 08:55:07.659869 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fb7b8647-2f8c-4508-8f11-555d31ad500b-catalog-content\") pod \"redhat-marketplace-bb4hs\" (UID: \"fb7b8647-2f8c-4508-8f11-555d31ad500b\") " pod="openshift-marketplace/redhat-marketplace-bb4hs" Dec 01 08:55:07 crc kubenswrapper[4813]: I1201 08:55:07.659874 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fb7b8647-2f8c-4508-8f11-555d31ad500b-utilities\") pod \"redhat-marketplace-bb4hs\" (UID: \"fb7b8647-2f8c-4508-8f11-555d31ad500b\") " pod="openshift-marketplace/redhat-marketplace-bb4hs" Dec 01 08:55:07 crc kubenswrapper[4813]: I1201 08:55:07.682256 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fz5pz\" (UniqueName: \"kubernetes.io/projected/fb7b8647-2f8c-4508-8f11-555d31ad500b-kube-api-access-fz5pz\") pod \"redhat-marketplace-bb4hs\" (UID: \"fb7b8647-2f8c-4508-8f11-555d31ad500b\") " pod="openshift-marketplace/redhat-marketplace-bb4hs" Dec 01 08:55:07 crc kubenswrapper[4813]: I1201 08:55:07.713420 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bb4hs" Dec 01 08:55:07 crc kubenswrapper[4813]: I1201 08:55:07.765722 4813 prober.go:107] "Probe failed" probeType="Readiness" pod="glance-kuttl-tests/openstack-galera-2" podUID="73988763-722d-4652-8fac-053ba8217547" containerName="galera" probeResult="failure" output=< Dec 01 08:55:07 crc kubenswrapper[4813]: wsrep_local_state_comment (Donor/Desynced) differs from Synced Dec 01 08:55:07 crc kubenswrapper[4813]: > Dec 01 08:55:08 crc kubenswrapper[4813]: I1201 08:55:08.125061 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-bb4hs"] Dec 01 08:55:08 crc kubenswrapper[4813]: W1201 08:55:08.130240 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfb7b8647_2f8c_4508_8f11_555d31ad500b.slice/crio-638479af90fa282f6631e725e113ccb3bbbcae729bf099409263f82099afeb56 WatchSource:0}: Error finding container 638479af90fa282f6631e725e113ccb3bbbcae729bf099409263f82099afeb56: Status 404 returned error can't find the container with id 638479af90fa282f6631e725e113ccb3bbbcae729bf099409263f82099afeb56 Dec 01 08:55:08 crc kubenswrapper[4813]: I1201 08:55:08.161118 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bb4hs" event={"ID":"fb7b8647-2f8c-4508-8f11-555d31ad500b","Type":"ContainerStarted","Data":"638479af90fa282f6631e725e113ccb3bbbcae729bf099409263f82099afeb56"} Dec 01 08:55:09 crc kubenswrapper[4813]: I1201 08:55:09.170239 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-d8vtz" event={"ID":"b3995a58-451f-458c-ba0c-ed66510679bf","Type":"ContainerStarted","Data":"6326e13817cba94738109a4820644c6095b3c303f37fa51b4e630cafaa142b83"} Dec 01 08:55:09 crc kubenswrapper[4813]: I1201 08:55:09.174258 4813 generic.go:334] "Generic (PLEG): container finished" podID="fb7b8647-2f8c-4508-8f11-555d31ad500b" containerID="562c36cfae2c8ec9e9d4f245aa9f3c8ebb1dabe78b07e01ef0f3c3f90577ecc6" exitCode=0 Dec 01 08:55:09 crc kubenswrapper[4813]: I1201 08:55:09.174311 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bb4hs" event={"ID":"fb7b8647-2f8c-4508-8f11-555d31ad500b","Type":"ContainerDied","Data":"562c36cfae2c8ec9e9d4f245aa9f3c8ebb1dabe78b07e01ef0f3c3f90577ecc6"} Dec 01 08:55:09 crc kubenswrapper[4813]: I1201 08:55:09.190196 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-d8vtz" podStartSLOduration=2.748542266 podStartE2EDuration="7.19017776s" podCreationTimestamp="2025-12-01 08:55:02 +0000 UTC" firstStartedPulling="2025-12-01 08:55:04.133655138 +0000 UTC m=+888.756576734" lastFinishedPulling="2025-12-01 08:55:08.575290642 +0000 UTC m=+893.198212228" observedRunningTime="2025-12-01 08:55:09.187796654 +0000 UTC m=+893.810718240" watchObservedRunningTime="2025-12-01 08:55:09.19017776 +0000 UTC m=+893.813099346" Dec 01 08:55:09 crc kubenswrapper[4813]: I1201 08:55:09.708314 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-grnss" Dec 01 08:55:09 crc kubenswrapper[4813]: I1201 08:55:09.708621 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-grnss" Dec 01 08:55:10 crc kubenswrapper[4813]: I1201 08:55:10.748504 4813 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-grnss" podUID="5a12a2ee-1fa9-4e27-b8b5-0b1c1f721c5b" containerName="registry-server" probeResult="failure" output=< Dec 01 08:55:10 crc kubenswrapper[4813]: timeout: failed to connect service ":50051" within 1s Dec 01 08:55:10 crc kubenswrapper[4813]: > Dec 01 08:55:11 crc kubenswrapper[4813]: I1201 08:55:11.188879 4813 generic.go:334] "Generic (PLEG): container finished" podID="fb7b8647-2f8c-4508-8f11-555d31ad500b" containerID="eb03b24cee58d54df865ab101f911285a410fe4bdefd75eb9566a679d247f114" exitCode=0 Dec 01 08:55:11 crc kubenswrapper[4813]: I1201 08:55:11.189092 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bb4hs" event={"ID":"fb7b8647-2f8c-4508-8f11-555d31ad500b","Type":"ContainerDied","Data":"eb03b24cee58d54df865ab101f911285a410fe4bdefd75eb9566a679d247f114"} Dec 01 08:55:12 crc kubenswrapper[4813]: I1201 08:55:12.196813 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bb4hs" event={"ID":"fb7b8647-2f8c-4508-8f11-555d31ad500b","Type":"ContainerStarted","Data":"c33a17bde30aa1c2a6cba8577969415d5485427d650335c0152de62d720b66cb"} Dec 01 08:55:12 crc kubenswrapper[4813]: I1201 08:55:12.216464 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-bb4hs" podStartSLOduration=2.464313302 podStartE2EDuration="5.216442115s" podCreationTimestamp="2025-12-01 08:55:07 +0000 UTC" firstStartedPulling="2025-12-01 08:55:09.175857004 +0000 UTC m=+893.798778590" lastFinishedPulling="2025-12-01 08:55:11.927985807 +0000 UTC m=+896.550907403" observedRunningTime="2025-12-01 08:55:12.215411887 +0000 UTC m=+896.838333483" watchObservedRunningTime="2025-12-01 08:55:12.216442115 +0000 UTC m=+896.839363701" Dec 01 08:55:12 crc kubenswrapper[4813]: I1201 08:55:12.902556 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-d8vtz" Dec 01 08:55:12 crc kubenswrapper[4813]: I1201 08:55:12.902907 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-d8vtz" Dec 01 08:55:12 crc kubenswrapper[4813]: I1201 08:55:12.969891 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-d8vtz" Dec 01 08:55:13 crc kubenswrapper[4813]: I1201 08:55:13.292945 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-d8vtz" Dec 01 08:55:13 crc kubenswrapper[4813]: I1201 08:55:13.386362 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/openstack-galera-0" Dec 01 08:55:13 crc kubenswrapper[4813]: I1201 08:55:13.473091 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/openstack-galera-0" Dec 01 08:55:14 crc kubenswrapper[4813]: I1201 08:55:14.899077 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-779fc9694b-c7nbx"] Dec 01 08:55:14 crc kubenswrapper[4813]: I1201 08:55:14.900576 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-c7nbx" Dec 01 08:55:14 crc kubenswrapper[4813]: I1201 08:55:14.903458 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-dockercfg-g29bm" Dec 01 08:55:14 crc kubenswrapper[4813]: I1201 08:55:14.925079 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-779fc9694b-c7nbx"] Dec 01 08:55:15 crc kubenswrapper[4813]: I1201 08:55:15.033700 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w994g\" (UniqueName: \"kubernetes.io/projected/3820379b-32dc-46ea-a899-0990fadcab5b-kube-api-access-w994g\") pod \"rabbitmq-cluster-operator-779fc9694b-c7nbx\" (UID: \"3820379b-32dc-46ea-a899-0990fadcab5b\") " pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-c7nbx" Dec 01 08:55:15 crc kubenswrapper[4813]: I1201 08:55:15.135238 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w994g\" (UniqueName: \"kubernetes.io/projected/3820379b-32dc-46ea-a899-0990fadcab5b-kube-api-access-w994g\") pod \"rabbitmq-cluster-operator-779fc9694b-c7nbx\" (UID: \"3820379b-32dc-46ea-a899-0990fadcab5b\") " pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-c7nbx" Dec 01 08:55:15 crc kubenswrapper[4813]: I1201 08:55:15.159022 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w994g\" (UniqueName: \"kubernetes.io/projected/3820379b-32dc-46ea-a899-0990fadcab5b-kube-api-access-w994g\") pod \"rabbitmq-cluster-operator-779fc9694b-c7nbx\" (UID: \"3820379b-32dc-46ea-a899-0990fadcab5b\") " pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-c7nbx" Dec 01 08:55:15 crc kubenswrapper[4813]: I1201 08:55:15.333464 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-c7nbx" Dec 01 08:55:15 crc kubenswrapper[4813]: I1201 08:55:15.769505 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-779fc9694b-c7nbx"] Dec 01 08:55:15 crc kubenswrapper[4813]: I1201 08:55:15.965340 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-htzpm"] Dec 01 08:55:15 crc kubenswrapper[4813]: I1201 08:55:15.967273 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-htzpm" Dec 01 08:55:15 crc kubenswrapper[4813]: I1201 08:55:15.976574 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-htzpm"] Dec 01 08:55:16 crc kubenswrapper[4813]: I1201 08:55:16.147265 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2a8a0c74-0097-498b-9602-2fa4c73873b3-catalog-content\") pod \"community-operators-htzpm\" (UID: \"2a8a0c74-0097-498b-9602-2fa4c73873b3\") " pod="openshift-marketplace/community-operators-htzpm" Dec 01 08:55:16 crc kubenswrapper[4813]: I1201 08:55:16.147363 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2a8a0c74-0097-498b-9602-2fa4c73873b3-utilities\") pod \"community-operators-htzpm\" (UID: \"2a8a0c74-0097-498b-9602-2fa4c73873b3\") " pod="openshift-marketplace/community-operators-htzpm" Dec 01 08:55:16 crc kubenswrapper[4813]: I1201 08:55:16.147422 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v8qp5\" (UniqueName: \"kubernetes.io/projected/2a8a0c74-0097-498b-9602-2fa4c73873b3-kube-api-access-v8qp5\") pod \"community-operators-htzpm\" (UID: \"2a8a0c74-0097-498b-9602-2fa4c73873b3\") " pod="openshift-marketplace/community-operators-htzpm" Dec 01 08:55:16 crc kubenswrapper[4813]: I1201 08:55:16.246012 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-c7nbx" event={"ID":"3820379b-32dc-46ea-a899-0990fadcab5b","Type":"ContainerStarted","Data":"b79485e00949fafaeab9ccc86e5e1c800b67c9120cdb041d64a117d83303402c"} Dec 01 08:55:16 crc kubenswrapper[4813]: I1201 08:55:16.248896 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v8qp5\" (UniqueName: \"kubernetes.io/projected/2a8a0c74-0097-498b-9602-2fa4c73873b3-kube-api-access-v8qp5\") pod \"community-operators-htzpm\" (UID: \"2a8a0c74-0097-498b-9602-2fa4c73873b3\") " pod="openshift-marketplace/community-operators-htzpm" Dec 01 08:55:16 crc kubenswrapper[4813]: I1201 08:55:16.249000 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2a8a0c74-0097-498b-9602-2fa4c73873b3-catalog-content\") pod \"community-operators-htzpm\" (UID: \"2a8a0c74-0097-498b-9602-2fa4c73873b3\") " pod="openshift-marketplace/community-operators-htzpm" Dec 01 08:55:16 crc kubenswrapper[4813]: I1201 08:55:16.249079 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2a8a0c74-0097-498b-9602-2fa4c73873b3-utilities\") pod \"community-operators-htzpm\" (UID: \"2a8a0c74-0097-498b-9602-2fa4c73873b3\") " pod="openshift-marketplace/community-operators-htzpm" Dec 01 08:55:16 crc kubenswrapper[4813]: I1201 08:55:16.249505 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2a8a0c74-0097-498b-9602-2fa4c73873b3-utilities\") pod \"community-operators-htzpm\" (UID: \"2a8a0c74-0097-498b-9602-2fa4c73873b3\") " pod="openshift-marketplace/community-operators-htzpm" Dec 01 08:55:16 crc kubenswrapper[4813]: I1201 08:55:16.249581 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2a8a0c74-0097-498b-9602-2fa4c73873b3-catalog-content\") pod \"community-operators-htzpm\" (UID: \"2a8a0c74-0097-498b-9602-2fa4c73873b3\") " pod="openshift-marketplace/community-operators-htzpm" Dec 01 08:55:16 crc kubenswrapper[4813]: I1201 08:55:16.274925 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v8qp5\" (UniqueName: \"kubernetes.io/projected/2a8a0c74-0097-498b-9602-2fa4c73873b3-kube-api-access-v8qp5\") pod \"community-operators-htzpm\" (UID: \"2a8a0c74-0097-498b-9602-2fa4c73873b3\") " pod="openshift-marketplace/community-operators-htzpm" Dec 01 08:55:16 crc kubenswrapper[4813]: I1201 08:55:16.293460 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-htzpm" Dec 01 08:55:16 crc kubenswrapper[4813]: I1201 08:55:16.597442 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/openstack-galera-1" Dec 01 08:55:16 crc kubenswrapper[4813]: I1201 08:55:16.700002 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/openstack-galera-1" Dec 01 08:55:16 crc kubenswrapper[4813]: I1201 08:55:16.817197 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-htzpm"] Dec 01 08:55:17 crc kubenswrapper[4813]: I1201 08:55:17.222049 4813 patch_prober.go:28] interesting pod/machine-config-daemon-9wn4q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 08:55:17 crc kubenswrapper[4813]: I1201 08:55:17.222394 4813 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9wn4q" podUID="3d423402-e08f-42e9-b60c-6eb9bf2f6d48" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 08:55:17 crc kubenswrapper[4813]: I1201 08:55:17.258669 4813 generic.go:334] "Generic (PLEG): container finished" podID="2a8a0c74-0097-498b-9602-2fa4c73873b3" containerID="ddc16dc13accc79cff7a425e620a8d4182519a67dc43b0204b381636dd83264d" exitCode=0 Dec 01 08:55:17 crc kubenswrapper[4813]: I1201 08:55:17.258839 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-htzpm" event={"ID":"2a8a0c74-0097-498b-9602-2fa4c73873b3","Type":"ContainerDied","Data":"ddc16dc13accc79cff7a425e620a8d4182519a67dc43b0204b381636dd83264d"} Dec 01 08:55:17 crc kubenswrapper[4813]: I1201 08:55:17.258899 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-htzpm" event={"ID":"2a8a0c74-0097-498b-9602-2fa4c73873b3","Type":"ContainerStarted","Data":"ee9738c25d3959b113e33f8357b6179c88ccb20f0b8ebc619f06470f1e8c73be"} Dec 01 08:55:17 crc kubenswrapper[4813]: I1201 08:55:17.713578 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-bb4hs" Dec 01 08:55:17 crc kubenswrapper[4813]: I1201 08:55:17.713622 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-bb4hs" Dec 01 08:55:17 crc kubenswrapper[4813]: I1201 08:55:17.773059 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-bb4hs" Dec 01 08:55:18 crc kubenswrapper[4813]: I1201 08:55:18.331268 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-bb4hs" Dec 01 08:55:19 crc kubenswrapper[4813]: I1201 08:55:19.763821 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-d8vtz"] Dec 01 08:55:19 crc kubenswrapper[4813]: I1201 08:55:19.764386 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-d8vtz" podUID="b3995a58-451f-458c-ba0c-ed66510679bf" containerName="registry-server" containerID="cri-o://6326e13817cba94738109a4820644c6095b3c303f37fa51b4e630cafaa142b83" gracePeriod=2 Dec 01 08:55:19 crc kubenswrapper[4813]: I1201 08:55:19.805762 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-grnss" Dec 01 08:55:19 crc kubenswrapper[4813]: I1201 08:55:19.856297 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-grnss" Dec 01 08:55:19 crc kubenswrapper[4813]: E1201 08:55:19.895427 4813 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb3995a58_451f_458c_ba0c_ed66510679bf.slice/crio-6326e13817cba94738109a4820644c6095b3c303f37fa51b4e630cafaa142b83.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2a8a0c74_0097_498b_9602_2fa4c73873b3.slice/crio-conmon-ab2ee4fe15992d4985bd93913356cf8395a7f9cf09ee8b544236dfb7aac2f901.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2a8a0c74_0097_498b_9602_2fa4c73873b3.slice/crio-ab2ee4fe15992d4985bd93913356cf8395a7f9cf09ee8b544236dfb7aac2f901.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb3995a58_451f_458c_ba0c_ed66510679bf.slice/crio-conmon-6326e13817cba94738109a4820644c6095b3c303f37fa51b4e630cafaa142b83.scope\": RecentStats: unable to find data in memory cache]" Dec 01 08:55:20 crc kubenswrapper[4813]: I1201 08:55:20.158566 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-d8vtz" Dec 01 08:55:20 crc kubenswrapper[4813]: I1201 08:55:20.287788 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b3995a58-451f-458c-ba0c-ed66510679bf-catalog-content\") pod \"b3995a58-451f-458c-ba0c-ed66510679bf\" (UID: \"b3995a58-451f-458c-ba0c-ed66510679bf\") " Dec 01 08:55:20 crc kubenswrapper[4813]: I1201 08:55:20.287896 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-952nc\" (UniqueName: \"kubernetes.io/projected/b3995a58-451f-458c-ba0c-ed66510679bf-kube-api-access-952nc\") pod \"b3995a58-451f-458c-ba0c-ed66510679bf\" (UID: \"b3995a58-451f-458c-ba0c-ed66510679bf\") " Dec 01 08:55:20 crc kubenswrapper[4813]: I1201 08:55:20.288061 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b3995a58-451f-458c-ba0c-ed66510679bf-utilities\") pod \"b3995a58-451f-458c-ba0c-ed66510679bf\" (UID: \"b3995a58-451f-458c-ba0c-ed66510679bf\") " Dec 01 08:55:20 crc kubenswrapper[4813]: I1201 08:55:20.289853 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b3995a58-451f-458c-ba0c-ed66510679bf-utilities" (OuterVolumeSpecName: "utilities") pod "b3995a58-451f-458c-ba0c-ed66510679bf" (UID: "b3995a58-451f-458c-ba0c-ed66510679bf"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 08:55:20 crc kubenswrapper[4813]: I1201 08:55:20.290614 4813 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b3995a58-451f-458c-ba0c-ed66510679bf-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 08:55:20 crc kubenswrapper[4813]: I1201 08:55:20.293190 4813 generic.go:334] "Generic (PLEG): container finished" podID="b3995a58-451f-458c-ba0c-ed66510679bf" containerID="6326e13817cba94738109a4820644c6095b3c303f37fa51b4e630cafaa142b83" exitCode=0 Dec 01 08:55:20 crc kubenswrapper[4813]: I1201 08:55:20.293565 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-d8vtz" event={"ID":"b3995a58-451f-458c-ba0c-ed66510679bf","Type":"ContainerDied","Data":"6326e13817cba94738109a4820644c6095b3c303f37fa51b4e630cafaa142b83"} Dec 01 08:55:20 crc kubenswrapper[4813]: I1201 08:55:20.293611 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-d8vtz" event={"ID":"b3995a58-451f-458c-ba0c-ed66510679bf","Type":"ContainerDied","Data":"8fa6779f65620513cac5a31cc67d6d03d2c99b0b303f2ffbee59154e4f982acc"} Dec 01 08:55:20 crc kubenswrapper[4813]: I1201 08:55:20.293697 4813 scope.go:117] "RemoveContainer" containerID="6326e13817cba94738109a4820644c6095b3c303f37fa51b4e630cafaa142b83" Dec 01 08:55:20 crc kubenswrapper[4813]: I1201 08:55:20.293934 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-d8vtz" Dec 01 08:55:20 crc kubenswrapper[4813]: I1201 08:55:20.298783 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b3995a58-451f-458c-ba0c-ed66510679bf-kube-api-access-952nc" (OuterVolumeSpecName: "kube-api-access-952nc") pod "b3995a58-451f-458c-ba0c-ed66510679bf" (UID: "b3995a58-451f-458c-ba0c-ed66510679bf"). InnerVolumeSpecName "kube-api-access-952nc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:55:20 crc kubenswrapper[4813]: I1201 08:55:20.301472 4813 generic.go:334] "Generic (PLEG): container finished" podID="2a8a0c74-0097-498b-9602-2fa4c73873b3" containerID="ab2ee4fe15992d4985bd93913356cf8395a7f9cf09ee8b544236dfb7aac2f901" exitCode=0 Dec 01 08:55:20 crc kubenswrapper[4813]: I1201 08:55:20.301898 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-htzpm" event={"ID":"2a8a0c74-0097-498b-9602-2fa4c73873b3","Type":"ContainerDied","Data":"ab2ee4fe15992d4985bd93913356cf8395a7f9cf09ee8b544236dfb7aac2f901"} Dec 01 08:55:20 crc kubenswrapper[4813]: I1201 08:55:20.307170 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-c7nbx" event={"ID":"3820379b-32dc-46ea-a899-0990fadcab5b","Type":"ContainerStarted","Data":"da30da264478e8e370b386bfa87faaa9724dd0a68166bca2ac7eb0b45c18fd0a"} Dec 01 08:55:20 crc kubenswrapper[4813]: I1201 08:55:20.361348 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-c7nbx" podStartSLOduration=2.854811661 podStartE2EDuration="6.361261858s" podCreationTimestamp="2025-12-01 08:55:14 +0000 UTC" firstStartedPulling="2025-12-01 08:55:15.783709454 +0000 UTC m=+900.406631080" lastFinishedPulling="2025-12-01 08:55:19.290159651 +0000 UTC m=+903.913081277" observedRunningTime="2025-12-01 08:55:20.351740725 +0000 UTC m=+904.974662351" watchObservedRunningTime="2025-12-01 08:55:20.361261858 +0000 UTC m=+904.984183484" Dec 01 08:55:20 crc kubenswrapper[4813]: I1201 08:55:20.366173 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b3995a58-451f-458c-ba0c-ed66510679bf-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b3995a58-451f-458c-ba0c-ed66510679bf" (UID: "b3995a58-451f-458c-ba0c-ed66510679bf"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 08:55:20 crc kubenswrapper[4813]: I1201 08:55:20.373352 4813 scope.go:117] "RemoveContainer" containerID="3c9d57a69b19794bfa5cb175ec2e43a99049622e30fcd0c7688a30f89cc31a4f" Dec 01 08:55:20 crc kubenswrapper[4813]: I1201 08:55:20.393766 4813 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b3995a58-451f-458c-ba0c-ed66510679bf-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 08:55:20 crc kubenswrapper[4813]: I1201 08:55:20.393838 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-952nc\" (UniqueName: \"kubernetes.io/projected/b3995a58-451f-458c-ba0c-ed66510679bf-kube-api-access-952nc\") on node \"crc\" DevicePath \"\"" Dec 01 08:55:20 crc kubenswrapper[4813]: I1201 08:55:20.405594 4813 scope.go:117] "RemoveContainer" containerID="e9de420dd4f5c2cdb6c61d178e5f1dd2b9c6ad8227e7bd5026ebfad0c24f263e" Dec 01 08:55:20 crc kubenswrapper[4813]: I1201 08:55:20.437578 4813 scope.go:117] "RemoveContainer" containerID="6326e13817cba94738109a4820644c6095b3c303f37fa51b4e630cafaa142b83" Dec 01 08:55:20 crc kubenswrapper[4813]: E1201 08:55:20.438212 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6326e13817cba94738109a4820644c6095b3c303f37fa51b4e630cafaa142b83\": container with ID starting with 6326e13817cba94738109a4820644c6095b3c303f37fa51b4e630cafaa142b83 not found: ID does not exist" containerID="6326e13817cba94738109a4820644c6095b3c303f37fa51b4e630cafaa142b83" Dec 01 08:55:20 crc kubenswrapper[4813]: I1201 08:55:20.438261 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6326e13817cba94738109a4820644c6095b3c303f37fa51b4e630cafaa142b83"} err="failed to get container status \"6326e13817cba94738109a4820644c6095b3c303f37fa51b4e630cafaa142b83\": rpc error: code = NotFound desc = could not find container \"6326e13817cba94738109a4820644c6095b3c303f37fa51b4e630cafaa142b83\": container with ID starting with 6326e13817cba94738109a4820644c6095b3c303f37fa51b4e630cafaa142b83 not found: ID does not exist" Dec 01 08:55:20 crc kubenswrapper[4813]: I1201 08:55:20.438287 4813 scope.go:117] "RemoveContainer" containerID="3c9d57a69b19794bfa5cb175ec2e43a99049622e30fcd0c7688a30f89cc31a4f" Dec 01 08:55:20 crc kubenswrapper[4813]: E1201 08:55:20.439188 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3c9d57a69b19794bfa5cb175ec2e43a99049622e30fcd0c7688a30f89cc31a4f\": container with ID starting with 3c9d57a69b19794bfa5cb175ec2e43a99049622e30fcd0c7688a30f89cc31a4f not found: ID does not exist" containerID="3c9d57a69b19794bfa5cb175ec2e43a99049622e30fcd0c7688a30f89cc31a4f" Dec 01 08:55:20 crc kubenswrapper[4813]: I1201 08:55:20.439218 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3c9d57a69b19794bfa5cb175ec2e43a99049622e30fcd0c7688a30f89cc31a4f"} err="failed to get container status \"3c9d57a69b19794bfa5cb175ec2e43a99049622e30fcd0c7688a30f89cc31a4f\": rpc error: code = NotFound desc = could not find container \"3c9d57a69b19794bfa5cb175ec2e43a99049622e30fcd0c7688a30f89cc31a4f\": container with ID starting with 3c9d57a69b19794bfa5cb175ec2e43a99049622e30fcd0c7688a30f89cc31a4f not found: ID does not exist" Dec 01 08:55:20 crc kubenswrapper[4813]: I1201 08:55:20.439239 4813 scope.go:117] "RemoveContainer" containerID="e9de420dd4f5c2cdb6c61d178e5f1dd2b9c6ad8227e7bd5026ebfad0c24f263e" Dec 01 08:55:20 crc kubenswrapper[4813]: E1201 08:55:20.440491 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e9de420dd4f5c2cdb6c61d178e5f1dd2b9c6ad8227e7bd5026ebfad0c24f263e\": container with ID starting with e9de420dd4f5c2cdb6c61d178e5f1dd2b9c6ad8227e7bd5026ebfad0c24f263e not found: ID does not exist" containerID="e9de420dd4f5c2cdb6c61d178e5f1dd2b9c6ad8227e7bd5026ebfad0c24f263e" Dec 01 08:55:20 crc kubenswrapper[4813]: I1201 08:55:20.440533 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e9de420dd4f5c2cdb6c61d178e5f1dd2b9c6ad8227e7bd5026ebfad0c24f263e"} err="failed to get container status \"e9de420dd4f5c2cdb6c61d178e5f1dd2b9c6ad8227e7bd5026ebfad0c24f263e\": rpc error: code = NotFound desc = could not find container \"e9de420dd4f5c2cdb6c61d178e5f1dd2b9c6ad8227e7bd5026ebfad0c24f263e\": container with ID starting with e9de420dd4f5c2cdb6c61d178e5f1dd2b9c6ad8227e7bd5026ebfad0c24f263e not found: ID does not exist" Dec 01 08:55:20 crc kubenswrapper[4813]: I1201 08:55:20.623950 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-d8vtz"] Dec 01 08:55:20 crc kubenswrapper[4813]: I1201 08:55:20.649483 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-d8vtz"] Dec 01 08:55:20 crc kubenswrapper[4813]: I1201 08:55:20.761144 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-bb4hs"] Dec 01 08:55:20 crc kubenswrapper[4813]: I1201 08:55:20.761864 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-bb4hs" podUID="fb7b8647-2f8c-4508-8f11-555d31ad500b" containerName="registry-server" containerID="cri-o://c33a17bde30aa1c2a6cba8577969415d5485427d650335c0152de62d720b66cb" gracePeriod=2 Dec 01 08:55:21 crc kubenswrapper[4813]: I1201 08:55:21.235237 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bb4hs" Dec 01 08:55:21 crc kubenswrapper[4813]: I1201 08:55:21.314522 4813 generic.go:334] "Generic (PLEG): container finished" podID="fb7b8647-2f8c-4508-8f11-555d31ad500b" containerID="c33a17bde30aa1c2a6cba8577969415d5485427d650335c0152de62d720b66cb" exitCode=0 Dec 01 08:55:21 crc kubenswrapper[4813]: I1201 08:55:21.314589 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bb4hs" Dec 01 08:55:21 crc kubenswrapper[4813]: I1201 08:55:21.314590 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bb4hs" event={"ID":"fb7b8647-2f8c-4508-8f11-555d31ad500b","Type":"ContainerDied","Data":"c33a17bde30aa1c2a6cba8577969415d5485427d650335c0152de62d720b66cb"} Dec 01 08:55:21 crc kubenswrapper[4813]: I1201 08:55:21.314992 4813 scope.go:117] "RemoveContainer" containerID="c33a17bde30aa1c2a6cba8577969415d5485427d650335c0152de62d720b66cb" Dec 01 08:55:21 crc kubenswrapper[4813]: I1201 08:55:21.315000 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bb4hs" event={"ID":"fb7b8647-2f8c-4508-8f11-555d31ad500b","Type":"ContainerDied","Data":"638479af90fa282f6631e725e113ccb3bbbcae729bf099409263f82099afeb56"} Dec 01 08:55:21 crc kubenswrapper[4813]: I1201 08:55:21.317409 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-htzpm" event={"ID":"2a8a0c74-0097-498b-9602-2fa4c73873b3","Type":"ContainerStarted","Data":"746e95def6d20f37461f69aaa5ba1a684908fd4ab7f3cf035a0de5db49931b61"} Dec 01 08:55:21 crc kubenswrapper[4813]: I1201 08:55:21.340632 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-htzpm" podStartSLOduration=2.817800456 podStartE2EDuration="6.340605645s" podCreationTimestamp="2025-12-01 08:55:15 +0000 UTC" firstStartedPulling="2025-12-01 08:55:17.260047319 +0000 UTC m=+901.882968905" lastFinishedPulling="2025-12-01 08:55:20.782852508 +0000 UTC m=+905.405774094" observedRunningTime="2025-12-01 08:55:21.33642271 +0000 UTC m=+905.959344306" watchObservedRunningTime="2025-12-01 08:55:21.340605645 +0000 UTC m=+905.963527231" Dec 01 08:55:21 crc kubenswrapper[4813]: I1201 08:55:21.342369 4813 scope.go:117] "RemoveContainer" containerID="eb03b24cee58d54df865ab101f911285a410fe4bdefd75eb9566a679d247f114" Dec 01 08:55:21 crc kubenswrapper[4813]: I1201 08:55:21.358333 4813 scope.go:117] "RemoveContainer" containerID="562c36cfae2c8ec9e9d4f245aa9f3c8ebb1dabe78b07e01ef0f3c3f90577ecc6" Dec 01 08:55:21 crc kubenswrapper[4813]: I1201 08:55:21.378898 4813 scope.go:117] "RemoveContainer" containerID="c33a17bde30aa1c2a6cba8577969415d5485427d650335c0152de62d720b66cb" Dec 01 08:55:21 crc kubenswrapper[4813]: E1201 08:55:21.379575 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c33a17bde30aa1c2a6cba8577969415d5485427d650335c0152de62d720b66cb\": container with ID starting with c33a17bde30aa1c2a6cba8577969415d5485427d650335c0152de62d720b66cb not found: ID does not exist" containerID="c33a17bde30aa1c2a6cba8577969415d5485427d650335c0152de62d720b66cb" Dec 01 08:55:21 crc kubenswrapper[4813]: I1201 08:55:21.379626 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c33a17bde30aa1c2a6cba8577969415d5485427d650335c0152de62d720b66cb"} err="failed to get container status \"c33a17bde30aa1c2a6cba8577969415d5485427d650335c0152de62d720b66cb\": rpc error: code = NotFound desc = could not find container \"c33a17bde30aa1c2a6cba8577969415d5485427d650335c0152de62d720b66cb\": container with ID starting with c33a17bde30aa1c2a6cba8577969415d5485427d650335c0152de62d720b66cb not found: ID does not exist" Dec 01 08:55:21 crc kubenswrapper[4813]: I1201 08:55:21.379659 4813 scope.go:117] "RemoveContainer" containerID="eb03b24cee58d54df865ab101f911285a410fe4bdefd75eb9566a679d247f114" Dec 01 08:55:21 crc kubenswrapper[4813]: E1201 08:55:21.380090 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"eb03b24cee58d54df865ab101f911285a410fe4bdefd75eb9566a679d247f114\": container with ID starting with eb03b24cee58d54df865ab101f911285a410fe4bdefd75eb9566a679d247f114 not found: ID does not exist" containerID="eb03b24cee58d54df865ab101f911285a410fe4bdefd75eb9566a679d247f114" Dec 01 08:55:21 crc kubenswrapper[4813]: I1201 08:55:21.380204 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eb03b24cee58d54df865ab101f911285a410fe4bdefd75eb9566a679d247f114"} err="failed to get container status \"eb03b24cee58d54df865ab101f911285a410fe4bdefd75eb9566a679d247f114\": rpc error: code = NotFound desc = could not find container \"eb03b24cee58d54df865ab101f911285a410fe4bdefd75eb9566a679d247f114\": container with ID starting with eb03b24cee58d54df865ab101f911285a410fe4bdefd75eb9566a679d247f114 not found: ID does not exist" Dec 01 08:55:21 crc kubenswrapper[4813]: I1201 08:55:21.380309 4813 scope.go:117] "RemoveContainer" containerID="562c36cfae2c8ec9e9d4f245aa9f3c8ebb1dabe78b07e01ef0f3c3f90577ecc6" Dec 01 08:55:21 crc kubenswrapper[4813]: E1201 08:55:21.380793 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"562c36cfae2c8ec9e9d4f245aa9f3c8ebb1dabe78b07e01ef0f3c3f90577ecc6\": container with ID starting with 562c36cfae2c8ec9e9d4f245aa9f3c8ebb1dabe78b07e01ef0f3c3f90577ecc6 not found: ID does not exist" containerID="562c36cfae2c8ec9e9d4f245aa9f3c8ebb1dabe78b07e01ef0f3c3f90577ecc6" Dec 01 08:55:21 crc kubenswrapper[4813]: I1201 08:55:21.380828 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"562c36cfae2c8ec9e9d4f245aa9f3c8ebb1dabe78b07e01ef0f3c3f90577ecc6"} err="failed to get container status \"562c36cfae2c8ec9e9d4f245aa9f3c8ebb1dabe78b07e01ef0f3c3f90577ecc6\": rpc error: code = NotFound desc = could not find container \"562c36cfae2c8ec9e9d4f245aa9f3c8ebb1dabe78b07e01ef0f3c3f90577ecc6\": container with ID starting with 562c36cfae2c8ec9e9d4f245aa9f3c8ebb1dabe78b07e01ef0f3c3f90577ecc6 not found: ID does not exist" Dec 01 08:55:21 crc kubenswrapper[4813]: I1201 08:55:21.405037 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fb7b8647-2f8c-4508-8f11-555d31ad500b-utilities\") pod \"fb7b8647-2f8c-4508-8f11-555d31ad500b\" (UID: \"fb7b8647-2f8c-4508-8f11-555d31ad500b\") " Dec 01 08:55:21 crc kubenswrapper[4813]: I1201 08:55:21.405112 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fz5pz\" (UniqueName: \"kubernetes.io/projected/fb7b8647-2f8c-4508-8f11-555d31ad500b-kube-api-access-fz5pz\") pod \"fb7b8647-2f8c-4508-8f11-555d31ad500b\" (UID: \"fb7b8647-2f8c-4508-8f11-555d31ad500b\") " Dec 01 08:55:21 crc kubenswrapper[4813]: I1201 08:55:21.405155 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fb7b8647-2f8c-4508-8f11-555d31ad500b-catalog-content\") pod \"fb7b8647-2f8c-4508-8f11-555d31ad500b\" (UID: \"fb7b8647-2f8c-4508-8f11-555d31ad500b\") " Dec 01 08:55:21 crc kubenswrapper[4813]: I1201 08:55:21.406829 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fb7b8647-2f8c-4508-8f11-555d31ad500b-utilities" (OuterVolumeSpecName: "utilities") pod "fb7b8647-2f8c-4508-8f11-555d31ad500b" (UID: "fb7b8647-2f8c-4508-8f11-555d31ad500b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 08:55:21 crc kubenswrapper[4813]: I1201 08:55:21.412187 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fb7b8647-2f8c-4508-8f11-555d31ad500b-kube-api-access-fz5pz" (OuterVolumeSpecName: "kube-api-access-fz5pz") pod "fb7b8647-2f8c-4508-8f11-555d31ad500b" (UID: "fb7b8647-2f8c-4508-8f11-555d31ad500b"). InnerVolumeSpecName "kube-api-access-fz5pz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:55:21 crc kubenswrapper[4813]: I1201 08:55:21.424065 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fb7b8647-2f8c-4508-8f11-555d31ad500b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "fb7b8647-2f8c-4508-8f11-555d31ad500b" (UID: "fb7b8647-2f8c-4508-8f11-555d31ad500b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 08:55:21 crc kubenswrapper[4813]: I1201 08:55:21.506758 4813 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fb7b8647-2f8c-4508-8f11-555d31ad500b-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 08:55:21 crc kubenswrapper[4813]: I1201 08:55:21.506804 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fz5pz\" (UniqueName: \"kubernetes.io/projected/fb7b8647-2f8c-4508-8f11-555d31ad500b-kube-api-access-fz5pz\") on node \"crc\" DevicePath \"\"" Dec 01 08:55:21 crc kubenswrapper[4813]: I1201 08:55:21.506817 4813 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fb7b8647-2f8c-4508-8f11-555d31ad500b-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 08:55:21 crc kubenswrapper[4813]: I1201 08:55:21.642916 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-bb4hs"] Dec 01 08:55:21 crc kubenswrapper[4813]: I1201 08:55:21.649358 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-bb4hs"] Dec 01 08:55:22 crc kubenswrapper[4813]: I1201 08:55:22.402713 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b3995a58-451f-458c-ba0c-ed66510679bf" path="/var/lib/kubelet/pods/b3995a58-451f-458c-ba0c-ed66510679bf/volumes" Dec 01 08:55:22 crc kubenswrapper[4813]: I1201 08:55:22.403655 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fb7b8647-2f8c-4508-8f11-555d31ad500b" path="/var/lib/kubelet/pods/fb7b8647-2f8c-4508-8f11-555d31ad500b/volumes" Dec 01 08:55:26 crc kubenswrapper[4813]: I1201 08:55:26.037055 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/rabbitmq-server-0"] Dec 01 08:55:26 crc kubenswrapper[4813]: E1201 08:55:26.037946 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fb7b8647-2f8c-4508-8f11-555d31ad500b" containerName="extract-content" Dec 01 08:55:26 crc kubenswrapper[4813]: I1201 08:55:26.037996 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="fb7b8647-2f8c-4508-8f11-555d31ad500b" containerName="extract-content" Dec 01 08:55:26 crc kubenswrapper[4813]: E1201 08:55:26.038030 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fb7b8647-2f8c-4508-8f11-555d31ad500b" containerName="registry-server" Dec 01 08:55:26 crc kubenswrapper[4813]: I1201 08:55:26.038040 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="fb7b8647-2f8c-4508-8f11-555d31ad500b" containerName="registry-server" Dec 01 08:55:26 crc kubenswrapper[4813]: E1201 08:55:26.038068 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b3995a58-451f-458c-ba0c-ed66510679bf" containerName="extract-content" Dec 01 08:55:26 crc kubenswrapper[4813]: I1201 08:55:26.038075 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="b3995a58-451f-458c-ba0c-ed66510679bf" containerName="extract-content" Dec 01 08:55:26 crc kubenswrapper[4813]: E1201 08:55:26.038085 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fb7b8647-2f8c-4508-8f11-555d31ad500b" containerName="extract-utilities" Dec 01 08:55:26 crc kubenswrapper[4813]: I1201 08:55:26.038094 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="fb7b8647-2f8c-4508-8f11-555d31ad500b" containerName="extract-utilities" Dec 01 08:55:26 crc kubenswrapper[4813]: E1201 08:55:26.038107 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b3995a58-451f-458c-ba0c-ed66510679bf" containerName="extract-utilities" Dec 01 08:55:26 crc kubenswrapper[4813]: I1201 08:55:26.038115 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="b3995a58-451f-458c-ba0c-ed66510679bf" containerName="extract-utilities" Dec 01 08:55:26 crc kubenswrapper[4813]: E1201 08:55:26.038126 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b3995a58-451f-458c-ba0c-ed66510679bf" containerName="registry-server" Dec 01 08:55:26 crc kubenswrapper[4813]: I1201 08:55:26.038133 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="b3995a58-451f-458c-ba0c-ed66510679bf" containerName="registry-server" Dec 01 08:55:26 crc kubenswrapper[4813]: I1201 08:55:26.038342 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="fb7b8647-2f8c-4508-8f11-555d31ad500b" containerName="registry-server" Dec 01 08:55:26 crc kubenswrapper[4813]: I1201 08:55:26.038363 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="b3995a58-451f-458c-ba0c-ed66510679bf" containerName="registry-server" Dec 01 08:55:26 crc kubenswrapper[4813]: I1201 08:55:26.039272 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/rabbitmq-server-0" Dec 01 08:55:26 crc kubenswrapper[4813]: I1201 08:55:26.043137 4813 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"rabbitmq-erlang-cookie" Dec 01 08:55:26 crc kubenswrapper[4813]: I1201 08:55:26.043140 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"rabbitmq-plugins-conf" Dec 01 08:55:26 crc kubenswrapper[4813]: I1201 08:55:26.043632 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"rabbitmq-server-conf" Dec 01 08:55:26 crc kubenswrapper[4813]: I1201 08:55:26.043719 4813 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"rabbitmq-default-user" Dec 01 08:55:26 crc kubenswrapper[4813]: I1201 08:55:26.044409 4813 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"rabbitmq-server-dockercfg-srstk" Dec 01 08:55:26 crc kubenswrapper[4813]: I1201 08:55:26.064664 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/rabbitmq-server-0"] Dec 01 08:55:26 crc kubenswrapper[4813]: I1201 08:55:26.143428 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-8bac5ff1-dca8-4c26-86da-88e356fbd887\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-8bac5ff1-dca8-4c26-86da-88e356fbd887\") pod \"rabbitmq-server-0\" (UID: \"a274574f-1dfe-4afd-8dfa-661d14669e73\") " pod="glance-kuttl-tests/rabbitmq-server-0" Dec 01 08:55:26 crc kubenswrapper[4813]: I1201 08:55:26.143505 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/a274574f-1dfe-4afd-8dfa-661d14669e73-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"a274574f-1dfe-4afd-8dfa-661d14669e73\") " pod="glance-kuttl-tests/rabbitmq-server-0" Dec 01 08:55:26 crc kubenswrapper[4813]: I1201 08:55:26.143551 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/a274574f-1dfe-4afd-8dfa-661d14669e73-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"a274574f-1dfe-4afd-8dfa-661d14669e73\") " pod="glance-kuttl-tests/rabbitmq-server-0" Dec 01 08:55:26 crc kubenswrapper[4813]: I1201 08:55:26.143574 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/a274574f-1dfe-4afd-8dfa-661d14669e73-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"a274574f-1dfe-4afd-8dfa-661d14669e73\") " pod="glance-kuttl-tests/rabbitmq-server-0" Dec 01 08:55:26 crc kubenswrapper[4813]: I1201 08:55:26.143624 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/a274574f-1dfe-4afd-8dfa-661d14669e73-pod-info\") pod \"rabbitmq-server-0\" (UID: \"a274574f-1dfe-4afd-8dfa-661d14669e73\") " pod="glance-kuttl-tests/rabbitmq-server-0" Dec 01 08:55:26 crc kubenswrapper[4813]: I1201 08:55:26.143810 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/a274574f-1dfe-4afd-8dfa-661d14669e73-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"a274574f-1dfe-4afd-8dfa-661d14669e73\") " pod="glance-kuttl-tests/rabbitmq-server-0" Dec 01 08:55:26 crc kubenswrapper[4813]: I1201 08:55:26.143845 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-slrjg\" (UniqueName: \"kubernetes.io/projected/a274574f-1dfe-4afd-8dfa-661d14669e73-kube-api-access-slrjg\") pod \"rabbitmq-server-0\" (UID: \"a274574f-1dfe-4afd-8dfa-661d14669e73\") " pod="glance-kuttl-tests/rabbitmq-server-0" Dec 01 08:55:26 crc kubenswrapper[4813]: I1201 08:55:26.143871 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/a274574f-1dfe-4afd-8dfa-661d14669e73-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"a274574f-1dfe-4afd-8dfa-661d14669e73\") " pod="glance-kuttl-tests/rabbitmq-server-0" Dec 01 08:55:26 crc kubenswrapper[4813]: I1201 08:55:26.244623 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-8bac5ff1-dca8-4c26-86da-88e356fbd887\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-8bac5ff1-dca8-4c26-86da-88e356fbd887\") pod \"rabbitmq-server-0\" (UID: \"a274574f-1dfe-4afd-8dfa-661d14669e73\") " pod="glance-kuttl-tests/rabbitmq-server-0" Dec 01 08:55:26 crc kubenswrapper[4813]: I1201 08:55:26.244710 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/a274574f-1dfe-4afd-8dfa-661d14669e73-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"a274574f-1dfe-4afd-8dfa-661d14669e73\") " pod="glance-kuttl-tests/rabbitmq-server-0" Dec 01 08:55:26 crc kubenswrapper[4813]: I1201 08:55:26.244739 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/a274574f-1dfe-4afd-8dfa-661d14669e73-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"a274574f-1dfe-4afd-8dfa-661d14669e73\") " pod="glance-kuttl-tests/rabbitmq-server-0" Dec 01 08:55:26 crc kubenswrapper[4813]: I1201 08:55:26.244767 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/a274574f-1dfe-4afd-8dfa-661d14669e73-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"a274574f-1dfe-4afd-8dfa-661d14669e73\") " pod="glance-kuttl-tests/rabbitmq-server-0" Dec 01 08:55:26 crc kubenswrapper[4813]: I1201 08:55:26.245756 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/a274574f-1dfe-4afd-8dfa-661d14669e73-pod-info\") pod \"rabbitmq-server-0\" (UID: \"a274574f-1dfe-4afd-8dfa-661d14669e73\") " pod="glance-kuttl-tests/rabbitmq-server-0" Dec 01 08:55:26 crc kubenswrapper[4813]: I1201 08:55:26.245826 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/a274574f-1dfe-4afd-8dfa-661d14669e73-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"a274574f-1dfe-4afd-8dfa-661d14669e73\") " pod="glance-kuttl-tests/rabbitmq-server-0" Dec 01 08:55:26 crc kubenswrapper[4813]: I1201 08:55:26.245853 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-slrjg\" (UniqueName: \"kubernetes.io/projected/a274574f-1dfe-4afd-8dfa-661d14669e73-kube-api-access-slrjg\") pod \"rabbitmq-server-0\" (UID: \"a274574f-1dfe-4afd-8dfa-661d14669e73\") " pod="glance-kuttl-tests/rabbitmq-server-0" Dec 01 08:55:26 crc kubenswrapper[4813]: I1201 08:55:26.245876 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/a274574f-1dfe-4afd-8dfa-661d14669e73-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"a274574f-1dfe-4afd-8dfa-661d14669e73\") " pod="glance-kuttl-tests/rabbitmq-server-0" Dec 01 08:55:26 crc kubenswrapper[4813]: I1201 08:55:26.246071 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/a274574f-1dfe-4afd-8dfa-661d14669e73-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"a274574f-1dfe-4afd-8dfa-661d14669e73\") " pod="glance-kuttl-tests/rabbitmq-server-0" Dec 01 08:55:26 crc kubenswrapper[4813]: I1201 08:55:26.246202 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/a274574f-1dfe-4afd-8dfa-661d14669e73-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"a274574f-1dfe-4afd-8dfa-661d14669e73\") " pod="glance-kuttl-tests/rabbitmq-server-0" Dec 01 08:55:26 crc kubenswrapper[4813]: I1201 08:55:26.246578 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/a274574f-1dfe-4afd-8dfa-661d14669e73-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"a274574f-1dfe-4afd-8dfa-661d14669e73\") " pod="glance-kuttl-tests/rabbitmq-server-0" Dec 01 08:55:26 crc kubenswrapper[4813]: I1201 08:55:26.250355 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/a274574f-1dfe-4afd-8dfa-661d14669e73-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"a274574f-1dfe-4afd-8dfa-661d14669e73\") " pod="glance-kuttl-tests/rabbitmq-server-0" Dec 01 08:55:26 crc kubenswrapper[4813]: I1201 08:55:26.250625 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/a274574f-1dfe-4afd-8dfa-661d14669e73-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"a274574f-1dfe-4afd-8dfa-661d14669e73\") " pod="glance-kuttl-tests/rabbitmq-server-0" Dec 01 08:55:26 crc kubenswrapper[4813]: I1201 08:55:26.256235 4813 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 01 08:55:26 crc kubenswrapper[4813]: I1201 08:55:26.256271 4813 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-8bac5ff1-dca8-4c26-86da-88e356fbd887\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-8bac5ff1-dca8-4c26-86da-88e356fbd887\") pod \"rabbitmq-server-0\" (UID: \"a274574f-1dfe-4afd-8dfa-661d14669e73\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/e1b48a400236afd23b05e965ea25f781608e41347943b26558a7696402de83a7/globalmount\"" pod="glance-kuttl-tests/rabbitmq-server-0" Dec 01 08:55:26 crc kubenswrapper[4813]: I1201 08:55:26.259550 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/a274574f-1dfe-4afd-8dfa-661d14669e73-pod-info\") pod \"rabbitmq-server-0\" (UID: \"a274574f-1dfe-4afd-8dfa-661d14669e73\") " pod="glance-kuttl-tests/rabbitmq-server-0" Dec 01 08:55:26 crc kubenswrapper[4813]: I1201 08:55:26.277752 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-slrjg\" (UniqueName: \"kubernetes.io/projected/a274574f-1dfe-4afd-8dfa-661d14669e73-kube-api-access-slrjg\") pod \"rabbitmq-server-0\" (UID: \"a274574f-1dfe-4afd-8dfa-661d14669e73\") " pod="glance-kuttl-tests/rabbitmq-server-0" Dec 01 08:55:26 crc kubenswrapper[4813]: I1201 08:55:26.294637 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-htzpm" Dec 01 08:55:26 crc kubenswrapper[4813]: I1201 08:55:26.349660 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-htzpm" Dec 01 08:55:26 crc kubenswrapper[4813]: I1201 08:55:26.359511 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-8bac5ff1-dca8-4c26-86da-88e356fbd887\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-8bac5ff1-dca8-4c26-86da-88e356fbd887\") pod \"rabbitmq-server-0\" (UID: \"a274574f-1dfe-4afd-8dfa-661d14669e73\") " pod="glance-kuttl-tests/rabbitmq-server-0" Dec 01 08:55:26 crc kubenswrapper[4813]: I1201 08:55:26.368182 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/rabbitmq-server-0" Dec 01 08:55:26 crc kubenswrapper[4813]: I1201 08:55:26.402106 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-htzpm" Dec 01 08:55:26 crc kubenswrapper[4813]: I1201 08:55:26.778797 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/rabbitmq-server-0"] Dec 01 08:55:27 crc kubenswrapper[4813]: I1201 08:55:27.383054 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/rabbitmq-server-0" event={"ID":"a274574f-1dfe-4afd-8dfa-661d14669e73","Type":"ContainerStarted","Data":"79eec5e6a0c1cdd813c88f2473920813cf0c85add05c31adb8bb9f3829f2f8a4"} Dec 01 08:55:27 crc kubenswrapper[4813]: I1201 08:55:27.430334 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-htzpm" Dec 01 08:55:28 crc kubenswrapper[4813]: I1201 08:55:28.558551 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-grnss"] Dec 01 08:55:28 crc kubenswrapper[4813]: I1201 08:55:28.558905 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-grnss" podUID="5a12a2ee-1fa9-4e27-b8b5-0b1c1f721c5b" containerName="registry-server" containerID="cri-o://8f54da59742c98d1ef046d47ca2d3d6ab5ed651541cbda700b84dbbe176420a1" gracePeriod=2 Dec 01 08:55:28 crc kubenswrapper[4813]: I1201 08:55:28.982478 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-grnss" Dec 01 08:55:29 crc kubenswrapper[4813]: I1201 08:55:29.004518 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5a12a2ee-1fa9-4e27-b8b5-0b1c1f721c5b-catalog-content\") pod \"5a12a2ee-1fa9-4e27-b8b5-0b1c1f721c5b\" (UID: \"5a12a2ee-1fa9-4e27-b8b5-0b1c1f721c5b\") " Dec 01 08:55:29 crc kubenswrapper[4813]: I1201 08:55:29.004569 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qdwd5\" (UniqueName: \"kubernetes.io/projected/5a12a2ee-1fa9-4e27-b8b5-0b1c1f721c5b-kube-api-access-qdwd5\") pod \"5a12a2ee-1fa9-4e27-b8b5-0b1c1f721c5b\" (UID: \"5a12a2ee-1fa9-4e27-b8b5-0b1c1f721c5b\") " Dec 01 08:55:29 crc kubenswrapper[4813]: I1201 08:55:29.004678 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5a12a2ee-1fa9-4e27-b8b5-0b1c1f721c5b-utilities\") pod \"5a12a2ee-1fa9-4e27-b8b5-0b1c1f721c5b\" (UID: \"5a12a2ee-1fa9-4e27-b8b5-0b1c1f721c5b\") " Dec 01 08:55:29 crc kubenswrapper[4813]: I1201 08:55:29.006675 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5a12a2ee-1fa9-4e27-b8b5-0b1c1f721c5b-utilities" (OuterVolumeSpecName: "utilities") pod "5a12a2ee-1fa9-4e27-b8b5-0b1c1f721c5b" (UID: "5a12a2ee-1fa9-4e27-b8b5-0b1c1f721c5b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 08:55:29 crc kubenswrapper[4813]: I1201 08:55:29.028899 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5a12a2ee-1fa9-4e27-b8b5-0b1c1f721c5b-kube-api-access-qdwd5" (OuterVolumeSpecName: "kube-api-access-qdwd5") pod "5a12a2ee-1fa9-4e27-b8b5-0b1c1f721c5b" (UID: "5a12a2ee-1fa9-4e27-b8b5-0b1c1f721c5b"). InnerVolumeSpecName "kube-api-access-qdwd5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:55:29 crc kubenswrapper[4813]: I1201 08:55:29.106653 4813 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5a12a2ee-1fa9-4e27-b8b5-0b1c1f721c5b-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 08:55:29 crc kubenswrapper[4813]: I1201 08:55:29.106688 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qdwd5\" (UniqueName: \"kubernetes.io/projected/5a12a2ee-1fa9-4e27-b8b5-0b1c1f721c5b-kube-api-access-qdwd5\") on node \"crc\" DevicePath \"\"" Dec 01 08:55:29 crc kubenswrapper[4813]: I1201 08:55:29.122375 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5a12a2ee-1fa9-4e27-b8b5-0b1c1f721c5b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5a12a2ee-1fa9-4e27-b8b5-0b1c1f721c5b" (UID: "5a12a2ee-1fa9-4e27-b8b5-0b1c1f721c5b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 08:55:29 crc kubenswrapper[4813]: I1201 08:55:29.208069 4813 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5a12a2ee-1fa9-4e27-b8b5-0b1c1f721c5b-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 08:55:29 crc kubenswrapper[4813]: I1201 08:55:29.404707 4813 generic.go:334] "Generic (PLEG): container finished" podID="5a12a2ee-1fa9-4e27-b8b5-0b1c1f721c5b" containerID="8f54da59742c98d1ef046d47ca2d3d6ab5ed651541cbda700b84dbbe176420a1" exitCode=0 Dec 01 08:55:29 crc kubenswrapper[4813]: I1201 08:55:29.404751 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-grnss" event={"ID":"5a12a2ee-1fa9-4e27-b8b5-0b1c1f721c5b","Type":"ContainerDied","Data":"8f54da59742c98d1ef046d47ca2d3d6ab5ed651541cbda700b84dbbe176420a1"} Dec 01 08:55:29 crc kubenswrapper[4813]: I1201 08:55:29.404762 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-grnss" Dec 01 08:55:29 crc kubenswrapper[4813]: I1201 08:55:29.404860 4813 scope.go:117] "RemoveContainer" containerID="8f54da59742c98d1ef046d47ca2d3d6ab5ed651541cbda700b84dbbe176420a1" Dec 01 08:55:29 crc kubenswrapper[4813]: I1201 08:55:29.404962 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-grnss" event={"ID":"5a12a2ee-1fa9-4e27-b8b5-0b1c1f721c5b","Type":"ContainerDied","Data":"2867de8145a3863ccb0a345d7e772141359217d8621d36a9866643e887338d1b"} Dec 01 08:55:29 crc kubenswrapper[4813]: I1201 08:55:29.436175 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-grnss"] Dec 01 08:55:29 crc kubenswrapper[4813]: I1201 08:55:29.441599 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-grnss"] Dec 01 08:55:30 crc kubenswrapper[4813]: I1201 08:55:30.423904 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5a12a2ee-1fa9-4e27-b8b5-0b1c1f721c5b" path="/var/lib/kubelet/pods/5a12a2ee-1fa9-4e27-b8b5-0b1c1f721c5b/volumes" Dec 01 08:55:31 crc kubenswrapper[4813]: I1201 08:55:31.069401 4813 scope.go:117] "RemoveContainer" containerID="fe936d6ad53b8083d4287dd8a478c15285de5e87babe01cdc5bbcbd8ab47a8cd" Dec 01 08:55:32 crc kubenswrapper[4813]: I1201 08:55:32.180033 4813 scope.go:117] "RemoveContainer" containerID="96227608cd663c5dc1fe44bd6cac947107c43e4749e4a97e836e0e805d33960c" Dec 01 08:55:32 crc kubenswrapper[4813]: I1201 08:55:32.772236 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-index-jvhf5"] Dec 01 08:55:32 crc kubenswrapper[4813]: E1201 08:55:32.772541 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5a12a2ee-1fa9-4e27-b8b5-0b1c1f721c5b" containerName="extract-content" Dec 01 08:55:32 crc kubenswrapper[4813]: I1201 08:55:32.772553 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="5a12a2ee-1fa9-4e27-b8b5-0b1c1f721c5b" containerName="extract-content" Dec 01 08:55:32 crc kubenswrapper[4813]: E1201 08:55:32.772576 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5a12a2ee-1fa9-4e27-b8b5-0b1c1f721c5b" containerName="registry-server" Dec 01 08:55:32 crc kubenswrapper[4813]: I1201 08:55:32.772582 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="5a12a2ee-1fa9-4e27-b8b5-0b1c1f721c5b" containerName="registry-server" Dec 01 08:55:32 crc kubenswrapper[4813]: E1201 08:55:32.772591 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5a12a2ee-1fa9-4e27-b8b5-0b1c1f721c5b" containerName="extract-utilities" Dec 01 08:55:32 crc kubenswrapper[4813]: I1201 08:55:32.772597 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="5a12a2ee-1fa9-4e27-b8b5-0b1c1f721c5b" containerName="extract-utilities" Dec 01 08:55:32 crc kubenswrapper[4813]: I1201 08:55:32.772700 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="5a12a2ee-1fa9-4e27-b8b5-0b1c1f721c5b" containerName="registry-server" Dec 01 08:55:32 crc kubenswrapper[4813]: I1201 08:55:32.773240 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-index-jvhf5" Dec 01 08:55:32 crc kubenswrapper[4813]: I1201 08:55:32.776663 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-index-dockercfg-4g7dp" Dec 01 08:55:32 crc kubenswrapper[4813]: I1201 08:55:32.780989 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-index-jvhf5"] Dec 01 08:55:32 crc kubenswrapper[4813]: I1201 08:55:32.809110 4813 scope.go:117] "RemoveContainer" containerID="8f54da59742c98d1ef046d47ca2d3d6ab5ed651541cbda700b84dbbe176420a1" Dec 01 08:55:32 crc kubenswrapper[4813]: E1201 08:55:32.809712 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8f54da59742c98d1ef046d47ca2d3d6ab5ed651541cbda700b84dbbe176420a1\": container with ID starting with 8f54da59742c98d1ef046d47ca2d3d6ab5ed651541cbda700b84dbbe176420a1 not found: ID does not exist" containerID="8f54da59742c98d1ef046d47ca2d3d6ab5ed651541cbda700b84dbbe176420a1" Dec 01 08:55:32 crc kubenswrapper[4813]: I1201 08:55:32.809760 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8f54da59742c98d1ef046d47ca2d3d6ab5ed651541cbda700b84dbbe176420a1"} err="failed to get container status \"8f54da59742c98d1ef046d47ca2d3d6ab5ed651541cbda700b84dbbe176420a1\": rpc error: code = NotFound desc = could not find container \"8f54da59742c98d1ef046d47ca2d3d6ab5ed651541cbda700b84dbbe176420a1\": container with ID starting with 8f54da59742c98d1ef046d47ca2d3d6ab5ed651541cbda700b84dbbe176420a1 not found: ID does not exist" Dec 01 08:55:32 crc kubenswrapper[4813]: I1201 08:55:32.809791 4813 scope.go:117] "RemoveContainer" containerID="fe936d6ad53b8083d4287dd8a478c15285de5e87babe01cdc5bbcbd8ab47a8cd" Dec 01 08:55:32 crc kubenswrapper[4813]: E1201 08:55:32.810028 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fe936d6ad53b8083d4287dd8a478c15285de5e87babe01cdc5bbcbd8ab47a8cd\": container with ID starting with fe936d6ad53b8083d4287dd8a478c15285de5e87babe01cdc5bbcbd8ab47a8cd not found: ID does not exist" containerID="fe936d6ad53b8083d4287dd8a478c15285de5e87babe01cdc5bbcbd8ab47a8cd" Dec 01 08:55:32 crc kubenswrapper[4813]: I1201 08:55:32.810068 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fe936d6ad53b8083d4287dd8a478c15285de5e87babe01cdc5bbcbd8ab47a8cd"} err="failed to get container status \"fe936d6ad53b8083d4287dd8a478c15285de5e87babe01cdc5bbcbd8ab47a8cd\": rpc error: code = NotFound desc = could not find container \"fe936d6ad53b8083d4287dd8a478c15285de5e87babe01cdc5bbcbd8ab47a8cd\": container with ID starting with fe936d6ad53b8083d4287dd8a478c15285de5e87babe01cdc5bbcbd8ab47a8cd not found: ID does not exist" Dec 01 08:55:32 crc kubenswrapper[4813]: I1201 08:55:32.810089 4813 scope.go:117] "RemoveContainer" containerID="96227608cd663c5dc1fe44bd6cac947107c43e4749e4a97e836e0e805d33960c" Dec 01 08:55:32 crc kubenswrapper[4813]: E1201 08:55:32.810301 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"96227608cd663c5dc1fe44bd6cac947107c43e4749e4a97e836e0e805d33960c\": container with ID starting with 96227608cd663c5dc1fe44bd6cac947107c43e4749e4a97e836e0e805d33960c not found: ID does not exist" containerID="96227608cd663c5dc1fe44bd6cac947107c43e4749e4a97e836e0e805d33960c" Dec 01 08:55:32 crc kubenswrapper[4813]: I1201 08:55:32.810329 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"96227608cd663c5dc1fe44bd6cac947107c43e4749e4a97e836e0e805d33960c"} err="failed to get container status \"96227608cd663c5dc1fe44bd6cac947107c43e4749e4a97e836e0e805d33960c\": rpc error: code = NotFound desc = could not find container \"96227608cd663c5dc1fe44bd6cac947107c43e4749e4a97e836e0e805d33960c\": container with ID starting with 96227608cd663c5dc1fe44bd6cac947107c43e4749e4a97e836e0e805d33960c not found: ID does not exist" Dec 01 08:55:32 crc kubenswrapper[4813]: I1201 08:55:32.868258 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hkctl\" (UniqueName: \"kubernetes.io/projected/6fee8f87-bd86-4153-81b5-fbfc619f609c-kube-api-access-hkctl\") pod \"keystone-operator-index-jvhf5\" (UID: \"6fee8f87-bd86-4153-81b5-fbfc619f609c\") " pod="openstack-operators/keystone-operator-index-jvhf5" Dec 01 08:55:32 crc kubenswrapper[4813]: I1201 08:55:32.969200 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hkctl\" (UniqueName: \"kubernetes.io/projected/6fee8f87-bd86-4153-81b5-fbfc619f609c-kube-api-access-hkctl\") pod \"keystone-operator-index-jvhf5\" (UID: \"6fee8f87-bd86-4153-81b5-fbfc619f609c\") " pod="openstack-operators/keystone-operator-index-jvhf5" Dec 01 08:55:32 crc kubenswrapper[4813]: I1201 08:55:32.999579 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hkctl\" (UniqueName: \"kubernetes.io/projected/6fee8f87-bd86-4153-81b5-fbfc619f609c-kube-api-access-hkctl\") pod \"keystone-operator-index-jvhf5\" (UID: \"6fee8f87-bd86-4153-81b5-fbfc619f609c\") " pod="openstack-operators/keystone-operator-index-jvhf5" Dec 01 08:55:33 crc kubenswrapper[4813]: I1201 08:55:33.094994 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-index-jvhf5" Dec 01 08:55:33 crc kubenswrapper[4813]: I1201 08:55:33.363679 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-htzpm"] Dec 01 08:55:33 crc kubenswrapper[4813]: I1201 08:55:33.364476 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-htzpm" podUID="2a8a0c74-0097-498b-9602-2fa4c73873b3" containerName="registry-server" containerID="cri-o://746e95def6d20f37461f69aaa5ba1a684908fd4ab7f3cf035a0de5db49931b61" gracePeriod=2 Dec 01 08:55:33 crc kubenswrapper[4813]: I1201 08:55:33.535953 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-index-jvhf5"] Dec 01 08:55:33 crc kubenswrapper[4813]: W1201 08:55:33.547535 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6fee8f87_bd86_4153_81b5_fbfc619f609c.slice/crio-b65512b2d2468332cfba7275c8f8d3d5165263e2a50fa1296e0b99620592f164 WatchSource:0}: Error finding container b65512b2d2468332cfba7275c8f8d3d5165263e2a50fa1296e0b99620592f164: Status 404 returned error can't find the container with id b65512b2d2468332cfba7275c8f8d3d5165263e2a50fa1296e0b99620592f164 Dec 01 08:55:33 crc kubenswrapper[4813]: I1201 08:55:33.883472 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-htzpm" Dec 01 08:55:34 crc kubenswrapper[4813]: I1201 08:55:34.032877 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v8qp5\" (UniqueName: \"kubernetes.io/projected/2a8a0c74-0097-498b-9602-2fa4c73873b3-kube-api-access-v8qp5\") pod \"2a8a0c74-0097-498b-9602-2fa4c73873b3\" (UID: \"2a8a0c74-0097-498b-9602-2fa4c73873b3\") " Dec 01 08:55:34 crc kubenswrapper[4813]: I1201 08:55:34.032948 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2a8a0c74-0097-498b-9602-2fa4c73873b3-catalog-content\") pod \"2a8a0c74-0097-498b-9602-2fa4c73873b3\" (UID: \"2a8a0c74-0097-498b-9602-2fa4c73873b3\") " Dec 01 08:55:34 crc kubenswrapper[4813]: I1201 08:55:34.033062 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2a8a0c74-0097-498b-9602-2fa4c73873b3-utilities\") pod \"2a8a0c74-0097-498b-9602-2fa4c73873b3\" (UID: \"2a8a0c74-0097-498b-9602-2fa4c73873b3\") " Dec 01 08:55:34 crc kubenswrapper[4813]: I1201 08:55:34.035059 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2a8a0c74-0097-498b-9602-2fa4c73873b3-utilities" (OuterVolumeSpecName: "utilities") pod "2a8a0c74-0097-498b-9602-2fa4c73873b3" (UID: "2a8a0c74-0097-498b-9602-2fa4c73873b3"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 08:55:34 crc kubenswrapper[4813]: I1201 08:55:34.039276 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2a8a0c74-0097-498b-9602-2fa4c73873b3-kube-api-access-v8qp5" (OuterVolumeSpecName: "kube-api-access-v8qp5") pod "2a8a0c74-0097-498b-9602-2fa4c73873b3" (UID: "2a8a0c74-0097-498b-9602-2fa4c73873b3"). InnerVolumeSpecName "kube-api-access-v8qp5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:55:34 crc kubenswrapper[4813]: I1201 08:55:34.086768 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2a8a0c74-0097-498b-9602-2fa4c73873b3-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2a8a0c74-0097-498b-9602-2fa4c73873b3" (UID: "2a8a0c74-0097-498b-9602-2fa4c73873b3"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 08:55:34 crc kubenswrapper[4813]: I1201 08:55:34.135250 4813 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2a8a0c74-0097-498b-9602-2fa4c73873b3-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 08:55:34 crc kubenswrapper[4813]: I1201 08:55:34.135298 4813 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2a8a0c74-0097-498b-9602-2fa4c73873b3-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 08:55:34 crc kubenswrapper[4813]: I1201 08:55:34.135327 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v8qp5\" (UniqueName: \"kubernetes.io/projected/2a8a0c74-0097-498b-9602-2fa4c73873b3-kube-api-access-v8qp5\") on node \"crc\" DevicePath \"\"" Dec 01 08:55:34 crc kubenswrapper[4813]: I1201 08:55:34.461126 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/rabbitmq-server-0" event={"ID":"a274574f-1dfe-4afd-8dfa-661d14669e73","Type":"ContainerStarted","Data":"a70c9b4b1ae94521445b57103c7df87633fe6b94256e1fa57ce08bd340754a02"} Dec 01 08:55:34 crc kubenswrapper[4813]: I1201 08:55:34.463537 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-index-jvhf5" event={"ID":"6fee8f87-bd86-4153-81b5-fbfc619f609c","Type":"ContainerStarted","Data":"b65512b2d2468332cfba7275c8f8d3d5165263e2a50fa1296e0b99620592f164"} Dec 01 08:55:34 crc kubenswrapper[4813]: I1201 08:55:34.468676 4813 generic.go:334] "Generic (PLEG): container finished" podID="2a8a0c74-0097-498b-9602-2fa4c73873b3" containerID="746e95def6d20f37461f69aaa5ba1a684908fd4ab7f3cf035a0de5db49931b61" exitCode=0 Dec 01 08:55:34 crc kubenswrapper[4813]: I1201 08:55:34.468795 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-htzpm" event={"ID":"2a8a0c74-0097-498b-9602-2fa4c73873b3","Type":"ContainerDied","Data":"746e95def6d20f37461f69aaa5ba1a684908fd4ab7f3cf035a0de5db49931b61"} Dec 01 08:55:34 crc kubenswrapper[4813]: I1201 08:55:34.468846 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-htzpm" event={"ID":"2a8a0c74-0097-498b-9602-2fa4c73873b3","Type":"ContainerDied","Data":"ee9738c25d3959b113e33f8357b6179c88ccb20f0b8ebc619f06470f1e8c73be"} Dec 01 08:55:34 crc kubenswrapper[4813]: I1201 08:55:34.468886 4813 scope.go:117] "RemoveContainer" containerID="746e95def6d20f37461f69aaa5ba1a684908fd4ab7f3cf035a0de5db49931b61" Dec 01 08:55:34 crc kubenswrapper[4813]: I1201 08:55:34.469375 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-htzpm" Dec 01 08:55:34 crc kubenswrapper[4813]: I1201 08:55:34.500875 4813 scope.go:117] "RemoveContainer" containerID="ab2ee4fe15992d4985bd93913356cf8395a7f9cf09ee8b544236dfb7aac2f901" Dec 01 08:55:34 crc kubenswrapper[4813]: I1201 08:55:34.522614 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-htzpm"] Dec 01 08:55:34 crc kubenswrapper[4813]: I1201 08:55:34.530769 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-htzpm"] Dec 01 08:55:34 crc kubenswrapper[4813]: I1201 08:55:34.531490 4813 scope.go:117] "RemoveContainer" containerID="ddc16dc13accc79cff7a425e620a8d4182519a67dc43b0204b381636dd83264d" Dec 01 08:55:34 crc kubenswrapper[4813]: I1201 08:55:34.555379 4813 scope.go:117] "RemoveContainer" containerID="746e95def6d20f37461f69aaa5ba1a684908fd4ab7f3cf035a0de5db49931b61" Dec 01 08:55:34 crc kubenswrapper[4813]: E1201 08:55:34.556064 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"746e95def6d20f37461f69aaa5ba1a684908fd4ab7f3cf035a0de5db49931b61\": container with ID starting with 746e95def6d20f37461f69aaa5ba1a684908fd4ab7f3cf035a0de5db49931b61 not found: ID does not exist" containerID="746e95def6d20f37461f69aaa5ba1a684908fd4ab7f3cf035a0de5db49931b61" Dec 01 08:55:34 crc kubenswrapper[4813]: I1201 08:55:34.556120 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"746e95def6d20f37461f69aaa5ba1a684908fd4ab7f3cf035a0de5db49931b61"} err="failed to get container status \"746e95def6d20f37461f69aaa5ba1a684908fd4ab7f3cf035a0de5db49931b61\": rpc error: code = NotFound desc = could not find container \"746e95def6d20f37461f69aaa5ba1a684908fd4ab7f3cf035a0de5db49931b61\": container with ID starting with 746e95def6d20f37461f69aaa5ba1a684908fd4ab7f3cf035a0de5db49931b61 not found: ID does not exist" Dec 01 08:55:34 crc kubenswrapper[4813]: I1201 08:55:34.556168 4813 scope.go:117] "RemoveContainer" containerID="ab2ee4fe15992d4985bd93913356cf8395a7f9cf09ee8b544236dfb7aac2f901" Dec 01 08:55:34 crc kubenswrapper[4813]: E1201 08:55:34.556516 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ab2ee4fe15992d4985bd93913356cf8395a7f9cf09ee8b544236dfb7aac2f901\": container with ID starting with ab2ee4fe15992d4985bd93913356cf8395a7f9cf09ee8b544236dfb7aac2f901 not found: ID does not exist" containerID="ab2ee4fe15992d4985bd93913356cf8395a7f9cf09ee8b544236dfb7aac2f901" Dec 01 08:55:34 crc kubenswrapper[4813]: I1201 08:55:34.556557 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ab2ee4fe15992d4985bd93913356cf8395a7f9cf09ee8b544236dfb7aac2f901"} err="failed to get container status \"ab2ee4fe15992d4985bd93913356cf8395a7f9cf09ee8b544236dfb7aac2f901\": rpc error: code = NotFound desc = could not find container \"ab2ee4fe15992d4985bd93913356cf8395a7f9cf09ee8b544236dfb7aac2f901\": container with ID starting with ab2ee4fe15992d4985bd93913356cf8395a7f9cf09ee8b544236dfb7aac2f901 not found: ID does not exist" Dec 01 08:55:34 crc kubenswrapper[4813]: I1201 08:55:34.556574 4813 scope.go:117] "RemoveContainer" containerID="ddc16dc13accc79cff7a425e620a8d4182519a67dc43b0204b381636dd83264d" Dec 01 08:55:34 crc kubenswrapper[4813]: E1201 08:55:34.557043 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ddc16dc13accc79cff7a425e620a8d4182519a67dc43b0204b381636dd83264d\": container with ID starting with ddc16dc13accc79cff7a425e620a8d4182519a67dc43b0204b381636dd83264d not found: ID does not exist" containerID="ddc16dc13accc79cff7a425e620a8d4182519a67dc43b0204b381636dd83264d" Dec 01 08:55:34 crc kubenswrapper[4813]: I1201 08:55:34.557067 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ddc16dc13accc79cff7a425e620a8d4182519a67dc43b0204b381636dd83264d"} err="failed to get container status \"ddc16dc13accc79cff7a425e620a8d4182519a67dc43b0204b381636dd83264d\": rpc error: code = NotFound desc = could not find container \"ddc16dc13accc79cff7a425e620a8d4182519a67dc43b0204b381636dd83264d\": container with ID starting with ddc16dc13accc79cff7a425e620a8d4182519a67dc43b0204b381636dd83264d not found: ID does not exist" Dec 01 08:55:35 crc kubenswrapper[4813]: I1201 08:55:35.483872 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-index-jvhf5" event={"ID":"6fee8f87-bd86-4153-81b5-fbfc619f609c","Type":"ContainerStarted","Data":"7419e96d359c1e84aaf7eb33c841e99f68729bb2e4bbfb0c84284f2139f8559d"} Dec 01 08:55:35 crc kubenswrapper[4813]: I1201 08:55:35.508308 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-index-jvhf5" podStartSLOduration=2.007188799 podStartE2EDuration="3.508252003s" podCreationTimestamp="2025-12-01 08:55:32 +0000 UTC" firstStartedPulling="2025-12-01 08:55:33.551643577 +0000 UTC m=+918.174565173" lastFinishedPulling="2025-12-01 08:55:35.052706751 +0000 UTC m=+919.675628377" observedRunningTime="2025-12-01 08:55:35.49886467 +0000 UTC m=+920.121786266" watchObservedRunningTime="2025-12-01 08:55:35.508252003 +0000 UTC m=+920.131173609" Dec 01 08:55:36 crc kubenswrapper[4813]: I1201 08:55:36.407196 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2a8a0c74-0097-498b-9602-2fa4c73873b3" path="/var/lib/kubelet/pods/2a8a0c74-0097-498b-9602-2fa4c73873b3/volumes" Dec 01 08:55:43 crc kubenswrapper[4813]: I1201 08:55:43.095946 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-index-jvhf5" Dec 01 08:55:43 crc kubenswrapper[4813]: I1201 08:55:43.096647 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/keystone-operator-index-jvhf5" Dec 01 08:55:43 crc kubenswrapper[4813]: I1201 08:55:43.207787 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/keystone-operator-index-jvhf5" Dec 01 08:55:43 crc kubenswrapper[4813]: I1201 08:55:43.587134 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-index-jvhf5" Dec 01 08:55:45 crc kubenswrapper[4813]: I1201 08:55:45.447393 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/49c083020ae5dfe237b73a6c6b807501660a323f061d7879268c43a121stbkr"] Dec 01 08:55:45 crc kubenswrapper[4813]: E1201 08:55:45.448239 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2a8a0c74-0097-498b-9602-2fa4c73873b3" containerName="registry-server" Dec 01 08:55:45 crc kubenswrapper[4813]: I1201 08:55:45.448303 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="2a8a0c74-0097-498b-9602-2fa4c73873b3" containerName="registry-server" Dec 01 08:55:45 crc kubenswrapper[4813]: E1201 08:55:45.448371 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2a8a0c74-0097-498b-9602-2fa4c73873b3" containerName="extract-utilities" Dec 01 08:55:45 crc kubenswrapper[4813]: I1201 08:55:45.448394 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="2a8a0c74-0097-498b-9602-2fa4c73873b3" containerName="extract-utilities" Dec 01 08:55:45 crc kubenswrapper[4813]: E1201 08:55:45.448420 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2a8a0c74-0097-498b-9602-2fa4c73873b3" containerName="extract-content" Dec 01 08:55:45 crc kubenswrapper[4813]: I1201 08:55:45.448439 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="2a8a0c74-0097-498b-9602-2fa4c73873b3" containerName="extract-content" Dec 01 08:55:45 crc kubenswrapper[4813]: I1201 08:55:45.448865 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="2a8a0c74-0097-498b-9602-2fa4c73873b3" containerName="registry-server" Dec 01 08:55:45 crc kubenswrapper[4813]: I1201 08:55:45.451488 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/49c083020ae5dfe237b73a6c6b807501660a323f061d7879268c43a121stbkr" Dec 01 08:55:45 crc kubenswrapper[4813]: I1201 08:55:45.454511 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-fml9r" Dec 01 08:55:45 crc kubenswrapper[4813]: I1201 08:55:45.489582 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/49c083020ae5dfe237b73a6c6b807501660a323f061d7879268c43a121stbkr"] Dec 01 08:55:45 crc kubenswrapper[4813]: I1201 08:55:45.631139 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/9f829ff8-1e05-44b9-85f2-570a3c0e6016-util\") pod \"49c083020ae5dfe237b73a6c6b807501660a323f061d7879268c43a121stbkr\" (UID: \"9f829ff8-1e05-44b9-85f2-570a3c0e6016\") " pod="openstack-operators/49c083020ae5dfe237b73a6c6b807501660a323f061d7879268c43a121stbkr" Dec 01 08:55:45 crc kubenswrapper[4813]: I1201 08:55:45.631264 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wbv4m\" (UniqueName: \"kubernetes.io/projected/9f829ff8-1e05-44b9-85f2-570a3c0e6016-kube-api-access-wbv4m\") pod \"49c083020ae5dfe237b73a6c6b807501660a323f061d7879268c43a121stbkr\" (UID: \"9f829ff8-1e05-44b9-85f2-570a3c0e6016\") " pod="openstack-operators/49c083020ae5dfe237b73a6c6b807501660a323f061d7879268c43a121stbkr" Dec 01 08:55:45 crc kubenswrapper[4813]: I1201 08:55:45.631308 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/9f829ff8-1e05-44b9-85f2-570a3c0e6016-bundle\") pod \"49c083020ae5dfe237b73a6c6b807501660a323f061d7879268c43a121stbkr\" (UID: \"9f829ff8-1e05-44b9-85f2-570a3c0e6016\") " pod="openstack-operators/49c083020ae5dfe237b73a6c6b807501660a323f061d7879268c43a121stbkr" Dec 01 08:55:45 crc kubenswrapper[4813]: I1201 08:55:45.733353 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/9f829ff8-1e05-44b9-85f2-570a3c0e6016-util\") pod \"49c083020ae5dfe237b73a6c6b807501660a323f061d7879268c43a121stbkr\" (UID: \"9f829ff8-1e05-44b9-85f2-570a3c0e6016\") " pod="openstack-operators/49c083020ae5dfe237b73a6c6b807501660a323f061d7879268c43a121stbkr" Dec 01 08:55:45 crc kubenswrapper[4813]: I1201 08:55:45.733551 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wbv4m\" (UniqueName: \"kubernetes.io/projected/9f829ff8-1e05-44b9-85f2-570a3c0e6016-kube-api-access-wbv4m\") pod \"49c083020ae5dfe237b73a6c6b807501660a323f061d7879268c43a121stbkr\" (UID: \"9f829ff8-1e05-44b9-85f2-570a3c0e6016\") " pod="openstack-operators/49c083020ae5dfe237b73a6c6b807501660a323f061d7879268c43a121stbkr" Dec 01 08:55:45 crc kubenswrapper[4813]: I1201 08:55:45.733620 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/9f829ff8-1e05-44b9-85f2-570a3c0e6016-bundle\") pod \"49c083020ae5dfe237b73a6c6b807501660a323f061d7879268c43a121stbkr\" (UID: \"9f829ff8-1e05-44b9-85f2-570a3c0e6016\") " pod="openstack-operators/49c083020ae5dfe237b73a6c6b807501660a323f061d7879268c43a121stbkr" Dec 01 08:55:45 crc kubenswrapper[4813]: I1201 08:55:45.734039 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/9f829ff8-1e05-44b9-85f2-570a3c0e6016-util\") pod \"49c083020ae5dfe237b73a6c6b807501660a323f061d7879268c43a121stbkr\" (UID: \"9f829ff8-1e05-44b9-85f2-570a3c0e6016\") " pod="openstack-operators/49c083020ae5dfe237b73a6c6b807501660a323f061d7879268c43a121stbkr" Dec 01 08:55:45 crc kubenswrapper[4813]: I1201 08:55:45.734581 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/9f829ff8-1e05-44b9-85f2-570a3c0e6016-bundle\") pod \"49c083020ae5dfe237b73a6c6b807501660a323f061d7879268c43a121stbkr\" (UID: \"9f829ff8-1e05-44b9-85f2-570a3c0e6016\") " pod="openstack-operators/49c083020ae5dfe237b73a6c6b807501660a323f061d7879268c43a121stbkr" Dec 01 08:55:45 crc kubenswrapper[4813]: I1201 08:55:45.757813 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wbv4m\" (UniqueName: \"kubernetes.io/projected/9f829ff8-1e05-44b9-85f2-570a3c0e6016-kube-api-access-wbv4m\") pod \"49c083020ae5dfe237b73a6c6b807501660a323f061d7879268c43a121stbkr\" (UID: \"9f829ff8-1e05-44b9-85f2-570a3c0e6016\") " pod="openstack-operators/49c083020ae5dfe237b73a6c6b807501660a323f061d7879268c43a121stbkr" Dec 01 08:55:45 crc kubenswrapper[4813]: I1201 08:55:45.790096 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/49c083020ae5dfe237b73a6c6b807501660a323f061d7879268c43a121stbkr" Dec 01 08:55:46 crc kubenswrapper[4813]: I1201 08:55:46.289107 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/49c083020ae5dfe237b73a6c6b807501660a323f061d7879268c43a121stbkr"] Dec 01 08:55:46 crc kubenswrapper[4813]: W1201 08:55:46.300756 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9f829ff8_1e05_44b9_85f2_570a3c0e6016.slice/crio-9de331c70280d01716d138654e3d83f5292e9c2d863a1f00d60c68702431d42a WatchSource:0}: Error finding container 9de331c70280d01716d138654e3d83f5292e9c2d863a1f00d60c68702431d42a: Status 404 returned error can't find the container with id 9de331c70280d01716d138654e3d83f5292e9c2d863a1f00d60c68702431d42a Dec 01 08:55:46 crc kubenswrapper[4813]: I1201 08:55:46.575268 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/49c083020ae5dfe237b73a6c6b807501660a323f061d7879268c43a121stbkr" event={"ID":"9f829ff8-1e05-44b9-85f2-570a3c0e6016","Type":"ContainerStarted","Data":"6b8b612de5a1e9011c313e31e4406c95924b8b9af3a343a761a6a6d23864b416"} Dec 01 08:55:46 crc kubenswrapper[4813]: I1201 08:55:46.575339 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/49c083020ae5dfe237b73a6c6b807501660a323f061d7879268c43a121stbkr" event={"ID":"9f829ff8-1e05-44b9-85f2-570a3c0e6016","Type":"ContainerStarted","Data":"9de331c70280d01716d138654e3d83f5292e9c2d863a1f00d60c68702431d42a"} Dec 01 08:55:47 crc kubenswrapper[4813]: I1201 08:55:47.237548 4813 patch_prober.go:28] interesting pod/machine-config-daemon-9wn4q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 08:55:47 crc kubenswrapper[4813]: I1201 08:55:47.237634 4813 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9wn4q" podUID="3d423402-e08f-42e9-b60c-6eb9bf2f6d48" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 08:55:47 crc kubenswrapper[4813]: I1201 08:55:47.237689 4813 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-9wn4q" Dec 01 08:55:47 crc kubenswrapper[4813]: I1201 08:55:47.238476 4813 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"bc619037491f7f2216e483229bcdbf7a96de1d588814d7688551705019230c50"} pod="openshift-machine-config-operator/machine-config-daemon-9wn4q" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 01 08:55:47 crc kubenswrapper[4813]: I1201 08:55:47.238563 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-9wn4q" podUID="3d423402-e08f-42e9-b60c-6eb9bf2f6d48" containerName="machine-config-daemon" containerID="cri-o://bc619037491f7f2216e483229bcdbf7a96de1d588814d7688551705019230c50" gracePeriod=600 Dec 01 08:55:47 crc kubenswrapper[4813]: I1201 08:55:47.583441 4813 generic.go:334] "Generic (PLEG): container finished" podID="9f829ff8-1e05-44b9-85f2-570a3c0e6016" containerID="6b8b612de5a1e9011c313e31e4406c95924b8b9af3a343a761a6a6d23864b416" exitCode=0 Dec 01 08:55:47 crc kubenswrapper[4813]: I1201 08:55:47.583551 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/49c083020ae5dfe237b73a6c6b807501660a323f061d7879268c43a121stbkr" event={"ID":"9f829ff8-1e05-44b9-85f2-570a3c0e6016","Type":"ContainerDied","Data":"6b8b612de5a1e9011c313e31e4406c95924b8b9af3a343a761a6a6d23864b416"} Dec 01 08:55:47 crc kubenswrapper[4813]: I1201 08:55:47.588059 4813 generic.go:334] "Generic (PLEG): container finished" podID="3d423402-e08f-42e9-b60c-6eb9bf2f6d48" containerID="bc619037491f7f2216e483229bcdbf7a96de1d588814d7688551705019230c50" exitCode=0 Dec 01 08:55:47 crc kubenswrapper[4813]: I1201 08:55:47.588101 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-9wn4q" event={"ID":"3d423402-e08f-42e9-b60c-6eb9bf2f6d48","Type":"ContainerDied","Data":"bc619037491f7f2216e483229bcdbf7a96de1d588814d7688551705019230c50"} Dec 01 08:55:47 crc kubenswrapper[4813]: I1201 08:55:47.588123 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-9wn4q" event={"ID":"3d423402-e08f-42e9-b60c-6eb9bf2f6d48","Type":"ContainerStarted","Data":"f46458d1e4ce790f7980bd9d4f88fad10c8bcf27bf165b2631ed78f8dc68e1b3"} Dec 01 08:55:47 crc kubenswrapper[4813]: I1201 08:55:47.588165 4813 scope.go:117] "RemoveContainer" containerID="c9bc69c1312504b715819da3ef57a1ae8b88ab55017f5060cf77c80a5fa0c1e8" Dec 01 08:55:49 crc kubenswrapper[4813]: I1201 08:55:49.617587 4813 generic.go:334] "Generic (PLEG): container finished" podID="9f829ff8-1e05-44b9-85f2-570a3c0e6016" containerID="0143f1788b1191f5c1d35c6e1a0986aa9fa3172f2a269ba224e731080c7c7a10" exitCode=0 Dec 01 08:55:49 crc kubenswrapper[4813]: I1201 08:55:49.617690 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/49c083020ae5dfe237b73a6c6b807501660a323f061d7879268c43a121stbkr" event={"ID":"9f829ff8-1e05-44b9-85f2-570a3c0e6016","Type":"ContainerDied","Data":"0143f1788b1191f5c1d35c6e1a0986aa9fa3172f2a269ba224e731080c7c7a10"} Dec 01 08:55:50 crc kubenswrapper[4813]: I1201 08:55:50.629255 4813 generic.go:334] "Generic (PLEG): container finished" podID="9f829ff8-1e05-44b9-85f2-570a3c0e6016" containerID="6eae32d4e0a5b9b7e4f9d222515b1e4685e708027fc970d304e51a7d82dc0ea7" exitCode=0 Dec 01 08:55:50 crc kubenswrapper[4813]: I1201 08:55:50.629326 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/49c083020ae5dfe237b73a6c6b807501660a323f061d7879268c43a121stbkr" event={"ID":"9f829ff8-1e05-44b9-85f2-570a3c0e6016","Type":"ContainerDied","Data":"6eae32d4e0a5b9b7e4f9d222515b1e4685e708027fc970d304e51a7d82dc0ea7"} Dec 01 08:55:51 crc kubenswrapper[4813]: I1201 08:55:51.944687 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/49c083020ae5dfe237b73a6c6b807501660a323f061d7879268c43a121stbkr" Dec 01 08:55:52 crc kubenswrapper[4813]: I1201 08:55:52.138934 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wbv4m\" (UniqueName: \"kubernetes.io/projected/9f829ff8-1e05-44b9-85f2-570a3c0e6016-kube-api-access-wbv4m\") pod \"9f829ff8-1e05-44b9-85f2-570a3c0e6016\" (UID: \"9f829ff8-1e05-44b9-85f2-570a3c0e6016\") " Dec 01 08:55:52 crc kubenswrapper[4813]: I1201 08:55:52.139089 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/9f829ff8-1e05-44b9-85f2-570a3c0e6016-util\") pod \"9f829ff8-1e05-44b9-85f2-570a3c0e6016\" (UID: \"9f829ff8-1e05-44b9-85f2-570a3c0e6016\") " Dec 01 08:55:52 crc kubenswrapper[4813]: I1201 08:55:52.139270 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/9f829ff8-1e05-44b9-85f2-570a3c0e6016-bundle\") pod \"9f829ff8-1e05-44b9-85f2-570a3c0e6016\" (UID: \"9f829ff8-1e05-44b9-85f2-570a3c0e6016\") " Dec 01 08:55:52 crc kubenswrapper[4813]: I1201 08:55:52.140048 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9f829ff8-1e05-44b9-85f2-570a3c0e6016-bundle" (OuterVolumeSpecName: "bundle") pod "9f829ff8-1e05-44b9-85f2-570a3c0e6016" (UID: "9f829ff8-1e05-44b9-85f2-570a3c0e6016"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 08:55:52 crc kubenswrapper[4813]: I1201 08:55:52.148398 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9f829ff8-1e05-44b9-85f2-570a3c0e6016-kube-api-access-wbv4m" (OuterVolumeSpecName: "kube-api-access-wbv4m") pod "9f829ff8-1e05-44b9-85f2-570a3c0e6016" (UID: "9f829ff8-1e05-44b9-85f2-570a3c0e6016"). InnerVolumeSpecName "kube-api-access-wbv4m". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:55:52 crc kubenswrapper[4813]: I1201 08:55:52.151402 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9f829ff8-1e05-44b9-85f2-570a3c0e6016-util" (OuterVolumeSpecName: "util") pod "9f829ff8-1e05-44b9-85f2-570a3c0e6016" (UID: "9f829ff8-1e05-44b9-85f2-570a3c0e6016"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 08:55:52 crc kubenswrapper[4813]: I1201 08:55:52.241217 4813 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/9f829ff8-1e05-44b9-85f2-570a3c0e6016-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 08:55:52 crc kubenswrapper[4813]: I1201 08:55:52.241274 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wbv4m\" (UniqueName: \"kubernetes.io/projected/9f829ff8-1e05-44b9-85f2-570a3c0e6016-kube-api-access-wbv4m\") on node \"crc\" DevicePath \"\"" Dec 01 08:55:52 crc kubenswrapper[4813]: I1201 08:55:52.241297 4813 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/9f829ff8-1e05-44b9-85f2-570a3c0e6016-util\") on node \"crc\" DevicePath \"\"" Dec 01 08:55:52 crc kubenswrapper[4813]: I1201 08:55:52.651496 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/49c083020ae5dfe237b73a6c6b807501660a323f061d7879268c43a121stbkr" event={"ID":"9f829ff8-1e05-44b9-85f2-570a3c0e6016","Type":"ContainerDied","Data":"9de331c70280d01716d138654e3d83f5292e9c2d863a1f00d60c68702431d42a"} Dec 01 08:55:52 crc kubenswrapper[4813]: I1201 08:55:52.651563 4813 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9de331c70280d01716d138654e3d83f5292e9c2d863a1f00d60c68702431d42a" Dec 01 08:55:52 crc kubenswrapper[4813]: I1201 08:55:52.651702 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/49c083020ae5dfe237b73a6c6b807501660a323f061d7879268c43a121stbkr" Dec 01 08:56:02 crc kubenswrapper[4813]: I1201 08:56:02.562685 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7789cf9588-hv6bg"] Dec 01 08:56:02 crc kubenswrapper[4813]: E1201 08:56:02.563661 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9f829ff8-1e05-44b9-85f2-570a3c0e6016" containerName="extract" Dec 01 08:56:02 crc kubenswrapper[4813]: I1201 08:56:02.563704 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="9f829ff8-1e05-44b9-85f2-570a3c0e6016" containerName="extract" Dec 01 08:56:02 crc kubenswrapper[4813]: E1201 08:56:02.563737 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9f829ff8-1e05-44b9-85f2-570a3c0e6016" containerName="pull" Dec 01 08:56:02 crc kubenswrapper[4813]: I1201 08:56:02.563745 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="9f829ff8-1e05-44b9-85f2-570a3c0e6016" containerName="pull" Dec 01 08:56:02 crc kubenswrapper[4813]: E1201 08:56:02.563758 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9f829ff8-1e05-44b9-85f2-570a3c0e6016" containerName="util" Dec 01 08:56:02 crc kubenswrapper[4813]: I1201 08:56:02.563766 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="9f829ff8-1e05-44b9-85f2-570a3c0e6016" containerName="util" Dec 01 08:56:02 crc kubenswrapper[4813]: I1201 08:56:02.564048 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="9f829ff8-1e05-44b9-85f2-570a3c0e6016" containerName="extract" Dec 01 08:56:02 crc kubenswrapper[4813]: I1201 08:56:02.564861 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-7789cf9588-hv6bg" Dec 01 08:56:02 crc kubenswrapper[4813]: I1201 08:56:02.570309 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-dockercfg-vnbvm" Dec 01 08:56:02 crc kubenswrapper[4813]: I1201 08:56:02.571877 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-service-cert" Dec 01 08:56:02 crc kubenswrapper[4813]: I1201 08:56:02.593359 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7789cf9588-hv6bg"] Dec 01 08:56:02 crc kubenswrapper[4813]: I1201 08:56:02.596508 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/2a681edc-eda3-4d33-a642-24898d63a97f-apiservice-cert\") pod \"keystone-operator-controller-manager-7789cf9588-hv6bg\" (UID: \"2a681edc-eda3-4d33-a642-24898d63a97f\") " pod="openstack-operators/keystone-operator-controller-manager-7789cf9588-hv6bg" Dec 01 08:56:02 crc kubenswrapper[4813]: I1201 08:56:02.596564 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/2a681edc-eda3-4d33-a642-24898d63a97f-webhook-cert\") pod \"keystone-operator-controller-manager-7789cf9588-hv6bg\" (UID: \"2a681edc-eda3-4d33-a642-24898d63a97f\") " pod="openstack-operators/keystone-operator-controller-manager-7789cf9588-hv6bg" Dec 01 08:56:02 crc kubenswrapper[4813]: I1201 08:56:02.596711 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9fwkb\" (UniqueName: \"kubernetes.io/projected/2a681edc-eda3-4d33-a642-24898d63a97f-kube-api-access-9fwkb\") pod \"keystone-operator-controller-manager-7789cf9588-hv6bg\" (UID: \"2a681edc-eda3-4d33-a642-24898d63a97f\") " pod="openstack-operators/keystone-operator-controller-manager-7789cf9588-hv6bg" Dec 01 08:56:02 crc kubenswrapper[4813]: I1201 08:56:02.698305 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/2a681edc-eda3-4d33-a642-24898d63a97f-apiservice-cert\") pod \"keystone-operator-controller-manager-7789cf9588-hv6bg\" (UID: \"2a681edc-eda3-4d33-a642-24898d63a97f\") " pod="openstack-operators/keystone-operator-controller-manager-7789cf9588-hv6bg" Dec 01 08:56:02 crc kubenswrapper[4813]: I1201 08:56:02.698390 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/2a681edc-eda3-4d33-a642-24898d63a97f-webhook-cert\") pod \"keystone-operator-controller-manager-7789cf9588-hv6bg\" (UID: \"2a681edc-eda3-4d33-a642-24898d63a97f\") " pod="openstack-operators/keystone-operator-controller-manager-7789cf9588-hv6bg" Dec 01 08:56:02 crc kubenswrapper[4813]: I1201 08:56:02.698470 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9fwkb\" (UniqueName: \"kubernetes.io/projected/2a681edc-eda3-4d33-a642-24898d63a97f-kube-api-access-9fwkb\") pod \"keystone-operator-controller-manager-7789cf9588-hv6bg\" (UID: \"2a681edc-eda3-4d33-a642-24898d63a97f\") " pod="openstack-operators/keystone-operator-controller-manager-7789cf9588-hv6bg" Dec 01 08:56:02 crc kubenswrapper[4813]: I1201 08:56:02.704289 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/2a681edc-eda3-4d33-a642-24898d63a97f-apiservice-cert\") pod \"keystone-operator-controller-manager-7789cf9588-hv6bg\" (UID: \"2a681edc-eda3-4d33-a642-24898d63a97f\") " pod="openstack-operators/keystone-operator-controller-manager-7789cf9588-hv6bg" Dec 01 08:56:02 crc kubenswrapper[4813]: I1201 08:56:02.704336 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/2a681edc-eda3-4d33-a642-24898d63a97f-webhook-cert\") pod \"keystone-operator-controller-manager-7789cf9588-hv6bg\" (UID: \"2a681edc-eda3-4d33-a642-24898d63a97f\") " pod="openstack-operators/keystone-operator-controller-manager-7789cf9588-hv6bg" Dec 01 08:56:02 crc kubenswrapper[4813]: I1201 08:56:02.720611 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9fwkb\" (UniqueName: \"kubernetes.io/projected/2a681edc-eda3-4d33-a642-24898d63a97f-kube-api-access-9fwkb\") pod \"keystone-operator-controller-manager-7789cf9588-hv6bg\" (UID: \"2a681edc-eda3-4d33-a642-24898d63a97f\") " pod="openstack-operators/keystone-operator-controller-manager-7789cf9588-hv6bg" Dec 01 08:56:02 crc kubenswrapper[4813]: I1201 08:56:02.884294 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-7789cf9588-hv6bg" Dec 01 08:56:03 crc kubenswrapper[4813]: I1201 08:56:03.144812 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7789cf9588-hv6bg"] Dec 01 08:56:03 crc kubenswrapper[4813]: W1201 08:56:03.156369 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2a681edc_eda3_4d33_a642_24898d63a97f.slice/crio-667fe2ff11c27639ba0b7a62d37da55fc2d9e3891451cf302e75f1de107f3a6c WatchSource:0}: Error finding container 667fe2ff11c27639ba0b7a62d37da55fc2d9e3891451cf302e75f1de107f3a6c: Status 404 returned error can't find the container with id 667fe2ff11c27639ba0b7a62d37da55fc2d9e3891451cf302e75f1de107f3a6c Dec 01 08:56:03 crc kubenswrapper[4813]: I1201 08:56:03.739208 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7789cf9588-hv6bg" event={"ID":"2a681edc-eda3-4d33-a642-24898d63a97f","Type":"ContainerStarted","Data":"667fe2ff11c27639ba0b7a62d37da55fc2d9e3891451cf302e75f1de107f3a6c"} Dec 01 08:56:05 crc kubenswrapper[4813]: I1201 08:56:05.766241 4813 generic.go:334] "Generic (PLEG): container finished" podID="a274574f-1dfe-4afd-8dfa-661d14669e73" containerID="a70c9b4b1ae94521445b57103c7df87633fe6b94256e1fa57ce08bd340754a02" exitCode=0 Dec 01 08:56:05 crc kubenswrapper[4813]: I1201 08:56:05.766273 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/rabbitmq-server-0" event={"ID":"a274574f-1dfe-4afd-8dfa-661d14669e73","Type":"ContainerDied","Data":"a70c9b4b1ae94521445b57103c7df87633fe6b94256e1fa57ce08bd340754a02"} Dec 01 08:56:07 crc kubenswrapper[4813]: I1201 08:56:07.786695 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7789cf9588-hv6bg" event={"ID":"2a681edc-eda3-4d33-a642-24898d63a97f","Type":"ContainerStarted","Data":"48dbd3ab6f02167f321a00621dc1663966530d13f29261a0304ea32c0e6001c9"} Dec 01 08:56:07 crc kubenswrapper[4813]: I1201 08:56:07.787115 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-7789cf9588-hv6bg" Dec 01 08:56:07 crc kubenswrapper[4813]: I1201 08:56:07.789703 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/rabbitmq-server-0" event={"ID":"a274574f-1dfe-4afd-8dfa-661d14669e73","Type":"ContainerStarted","Data":"e9eef734cf66c90ff9f0f04280a9c11fc9ecb34a8b6da3055227e88543656ffe"} Dec 01 08:56:07 crc kubenswrapper[4813]: I1201 08:56:07.790002 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/rabbitmq-server-0" Dec 01 08:56:07 crc kubenswrapper[4813]: I1201 08:56:07.811704 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-controller-manager-7789cf9588-hv6bg" podStartSLOduration=1.541476236 podStartE2EDuration="5.811641216s" podCreationTimestamp="2025-12-01 08:56:02 +0000 UTC" firstStartedPulling="2025-12-01 08:56:03.160610938 +0000 UTC m=+947.783532524" lastFinishedPulling="2025-12-01 08:56:07.430775918 +0000 UTC m=+952.053697504" observedRunningTime="2025-12-01 08:56:07.808024644 +0000 UTC m=+952.430946240" watchObservedRunningTime="2025-12-01 08:56:07.811641216 +0000 UTC m=+952.434562822" Dec 01 08:56:07 crc kubenswrapper[4813]: I1201 08:56:07.840497 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/rabbitmq-server-0" podStartSLOduration=36.772140269 podStartE2EDuration="42.840470124s" podCreationTimestamp="2025-12-01 08:55:25 +0000 UTC" firstStartedPulling="2025-12-01 08:55:26.802804173 +0000 UTC m=+911.425725759" lastFinishedPulling="2025-12-01 08:55:32.871134028 +0000 UTC m=+917.494055614" observedRunningTime="2025-12-01 08:56:07.831951875 +0000 UTC m=+952.454873471" watchObservedRunningTime="2025-12-01 08:56:07.840470124 +0000 UTC m=+952.463391730" Dec 01 08:56:12 crc kubenswrapper[4813]: I1201 08:56:12.888633 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-7789cf9588-hv6bg" Dec 01 08:56:17 crc kubenswrapper[4813]: I1201 08:56:17.626024 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/keystone-05f1-account-create-update-r6np6"] Dec 01 08:56:17 crc kubenswrapper[4813]: I1201 08:56:17.627409 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-05f1-account-create-update-r6np6" Dec 01 08:56:17 crc kubenswrapper[4813]: I1201 08:56:17.629347 4813 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"keystone-db-secret" Dec 01 08:56:17 crc kubenswrapper[4813]: I1201 08:56:17.638194 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/keystone-05f1-account-create-update-r6np6"] Dec 01 08:56:17 crc kubenswrapper[4813]: I1201 08:56:17.699561 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6xzh2\" (UniqueName: \"kubernetes.io/projected/a9912b4c-f609-4897-bb27-459294633dc7-kube-api-access-6xzh2\") pod \"keystone-05f1-account-create-update-r6np6\" (UID: \"a9912b4c-f609-4897-bb27-459294633dc7\") " pod="glance-kuttl-tests/keystone-05f1-account-create-update-r6np6" Dec 01 08:56:17 crc kubenswrapper[4813]: I1201 08:56:17.699631 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a9912b4c-f609-4897-bb27-459294633dc7-operator-scripts\") pod \"keystone-05f1-account-create-update-r6np6\" (UID: \"a9912b4c-f609-4897-bb27-459294633dc7\") " pod="glance-kuttl-tests/keystone-05f1-account-create-update-r6np6" Dec 01 08:56:17 crc kubenswrapper[4813]: I1201 08:56:17.724275 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/keystone-db-create-lv649"] Dec 01 08:56:17 crc kubenswrapper[4813]: I1201 08:56:17.725299 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-db-create-lv649" Dec 01 08:56:17 crc kubenswrapper[4813]: I1201 08:56:17.738213 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/keystone-db-create-lv649"] Dec 01 08:56:17 crc kubenswrapper[4813]: I1201 08:56:17.769596 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-index-jn2z5"] Dec 01 08:56:17 crc kubenswrapper[4813]: I1201 08:56:17.770381 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-index-jn2z5" Dec 01 08:56:17 crc kubenswrapper[4813]: I1201 08:56:17.775280 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-index-dockercfg-8mszd" Dec 01 08:56:17 crc kubenswrapper[4813]: I1201 08:56:17.776618 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-index-jn2z5"] Dec 01 08:56:17 crc kubenswrapper[4813]: I1201 08:56:17.802326 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hdwq4\" (UniqueName: \"kubernetes.io/projected/387b5367-4ea4-421c-98e9-0c71fca3084c-kube-api-access-hdwq4\") pod \"keystone-db-create-lv649\" (UID: \"387b5367-4ea4-421c-98e9-0c71fca3084c\") " pod="glance-kuttl-tests/keystone-db-create-lv649" Dec 01 08:56:17 crc kubenswrapper[4813]: I1201 08:56:17.802413 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6xzh2\" (UniqueName: \"kubernetes.io/projected/a9912b4c-f609-4897-bb27-459294633dc7-kube-api-access-6xzh2\") pod \"keystone-05f1-account-create-update-r6np6\" (UID: \"a9912b4c-f609-4897-bb27-459294633dc7\") " pod="glance-kuttl-tests/keystone-05f1-account-create-update-r6np6" Dec 01 08:56:17 crc kubenswrapper[4813]: I1201 08:56:17.802479 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-429mb\" (UniqueName: \"kubernetes.io/projected/488d9561-dccc-4b1f-b62f-7af825416efb-kube-api-access-429mb\") pod \"horizon-operator-index-jn2z5\" (UID: \"488d9561-dccc-4b1f-b62f-7af825416efb\") " pod="openstack-operators/horizon-operator-index-jn2z5" Dec 01 08:56:17 crc kubenswrapper[4813]: I1201 08:56:17.802522 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a9912b4c-f609-4897-bb27-459294633dc7-operator-scripts\") pod \"keystone-05f1-account-create-update-r6np6\" (UID: \"a9912b4c-f609-4897-bb27-459294633dc7\") " pod="glance-kuttl-tests/keystone-05f1-account-create-update-r6np6" Dec 01 08:56:17 crc kubenswrapper[4813]: I1201 08:56:17.802543 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/387b5367-4ea4-421c-98e9-0c71fca3084c-operator-scripts\") pod \"keystone-db-create-lv649\" (UID: \"387b5367-4ea4-421c-98e9-0c71fca3084c\") " pod="glance-kuttl-tests/keystone-db-create-lv649" Dec 01 08:56:17 crc kubenswrapper[4813]: I1201 08:56:17.803997 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a9912b4c-f609-4897-bb27-459294633dc7-operator-scripts\") pod \"keystone-05f1-account-create-update-r6np6\" (UID: \"a9912b4c-f609-4897-bb27-459294633dc7\") " pod="glance-kuttl-tests/keystone-05f1-account-create-update-r6np6" Dec 01 08:56:17 crc kubenswrapper[4813]: I1201 08:56:17.820386 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6xzh2\" (UniqueName: \"kubernetes.io/projected/a9912b4c-f609-4897-bb27-459294633dc7-kube-api-access-6xzh2\") pod \"keystone-05f1-account-create-update-r6np6\" (UID: \"a9912b4c-f609-4897-bb27-459294633dc7\") " pod="glance-kuttl-tests/keystone-05f1-account-create-update-r6np6" Dec 01 08:56:17 crc kubenswrapper[4813]: I1201 08:56:17.903893 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/387b5367-4ea4-421c-98e9-0c71fca3084c-operator-scripts\") pod \"keystone-db-create-lv649\" (UID: \"387b5367-4ea4-421c-98e9-0c71fca3084c\") " pod="glance-kuttl-tests/keystone-db-create-lv649" Dec 01 08:56:17 crc kubenswrapper[4813]: I1201 08:56:17.904149 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hdwq4\" (UniqueName: \"kubernetes.io/projected/387b5367-4ea4-421c-98e9-0c71fca3084c-kube-api-access-hdwq4\") pod \"keystone-db-create-lv649\" (UID: \"387b5367-4ea4-421c-98e9-0c71fca3084c\") " pod="glance-kuttl-tests/keystone-db-create-lv649" Dec 01 08:56:17 crc kubenswrapper[4813]: I1201 08:56:17.904297 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-429mb\" (UniqueName: \"kubernetes.io/projected/488d9561-dccc-4b1f-b62f-7af825416efb-kube-api-access-429mb\") pod \"horizon-operator-index-jn2z5\" (UID: \"488d9561-dccc-4b1f-b62f-7af825416efb\") " pod="openstack-operators/horizon-operator-index-jn2z5" Dec 01 08:56:17 crc kubenswrapper[4813]: I1201 08:56:17.904702 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/387b5367-4ea4-421c-98e9-0c71fca3084c-operator-scripts\") pod \"keystone-db-create-lv649\" (UID: \"387b5367-4ea4-421c-98e9-0c71fca3084c\") " pod="glance-kuttl-tests/keystone-db-create-lv649" Dec 01 08:56:17 crc kubenswrapper[4813]: I1201 08:56:17.934452 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hdwq4\" (UniqueName: \"kubernetes.io/projected/387b5367-4ea4-421c-98e9-0c71fca3084c-kube-api-access-hdwq4\") pod \"keystone-db-create-lv649\" (UID: \"387b5367-4ea4-421c-98e9-0c71fca3084c\") " pod="glance-kuttl-tests/keystone-db-create-lv649" Dec 01 08:56:17 crc kubenswrapper[4813]: I1201 08:56:17.936002 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-429mb\" (UniqueName: \"kubernetes.io/projected/488d9561-dccc-4b1f-b62f-7af825416efb-kube-api-access-429mb\") pod \"horizon-operator-index-jn2z5\" (UID: \"488d9561-dccc-4b1f-b62f-7af825416efb\") " pod="openstack-operators/horizon-operator-index-jn2z5" Dec 01 08:56:17 crc kubenswrapper[4813]: I1201 08:56:17.945720 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-05f1-account-create-update-r6np6" Dec 01 08:56:18 crc kubenswrapper[4813]: I1201 08:56:18.042603 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-db-create-lv649" Dec 01 08:56:18 crc kubenswrapper[4813]: I1201 08:56:18.090109 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-index-jn2z5" Dec 01 08:56:18 crc kubenswrapper[4813]: I1201 08:56:18.380789 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/keystone-05f1-account-create-update-r6np6"] Dec 01 08:56:18 crc kubenswrapper[4813]: I1201 08:56:18.531897 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/keystone-db-create-lv649"] Dec 01 08:56:18 crc kubenswrapper[4813]: W1201 08:56:18.537652 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod387b5367_4ea4_421c_98e9_0c71fca3084c.slice/crio-d2e3917ddd358066c7a93902a5d1bdda5f6f06f0bb9e03dea72f117f67efa796 WatchSource:0}: Error finding container d2e3917ddd358066c7a93902a5d1bdda5f6f06f0bb9e03dea72f117f67efa796: Status 404 returned error can't find the container with id d2e3917ddd358066c7a93902a5d1bdda5f6f06f0bb9e03dea72f117f67efa796 Dec 01 08:56:18 crc kubenswrapper[4813]: I1201 08:56:18.540361 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-index-jn2z5"] Dec 01 08:56:18 crc kubenswrapper[4813]: W1201 08:56:18.553914 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod488d9561_dccc_4b1f_b62f_7af825416efb.slice/crio-0d541a32d927bd86245705400b49439edbe5d8459555a548c7b4b454d3a56cf8 WatchSource:0}: Error finding container 0d541a32d927bd86245705400b49439edbe5d8459555a548c7b4b454d3a56cf8: Status 404 returned error can't find the container with id 0d541a32d927bd86245705400b49439edbe5d8459555a548c7b4b454d3a56cf8 Dec 01 08:56:18 crc kubenswrapper[4813]: I1201 08:56:18.868006 4813 generic.go:334] "Generic (PLEG): container finished" podID="387b5367-4ea4-421c-98e9-0c71fca3084c" containerID="eac1c2f1c7e233613b531497d81712a4b61aed431ac7cad1204910e5e2633b87" exitCode=0 Dec 01 08:56:18 crc kubenswrapper[4813]: I1201 08:56:18.868072 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-db-create-lv649" event={"ID":"387b5367-4ea4-421c-98e9-0c71fca3084c","Type":"ContainerDied","Data":"eac1c2f1c7e233613b531497d81712a4b61aed431ac7cad1204910e5e2633b87"} Dec 01 08:56:18 crc kubenswrapper[4813]: I1201 08:56:18.868103 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-db-create-lv649" event={"ID":"387b5367-4ea4-421c-98e9-0c71fca3084c","Type":"ContainerStarted","Data":"d2e3917ddd358066c7a93902a5d1bdda5f6f06f0bb9e03dea72f117f67efa796"} Dec 01 08:56:18 crc kubenswrapper[4813]: I1201 08:56:18.869487 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-index-jn2z5" event={"ID":"488d9561-dccc-4b1f-b62f-7af825416efb","Type":"ContainerStarted","Data":"0d541a32d927bd86245705400b49439edbe5d8459555a548c7b4b454d3a56cf8"} Dec 01 08:56:18 crc kubenswrapper[4813]: I1201 08:56:18.871224 4813 generic.go:334] "Generic (PLEG): container finished" podID="a9912b4c-f609-4897-bb27-459294633dc7" containerID="c77365be1f5b03ab4bf99c3ffbe4da8a4e301ddcea7a11b9f74f9f10847510e6" exitCode=0 Dec 01 08:56:18 crc kubenswrapper[4813]: I1201 08:56:18.871303 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-05f1-account-create-update-r6np6" event={"ID":"a9912b4c-f609-4897-bb27-459294633dc7","Type":"ContainerDied","Data":"c77365be1f5b03ab4bf99c3ffbe4da8a4e301ddcea7a11b9f74f9f10847510e6"} Dec 01 08:56:18 crc kubenswrapper[4813]: I1201 08:56:18.871352 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-05f1-account-create-update-r6np6" event={"ID":"a9912b4c-f609-4897-bb27-459294633dc7","Type":"ContainerStarted","Data":"92ee0e0f359078410e0824bf536f6dcdb82e04cc8da46b307a6cd20ae2377759"} Dec 01 08:56:19 crc kubenswrapper[4813]: I1201 08:56:19.881483 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-index-jn2z5" event={"ID":"488d9561-dccc-4b1f-b62f-7af825416efb","Type":"ContainerStarted","Data":"56b820164aa9ecb9eb292b5ec0d593d2085aae0f108fc1d8c09e93a315846247"} Dec 01 08:56:19 crc kubenswrapper[4813]: I1201 08:56:19.901647 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-index-jn2z5" podStartSLOduration=2.087818461 podStartE2EDuration="2.901622077s" podCreationTimestamp="2025-12-01 08:56:17 +0000 UTC" firstStartedPulling="2025-12-01 08:56:18.566227797 +0000 UTC m=+963.189149383" lastFinishedPulling="2025-12-01 08:56:19.380031413 +0000 UTC m=+964.002952999" observedRunningTime="2025-12-01 08:56:19.901178164 +0000 UTC m=+964.524099820" watchObservedRunningTime="2025-12-01 08:56:19.901622077 +0000 UTC m=+964.524543663" Dec 01 08:56:20 crc kubenswrapper[4813]: I1201 08:56:20.265912 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-05f1-account-create-update-r6np6" Dec 01 08:56:20 crc kubenswrapper[4813]: I1201 08:56:20.271379 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-db-create-lv649" Dec 01 08:56:20 crc kubenswrapper[4813]: I1201 08:56:20.441472 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a9912b4c-f609-4897-bb27-459294633dc7-operator-scripts\") pod \"a9912b4c-f609-4897-bb27-459294633dc7\" (UID: \"a9912b4c-f609-4897-bb27-459294633dc7\") " Dec 01 08:56:20 crc kubenswrapper[4813]: I1201 08:56:20.441531 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6xzh2\" (UniqueName: \"kubernetes.io/projected/a9912b4c-f609-4897-bb27-459294633dc7-kube-api-access-6xzh2\") pod \"a9912b4c-f609-4897-bb27-459294633dc7\" (UID: \"a9912b4c-f609-4897-bb27-459294633dc7\") " Dec 01 08:56:20 crc kubenswrapper[4813]: I1201 08:56:20.441621 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/387b5367-4ea4-421c-98e9-0c71fca3084c-operator-scripts\") pod \"387b5367-4ea4-421c-98e9-0c71fca3084c\" (UID: \"387b5367-4ea4-421c-98e9-0c71fca3084c\") " Dec 01 08:56:20 crc kubenswrapper[4813]: I1201 08:56:20.441673 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hdwq4\" (UniqueName: \"kubernetes.io/projected/387b5367-4ea4-421c-98e9-0c71fca3084c-kube-api-access-hdwq4\") pod \"387b5367-4ea4-421c-98e9-0c71fca3084c\" (UID: \"387b5367-4ea4-421c-98e9-0c71fca3084c\") " Dec 01 08:56:20 crc kubenswrapper[4813]: I1201 08:56:20.442432 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a9912b4c-f609-4897-bb27-459294633dc7-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "a9912b4c-f609-4897-bb27-459294633dc7" (UID: "a9912b4c-f609-4897-bb27-459294633dc7"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:56:20 crc kubenswrapper[4813]: I1201 08:56:20.442540 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/387b5367-4ea4-421c-98e9-0c71fca3084c-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "387b5367-4ea4-421c-98e9-0c71fca3084c" (UID: "387b5367-4ea4-421c-98e9-0c71fca3084c"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:56:20 crc kubenswrapper[4813]: I1201 08:56:20.448539 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/387b5367-4ea4-421c-98e9-0c71fca3084c-kube-api-access-hdwq4" (OuterVolumeSpecName: "kube-api-access-hdwq4") pod "387b5367-4ea4-421c-98e9-0c71fca3084c" (UID: "387b5367-4ea4-421c-98e9-0c71fca3084c"). InnerVolumeSpecName "kube-api-access-hdwq4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:56:20 crc kubenswrapper[4813]: I1201 08:56:20.449517 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a9912b4c-f609-4897-bb27-459294633dc7-kube-api-access-6xzh2" (OuterVolumeSpecName: "kube-api-access-6xzh2") pod "a9912b4c-f609-4897-bb27-459294633dc7" (UID: "a9912b4c-f609-4897-bb27-459294633dc7"). InnerVolumeSpecName "kube-api-access-6xzh2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:56:20 crc kubenswrapper[4813]: I1201 08:56:20.544056 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hdwq4\" (UniqueName: \"kubernetes.io/projected/387b5367-4ea4-421c-98e9-0c71fca3084c-kube-api-access-hdwq4\") on node \"crc\" DevicePath \"\"" Dec 01 08:56:20 crc kubenswrapper[4813]: I1201 08:56:20.544520 4813 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a9912b4c-f609-4897-bb27-459294633dc7-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 08:56:20 crc kubenswrapper[4813]: I1201 08:56:20.544583 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6xzh2\" (UniqueName: \"kubernetes.io/projected/a9912b4c-f609-4897-bb27-459294633dc7-kube-api-access-6xzh2\") on node \"crc\" DevicePath \"\"" Dec 01 08:56:20 crc kubenswrapper[4813]: I1201 08:56:20.544660 4813 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/387b5367-4ea4-421c-98e9-0c71fca3084c-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 08:56:20 crc kubenswrapper[4813]: I1201 08:56:20.888795 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-05f1-account-create-update-r6np6" event={"ID":"a9912b4c-f609-4897-bb27-459294633dc7","Type":"ContainerDied","Data":"92ee0e0f359078410e0824bf536f6dcdb82e04cc8da46b307a6cd20ae2377759"} Dec 01 08:56:20 crc kubenswrapper[4813]: I1201 08:56:20.888897 4813 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="92ee0e0f359078410e0824bf536f6dcdb82e04cc8da46b307a6cd20ae2377759" Dec 01 08:56:20 crc kubenswrapper[4813]: I1201 08:56:20.888916 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-05f1-account-create-update-r6np6" Dec 01 08:56:20 crc kubenswrapper[4813]: I1201 08:56:20.891674 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-db-create-lv649" Dec 01 08:56:20 crc kubenswrapper[4813]: I1201 08:56:20.891660 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-db-create-lv649" event={"ID":"387b5367-4ea4-421c-98e9-0c71fca3084c","Type":"ContainerDied","Data":"d2e3917ddd358066c7a93902a5d1bdda5f6f06f0bb9e03dea72f117f67efa796"} Dec 01 08:56:20 crc kubenswrapper[4813]: I1201 08:56:20.891725 4813 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d2e3917ddd358066c7a93902a5d1bdda5f6f06f0bb9e03dea72f117f67efa796" Dec 01 08:56:21 crc kubenswrapper[4813]: I1201 08:56:21.365302 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-index-vkcnv"] Dec 01 08:56:21 crc kubenswrapper[4813]: E1201 08:56:21.365848 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a9912b4c-f609-4897-bb27-459294633dc7" containerName="mariadb-account-create-update" Dec 01 08:56:21 crc kubenswrapper[4813]: I1201 08:56:21.365864 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="a9912b4c-f609-4897-bb27-459294633dc7" containerName="mariadb-account-create-update" Dec 01 08:56:21 crc kubenswrapper[4813]: E1201 08:56:21.365875 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="387b5367-4ea4-421c-98e9-0c71fca3084c" containerName="mariadb-database-create" Dec 01 08:56:21 crc kubenswrapper[4813]: I1201 08:56:21.365883 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="387b5367-4ea4-421c-98e9-0c71fca3084c" containerName="mariadb-database-create" Dec 01 08:56:21 crc kubenswrapper[4813]: I1201 08:56:21.366131 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="a9912b4c-f609-4897-bb27-459294633dc7" containerName="mariadb-account-create-update" Dec 01 08:56:21 crc kubenswrapper[4813]: I1201 08:56:21.366164 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="387b5367-4ea4-421c-98e9-0c71fca3084c" containerName="mariadb-database-create" Dec 01 08:56:21 crc kubenswrapper[4813]: I1201 08:56:21.368613 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-index-vkcnv" Dec 01 08:56:21 crc kubenswrapper[4813]: I1201 08:56:21.371030 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-index-dockercfg-4hjzt" Dec 01 08:56:21 crc kubenswrapper[4813]: I1201 08:56:21.374107 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-index-vkcnv"] Dec 01 08:56:21 crc kubenswrapper[4813]: I1201 08:56:21.558643 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zpvwb\" (UniqueName: \"kubernetes.io/projected/174d5e2e-9c14-48e0-af03-91a15c96ad66-kube-api-access-zpvwb\") pod \"swift-operator-index-vkcnv\" (UID: \"174d5e2e-9c14-48e0-af03-91a15c96ad66\") " pod="openstack-operators/swift-operator-index-vkcnv" Dec 01 08:56:21 crc kubenswrapper[4813]: I1201 08:56:21.660375 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zpvwb\" (UniqueName: \"kubernetes.io/projected/174d5e2e-9c14-48e0-af03-91a15c96ad66-kube-api-access-zpvwb\") pod \"swift-operator-index-vkcnv\" (UID: \"174d5e2e-9c14-48e0-af03-91a15c96ad66\") " pod="openstack-operators/swift-operator-index-vkcnv" Dec 01 08:56:21 crc kubenswrapper[4813]: I1201 08:56:21.682081 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zpvwb\" (UniqueName: \"kubernetes.io/projected/174d5e2e-9c14-48e0-af03-91a15c96ad66-kube-api-access-zpvwb\") pod \"swift-operator-index-vkcnv\" (UID: \"174d5e2e-9c14-48e0-af03-91a15c96ad66\") " pod="openstack-operators/swift-operator-index-vkcnv" Dec 01 08:56:21 crc kubenswrapper[4813]: I1201 08:56:21.694863 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-index-vkcnv" Dec 01 08:56:22 crc kubenswrapper[4813]: I1201 08:56:22.220522 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-index-vkcnv"] Dec 01 08:56:22 crc kubenswrapper[4813]: I1201 08:56:22.918600 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-index-vkcnv" event={"ID":"174d5e2e-9c14-48e0-af03-91a15c96ad66","Type":"ContainerStarted","Data":"911e826c96baa80392dfebaeb15f4b76cc73cf02b7afdaca79a2125cc94a17c1"} Dec 01 08:56:23 crc kubenswrapper[4813]: I1201 08:56:23.927486 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-index-vkcnv" event={"ID":"174d5e2e-9c14-48e0-af03-91a15c96ad66","Type":"ContainerStarted","Data":"f9c5ee37aef3819d0bdc31bfaff6bdcdb7bcbce9deff27949a932cf2c0e39352"} Dec 01 08:56:23 crc kubenswrapper[4813]: I1201 08:56:23.948847 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-index-vkcnv" podStartSLOduration=1.936487074 podStartE2EDuration="2.948824796s" podCreationTimestamp="2025-12-01 08:56:21 +0000 UTC" firstStartedPulling="2025-12-01 08:56:22.223033461 +0000 UTC m=+966.845955067" lastFinishedPulling="2025-12-01 08:56:23.235371163 +0000 UTC m=+967.858292789" observedRunningTime="2025-12-01 08:56:23.944384862 +0000 UTC m=+968.567306448" watchObservedRunningTime="2025-12-01 08:56:23.948824796 +0000 UTC m=+968.571746382" Dec 01 08:56:26 crc kubenswrapper[4813]: I1201 08:56:26.371798 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/rabbitmq-server-0" Dec 01 08:56:26 crc kubenswrapper[4813]: I1201 08:56:26.956839 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/keystone-db-sync-gjdp8"] Dec 01 08:56:26 crc kubenswrapper[4813]: I1201 08:56:26.958093 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-db-sync-gjdp8" Dec 01 08:56:26 crc kubenswrapper[4813]: I1201 08:56:26.960414 4813 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"keystone-scripts" Dec 01 08:56:26 crc kubenswrapper[4813]: I1201 08:56:26.964541 4813 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"keystone-keystone-dockercfg-cwrjn" Dec 01 08:56:26 crc kubenswrapper[4813]: I1201 08:56:26.964598 4813 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"keystone" Dec 01 08:56:26 crc kubenswrapper[4813]: I1201 08:56:26.965014 4813 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"keystone-config-data" Dec 01 08:56:26 crc kubenswrapper[4813]: I1201 08:56:26.968267 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/keystone-db-sync-gjdp8"] Dec 01 08:56:27 crc kubenswrapper[4813]: I1201 08:56:27.156790 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dqdhj\" (UniqueName: \"kubernetes.io/projected/6edb6678-d60d-4e1e-92aa-5018a9c1c58d-kube-api-access-dqdhj\") pod \"keystone-db-sync-gjdp8\" (UID: \"6edb6678-d60d-4e1e-92aa-5018a9c1c58d\") " pod="glance-kuttl-tests/keystone-db-sync-gjdp8" Dec 01 08:56:27 crc kubenswrapper[4813]: I1201 08:56:27.157172 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6edb6678-d60d-4e1e-92aa-5018a9c1c58d-config-data\") pod \"keystone-db-sync-gjdp8\" (UID: \"6edb6678-d60d-4e1e-92aa-5018a9c1c58d\") " pod="glance-kuttl-tests/keystone-db-sync-gjdp8" Dec 01 08:56:27 crc kubenswrapper[4813]: I1201 08:56:27.258135 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dqdhj\" (UniqueName: \"kubernetes.io/projected/6edb6678-d60d-4e1e-92aa-5018a9c1c58d-kube-api-access-dqdhj\") pod \"keystone-db-sync-gjdp8\" (UID: \"6edb6678-d60d-4e1e-92aa-5018a9c1c58d\") " pod="glance-kuttl-tests/keystone-db-sync-gjdp8" Dec 01 08:56:27 crc kubenswrapper[4813]: I1201 08:56:27.258221 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6edb6678-d60d-4e1e-92aa-5018a9c1c58d-config-data\") pod \"keystone-db-sync-gjdp8\" (UID: \"6edb6678-d60d-4e1e-92aa-5018a9c1c58d\") " pod="glance-kuttl-tests/keystone-db-sync-gjdp8" Dec 01 08:56:27 crc kubenswrapper[4813]: I1201 08:56:27.265047 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6edb6678-d60d-4e1e-92aa-5018a9c1c58d-config-data\") pod \"keystone-db-sync-gjdp8\" (UID: \"6edb6678-d60d-4e1e-92aa-5018a9c1c58d\") " pod="glance-kuttl-tests/keystone-db-sync-gjdp8" Dec 01 08:56:27 crc kubenswrapper[4813]: I1201 08:56:27.294161 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dqdhj\" (UniqueName: \"kubernetes.io/projected/6edb6678-d60d-4e1e-92aa-5018a9c1c58d-kube-api-access-dqdhj\") pod \"keystone-db-sync-gjdp8\" (UID: \"6edb6678-d60d-4e1e-92aa-5018a9c1c58d\") " pod="glance-kuttl-tests/keystone-db-sync-gjdp8" Dec 01 08:56:27 crc kubenswrapper[4813]: I1201 08:56:27.575692 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-db-sync-gjdp8" Dec 01 08:56:27 crc kubenswrapper[4813]: I1201 08:56:27.993493 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/keystone-db-sync-gjdp8"] Dec 01 08:56:28 crc kubenswrapper[4813]: I1201 08:56:28.002195 4813 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 01 08:56:28 crc kubenswrapper[4813]: I1201 08:56:28.091010 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/horizon-operator-index-jn2z5" Dec 01 08:56:28 crc kubenswrapper[4813]: I1201 08:56:28.091069 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-index-jn2z5" Dec 01 08:56:28 crc kubenswrapper[4813]: I1201 08:56:28.118476 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/horizon-operator-index-jn2z5" Dec 01 08:56:29 crc kubenswrapper[4813]: I1201 08:56:29.013288 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-db-sync-gjdp8" event={"ID":"6edb6678-d60d-4e1e-92aa-5018a9c1c58d","Type":"ContainerStarted","Data":"a4e36f66500f0c63138cb380060fdd1d5420c363dd05eb2a625d84d943d8b1cd"} Dec 01 08:56:29 crc kubenswrapper[4813]: I1201 08:56:29.045418 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-index-jn2z5" Dec 01 08:56:31 crc kubenswrapper[4813]: I1201 08:56:31.695883 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-index-vkcnv" Dec 01 08:56:31 crc kubenswrapper[4813]: I1201 08:56:31.696243 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/swift-operator-index-vkcnv" Dec 01 08:56:31 crc kubenswrapper[4813]: I1201 08:56:31.729096 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/swift-operator-index-vkcnv" Dec 01 08:56:32 crc kubenswrapper[4813]: I1201 08:56:32.065469 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-index-vkcnv" Dec 01 08:56:36 crc kubenswrapper[4813]: I1201 08:56:36.413250 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368cx94s5"] Dec 01 08:56:36 crc kubenswrapper[4813]: I1201 08:56:36.416251 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368cx94s5"] Dec 01 08:56:36 crc kubenswrapper[4813]: I1201 08:56:36.416532 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368cx94s5" Dec 01 08:56:36 crc kubenswrapper[4813]: I1201 08:56:36.418726 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/071eb401-2f12-41e4-b52e-312ca4152da5-util\") pod \"9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368cx94s5\" (UID: \"071eb401-2f12-41e4-b52e-312ca4152da5\") " pod="openstack-operators/9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368cx94s5" Dec 01 08:56:36 crc kubenswrapper[4813]: I1201 08:56:36.418787 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/071eb401-2f12-41e4-b52e-312ca4152da5-bundle\") pod \"9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368cx94s5\" (UID: \"071eb401-2f12-41e4-b52e-312ca4152da5\") " pod="openstack-operators/9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368cx94s5" Dec 01 08:56:36 crc kubenswrapper[4813]: I1201 08:56:36.418824 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xn5qf\" (UniqueName: \"kubernetes.io/projected/071eb401-2f12-41e4-b52e-312ca4152da5-kube-api-access-xn5qf\") pod \"9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368cx94s5\" (UID: \"071eb401-2f12-41e4-b52e-312ca4152da5\") " pod="openstack-operators/9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368cx94s5" Dec 01 08:56:36 crc kubenswrapper[4813]: I1201 08:56:36.419350 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-fml9r" Dec 01 08:56:36 crc kubenswrapper[4813]: I1201 08:56:36.520180 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/071eb401-2f12-41e4-b52e-312ca4152da5-util\") pod \"9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368cx94s5\" (UID: \"071eb401-2f12-41e4-b52e-312ca4152da5\") " pod="openstack-operators/9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368cx94s5" Dec 01 08:56:36 crc kubenswrapper[4813]: I1201 08:56:36.520223 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/071eb401-2f12-41e4-b52e-312ca4152da5-bundle\") pod \"9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368cx94s5\" (UID: \"071eb401-2f12-41e4-b52e-312ca4152da5\") " pod="openstack-operators/9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368cx94s5" Dec 01 08:56:36 crc kubenswrapper[4813]: I1201 08:56:36.520256 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xn5qf\" (UniqueName: \"kubernetes.io/projected/071eb401-2f12-41e4-b52e-312ca4152da5-kube-api-access-xn5qf\") pod \"9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368cx94s5\" (UID: \"071eb401-2f12-41e4-b52e-312ca4152da5\") " pod="openstack-operators/9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368cx94s5" Dec 01 08:56:36 crc kubenswrapper[4813]: I1201 08:56:36.521088 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/071eb401-2f12-41e4-b52e-312ca4152da5-bundle\") pod \"9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368cx94s5\" (UID: \"071eb401-2f12-41e4-b52e-312ca4152da5\") " pod="openstack-operators/9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368cx94s5" Dec 01 08:56:36 crc kubenswrapper[4813]: I1201 08:56:36.521187 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/071eb401-2f12-41e4-b52e-312ca4152da5-util\") pod \"9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368cx94s5\" (UID: \"071eb401-2f12-41e4-b52e-312ca4152da5\") " pod="openstack-operators/9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368cx94s5" Dec 01 08:56:36 crc kubenswrapper[4813]: I1201 08:56:36.537583 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xn5qf\" (UniqueName: \"kubernetes.io/projected/071eb401-2f12-41e4-b52e-312ca4152da5-kube-api-access-xn5qf\") pod \"9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368cx94s5\" (UID: \"071eb401-2f12-41e4-b52e-312ca4152da5\") " pod="openstack-operators/9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368cx94s5" Dec 01 08:56:36 crc kubenswrapper[4813]: I1201 08:56:36.796887 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368cx94s5" Dec 01 08:56:37 crc kubenswrapper[4813]: I1201 08:56:37.421941 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240btqv7m"] Dec 01 08:56:37 crc kubenswrapper[4813]: I1201 08:56:37.423868 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240btqv7m" Dec 01 08:56:37 crc kubenswrapper[4813]: I1201 08:56:37.429394 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240btqv7m"] Dec 01 08:56:37 crc kubenswrapper[4813]: I1201 08:56:37.533515 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xdkhm\" (UniqueName: \"kubernetes.io/projected/323401ed-8df8-4341-8d07-ab27d5162fbd-kube-api-access-xdkhm\") pod \"87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240btqv7m\" (UID: \"323401ed-8df8-4341-8d07-ab27d5162fbd\") " pod="openstack-operators/87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240btqv7m" Dec 01 08:56:37 crc kubenswrapper[4813]: I1201 08:56:37.533603 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/323401ed-8df8-4341-8d07-ab27d5162fbd-bundle\") pod \"87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240btqv7m\" (UID: \"323401ed-8df8-4341-8d07-ab27d5162fbd\") " pod="openstack-operators/87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240btqv7m" Dec 01 08:56:37 crc kubenswrapper[4813]: I1201 08:56:37.533737 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/323401ed-8df8-4341-8d07-ab27d5162fbd-util\") pod \"87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240btqv7m\" (UID: \"323401ed-8df8-4341-8d07-ab27d5162fbd\") " pod="openstack-operators/87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240btqv7m" Dec 01 08:56:37 crc kubenswrapper[4813]: I1201 08:56:37.635896 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xdkhm\" (UniqueName: \"kubernetes.io/projected/323401ed-8df8-4341-8d07-ab27d5162fbd-kube-api-access-xdkhm\") pod \"87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240btqv7m\" (UID: \"323401ed-8df8-4341-8d07-ab27d5162fbd\") " pod="openstack-operators/87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240btqv7m" Dec 01 08:56:37 crc kubenswrapper[4813]: I1201 08:56:37.636216 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/323401ed-8df8-4341-8d07-ab27d5162fbd-bundle\") pod \"87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240btqv7m\" (UID: \"323401ed-8df8-4341-8d07-ab27d5162fbd\") " pod="openstack-operators/87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240btqv7m" Dec 01 08:56:37 crc kubenswrapper[4813]: I1201 08:56:37.636274 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/323401ed-8df8-4341-8d07-ab27d5162fbd-util\") pod \"87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240btqv7m\" (UID: \"323401ed-8df8-4341-8d07-ab27d5162fbd\") " pod="openstack-operators/87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240btqv7m" Dec 01 08:56:37 crc kubenswrapper[4813]: I1201 08:56:37.636803 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/323401ed-8df8-4341-8d07-ab27d5162fbd-bundle\") pod \"87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240btqv7m\" (UID: \"323401ed-8df8-4341-8d07-ab27d5162fbd\") " pod="openstack-operators/87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240btqv7m" Dec 01 08:56:37 crc kubenswrapper[4813]: I1201 08:56:37.636863 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/323401ed-8df8-4341-8d07-ab27d5162fbd-util\") pod \"87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240btqv7m\" (UID: \"323401ed-8df8-4341-8d07-ab27d5162fbd\") " pod="openstack-operators/87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240btqv7m" Dec 01 08:56:37 crc kubenswrapper[4813]: I1201 08:56:37.657803 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xdkhm\" (UniqueName: \"kubernetes.io/projected/323401ed-8df8-4341-8d07-ab27d5162fbd-kube-api-access-xdkhm\") pod \"87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240btqv7m\" (UID: \"323401ed-8df8-4341-8d07-ab27d5162fbd\") " pod="openstack-operators/87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240btqv7m" Dec 01 08:56:37 crc kubenswrapper[4813]: I1201 08:56:37.739869 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240btqv7m" Dec 01 08:56:39 crc kubenswrapper[4813]: I1201 08:56:39.116699 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-db-sync-gjdp8" event={"ID":"6edb6678-d60d-4e1e-92aa-5018a9c1c58d","Type":"ContainerStarted","Data":"45a618c9430ab1f255c16e35c6c18bd5d07f8286d9619ee72c3673628cc81da7"} Dec 01 08:56:39 crc kubenswrapper[4813]: I1201 08:56:39.136681 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/keystone-db-sync-gjdp8" podStartSLOduration=2.222048282 podStartE2EDuration="13.136616388s" podCreationTimestamp="2025-12-01 08:56:26 +0000 UTC" firstStartedPulling="2025-12-01 08:56:28.001800187 +0000 UTC m=+972.624721773" lastFinishedPulling="2025-12-01 08:56:38.916368293 +0000 UTC m=+983.539289879" observedRunningTime="2025-12-01 08:56:39.12883614 +0000 UTC m=+983.751757746" watchObservedRunningTime="2025-12-01 08:56:39.136616388 +0000 UTC m=+983.759537994" Dec 01 08:56:39 crc kubenswrapper[4813]: I1201 08:56:39.303162 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368cx94s5"] Dec 01 08:56:39 crc kubenswrapper[4813]: W1201 08:56:39.312675 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod071eb401_2f12_41e4_b52e_312ca4152da5.slice/crio-ec1f5ef5c5f0ac817fb3b72de3ef1e1835929b884abb19cc7b49e30f81b1d134 WatchSource:0}: Error finding container ec1f5ef5c5f0ac817fb3b72de3ef1e1835929b884abb19cc7b49e30f81b1d134: Status 404 returned error can't find the container with id ec1f5ef5c5f0ac817fb3b72de3ef1e1835929b884abb19cc7b49e30f81b1d134 Dec 01 08:56:39 crc kubenswrapper[4813]: I1201 08:56:39.332283 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240btqv7m"] Dec 01 08:56:39 crc kubenswrapper[4813]: W1201 08:56:39.339805 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod323401ed_8df8_4341_8d07_ab27d5162fbd.slice/crio-28c89dd0520e03bbcc3600cdf623348d5412ab3265c099940a9337bfe57c85bf WatchSource:0}: Error finding container 28c89dd0520e03bbcc3600cdf623348d5412ab3265c099940a9337bfe57c85bf: Status 404 returned error can't find the container with id 28c89dd0520e03bbcc3600cdf623348d5412ab3265c099940a9337bfe57c85bf Dec 01 08:56:40 crc kubenswrapper[4813]: I1201 08:56:40.131991 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368cx94s5" event={"ID":"071eb401-2f12-41e4-b52e-312ca4152da5","Type":"ContainerStarted","Data":"545ad5db9ae89f77a4a89cc08c55c301ea9bfcbd6b5484bbf5c0087be00d76d0"} Dec 01 08:56:40 crc kubenswrapper[4813]: I1201 08:56:40.132047 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368cx94s5" event={"ID":"071eb401-2f12-41e4-b52e-312ca4152da5","Type":"ContainerStarted","Data":"ec1f5ef5c5f0ac817fb3b72de3ef1e1835929b884abb19cc7b49e30f81b1d134"} Dec 01 08:56:40 crc kubenswrapper[4813]: I1201 08:56:40.136367 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240btqv7m" event={"ID":"323401ed-8df8-4341-8d07-ab27d5162fbd","Type":"ContainerStarted","Data":"1919ffcab60a2c9578329d6abb86defb8b5f6220be98ab8d091ef47b8e44420f"} Dec 01 08:56:40 crc kubenswrapper[4813]: I1201 08:56:40.136517 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240btqv7m" event={"ID":"323401ed-8df8-4341-8d07-ab27d5162fbd","Type":"ContainerStarted","Data":"28c89dd0520e03bbcc3600cdf623348d5412ab3265c099940a9337bfe57c85bf"} Dec 01 08:56:41 crc kubenswrapper[4813]: I1201 08:56:41.146046 4813 generic.go:334] "Generic (PLEG): container finished" podID="071eb401-2f12-41e4-b52e-312ca4152da5" containerID="545ad5db9ae89f77a4a89cc08c55c301ea9bfcbd6b5484bbf5c0087be00d76d0" exitCode=0 Dec 01 08:56:41 crc kubenswrapper[4813]: I1201 08:56:41.146057 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368cx94s5" event={"ID":"071eb401-2f12-41e4-b52e-312ca4152da5","Type":"ContainerDied","Data":"545ad5db9ae89f77a4a89cc08c55c301ea9bfcbd6b5484bbf5c0087be00d76d0"} Dec 01 08:56:42 crc kubenswrapper[4813]: I1201 08:56:42.157614 4813 generic.go:334] "Generic (PLEG): container finished" podID="323401ed-8df8-4341-8d07-ab27d5162fbd" containerID="1919ffcab60a2c9578329d6abb86defb8b5f6220be98ab8d091ef47b8e44420f" exitCode=0 Dec 01 08:56:42 crc kubenswrapper[4813]: I1201 08:56:42.157721 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240btqv7m" event={"ID":"323401ed-8df8-4341-8d07-ab27d5162fbd","Type":"ContainerDied","Data":"1919ffcab60a2c9578329d6abb86defb8b5f6220be98ab8d091ef47b8e44420f"} Dec 01 08:56:44 crc kubenswrapper[4813]: I1201 08:56:44.177383 4813 generic.go:334] "Generic (PLEG): container finished" podID="071eb401-2f12-41e4-b52e-312ca4152da5" containerID="b73d6c92fdbaa26bf40a058f166397dbf2cdf6371270317dbdcc0af7775b4813" exitCode=0 Dec 01 08:56:44 crc kubenswrapper[4813]: I1201 08:56:44.177459 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368cx94s5" event={"ID":"071eb401-2f12-41e4-b52e-312ca4152da5","Type":"ContainerDied","Data":"b73d6c92fdbaa26bf40a058f166397dbf2cdf6371270317dbdcc0af7775b4813"} Dec 01 08:56:44 crc kubenswrapper[4813]: I1201 08:56:44.181279 4813 generic.go:334] "Generic (PLEG): container finished" podID="323401ed-8df8-4341-8d07-ab27d5162fbd" containerID="0b87db48124c7f7feff803119917724c19e455832f19c3f2431969766e9d77a7" exitCode=0 Dec 01 08:56:44 crc kubenswrapper[4813]: I1201 08:56:44.181379 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240btqv7m" event={"ID":"323401ed-8df8-4341-8d07-ab27d5162fbd","Type":"ContainerDied","Data":"0b87db48124c7f7feff803119917724c19e455832f19c3f2431969766e9d77a7"} Dec 01 08:56:44 crc kubenswrapper[4813]: I1201 08:56:44.185586 4813 generic.go:334] "Generic (PLEG): container finished" podID="6edb6678-d60d-4e1e-92aa-5018a9c1c58d" containerID="45a618c9430ab1f255c16e35c6c18bd5d07f8286d9619ee72c3673628cc81da7" exitCode=0 Dec 01 08:56:44 crc kubenswrapper[4813]: I1201 08:56:44.185644 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-db-sync-gjdp8" event={"ID":"6edb6678-d60d-4e1e-92aa-5018a9c1c58d","Type":"ContainerDied","Data":"45a618c9430ab1f255c16e35c6c18bd5d07f8286d9619ee72c3673628cc81da7"} Dec 01 08:56:45 crc kubenswrapper[4813]: I1201 08:56:45.193880 4813 generic.go:334] "Generic (PLEG): container finished" podID="071eb401-2f12-41e4-b52e-312ca4152da5" containerID="14fe56de24e1d33dbc064fc7bc590e32e2f87c3e7fa4cbda4c3bc26012b961f2" exitCode=0 Dec 01 08:56:45 crc kubenswrapper[4813]: I1201 08:56:45.193952 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368cx94s5" event={"ID":"071eb401-2f12-41e4-b52e-312ca4152da5","Type":"ContainerDied","Data":"14fe56de24e1d33dbc064fc7bc590e32e2f87c3e7fa4cbda4c3bc26012b961f2"} Dec 01 08:56:45 crc kubenswrapper[4813]: I1201 08:56:45.195922 4813 generic.go:334] "Generic (PLEG): container finished" podID="323401ed-8df8-4341-8d07-ab27d5162fbd" containerID="b3764de15eb2105efb977ba179c0526333622e4ab41c6c22974988ef16451ed7" exitCode=0 Dec 01 08:56:45 crc kubenswrapper[4813]: I1201 08:56:45.195945 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240btqv7m" event={"ID":"323401ed-8df8-4341-8d07-ab27d5162fbd","Type":"ContainerDied","Data":"b3764de15eb2105efb977ba179c0526333622e4ab41c6c22974988ef16451ed7"} Dec 01 08:56:45 crc kubenswrapper[4813]: I1201 08:56:45.498310 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-db-sync-gjdp8" Dec 01 08:56:45 crc kubenswrapper[4813]: I1201 08:56:45.502770 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6edb6678-d60d-4e1e-92aa-5018a9c1c58d-config-data\") pod \"6edb6678-d60d-4e1e-92aa-5018a9c1c58d\" (UID: \"6edb6678-d60d-4e1e-92aa-5018a9c1c58d\") " Dec 01 08:56:45 crc kubenswrapper[4813]: I1201 08:56:45.561685 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6edb6678-d60d-4e1e-92aa-5018a9c1c58d-config-data" (OuterVolumeSpecName: "config-data") pod "6edb6678-d60d-4e1e-92aa-5018a9c1c58d" (UID: "6edb6678-d60d-4e1e-92aa-5018a9c1c58d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:56:45 crc kubenswrapper[4813]: I1201 08:56:45.603499 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dqdhj\" (UniqueName: \"kubernetes.io/projected/6edb6678-d60d-4e1e-92aa-5018a9c1c58d-kube-api-access-dqdhj\") pod \"6edb6678-d60d-4e1e-92aa-5018a9c1c58d\" (UID: \"6edb6678-d60d-4e1e-92aa-5018a9c1c58d\") " Dec 01 08:56:45 crc kubenswrapper[4813]: I1201 08:56:45.604539 4813 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6edb6678-d60d-4e1e-92aa-5018a9c1c58d-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 08:56:45 crc kubenswrapper[4813]: I1201 08:56:45.607131 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6edb6678-d60d-4e1e-92aa-5018a9c1c58d-kube-api-access-dqdhj" (OuterVolumeSpecName: "kube-api-access-dqdhj") pod "6edb6678-d60d-4e1e-92aa-5018a9c1c58d" (UID: "6edb6678-d60d-4e1e-92aa-5018a9c1c58d"). InnerVolumeSpecName "kube-api-access-dqdhj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:56:45 crc kubenswrapper[4813]: I1201 08:56:45.706513 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dqdhj\" (UniqueName: \"kubernetes.io/projected/6edb6678-d60d-4e1e-92aa-5018a9c1c58d-kube-api-access-dqdhj\") on node \"crc\" DevicePath \"\"" Dec 01 08:56:46 crc kubenswrapper[4813]: I1201 08:56:46.204225 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-db-sync-gjdp8" event={"ID":"6edb6678-d60d-4e1e-92aa-5018a9c1c58d","Type":"ContainerDied","Data":"a4e36f66500f0c63138cb380060fdd1d5420c363dd05eb2a625d84d943d8b1cd"} Dec 01 08:56:46 crc kubenswrapper[4813]: I1201 08:56:46.204648 4813 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a4e36f66500f0c63138cb380060fdd1d5420c363dd05eb2a625d84d943d8b1cd" Dec 01 08:56:46 crc kubenswrapper[4813]: I1201 08:56:46.204311 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-db-sync-gjdp8" Dec 01 08:56:46 crc kubenswrapper[4813]: I1201 08:56:46.473624 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/keystone-bootstrap-sv5w6"] Dec 01 08:56:46 crc kubenswrapper[4813]: E1201 08:56:46.474100 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6edb6678-d60d-4e1e-92aa-5018a9c1c58d" containerName="keystone-db-sync" Dec 01 08:56:46 crc kubenswrapper[4813]: I1201 08:56:46.474131 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="6edb6678-d60d-4e1e-92aa-5018a9c1c58d" containerName="keystone-db-sync" Dec 01 08:56:46 crc kubenswrapper[4813]: I1201 08:56:46.474370 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="6edb6678-d60d-4e1e-92aa-5018a9c1c58d" containerName="keystone-db-sync" Dec 01 08:56:46 crc kubenswrapper[4813]: I1201 08:56:46.475058 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-bootstrap-sv5w6" Dec 01 08:56:46 crc kubenswrapper[4813]: I1201 08:56:46.477541 4813 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"osp-secret" Dec 01 08:56:46 crc kubenswrapper[4813]: I1201 08:56:46.477768 4813 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"keystone-config-data" Dec 01 08:56:46 crc kubenswrapper[4813]: I1201 08:56:46.480702 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/keystone-bootstrap-sv5w6"] Dec 01 08:56:46 crc kubenswrapper[4813]: I1201 08:56:46.481587 4813 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"keystone" Dec 01 08:56:46 crc kubenswrapper[4813]: I1201 08:56:46.481843 4813 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"keystone-keystone-dockercfg-cwrjn" Dec 01 08:56:46 crc kubenswrapper[4813]: I1201 08:56:46.482814 4813 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"keystone-scripts" Dec 01 08:56:46 crc kubenswrapper[4813]: I1201 08:56:46.535405 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/56b5c9f0-c484-4d62-88d9-67c039dd7bca-scripts\") pod \"keystone-bootstrap-sv5w6\" (UID: \"56b5c9f0-c484-4d62-88d9-67c039dd7bca\") " pod="glance-kuttl-tests/keystone-bootstrap-sv5w6" Dec 01 08:56:46 crc kubenswrapper[4813]: I1201 08:56:46.535481 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/56b5c9f0-c484-4d62-88d9-67c039dd7bca-fernet-keys\") pod \"keystone-bootstrap-sv5w6\" (UID: \"56b5c9f0-c484-4d62-88d9-67c039dd7bca\") " pod="glance-kuttl-tests/keystone-bootstrap-sv5w6" Dec 01 08:56:46 crc kubenswrapper[4813]: I1201 08:56:46.535561 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rlkg2\" (UniqueName: \"kubernetes.io/projected/56b5c9f0-c484-4d62-88d9-67c039dd7bca-kube-api-access-rlkg2\") pod \"keystone-bootstrap-sv5w6\" (UID: \"56b5c9f0-c484-4d62-88d9-67c039dd7bca\") " pod="glance-kuttl-tests/keystone-bootstrap-sv5w6" Dec 01 08:56:46 crc kubenswrapper[4813]: I1201 08:56:46.535640 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/56b5c9f0-c484-4d62-88d9-67c039dd7bca-config-data\") pod \"keystone-bootstrap-sv5w6\" (UID: \"56b5c9f0-c484-4d62-88d9-67c039dd7bca\") " pod="glance-kuttl-tests/keystone-bootstrap-sv5w6" Dec 01 08:56:46 crc kubenswrapper[4813]: I1201 08:56:46.541882 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/56b5c9f0-c484-4d62-88d9-67c039dd7bca-credential-keys\") pod \"keystone-bootstrap-sv5w6\" (UID: \"56b5c9f0-c484-4d62-88d9-67c039dd7bca\") " pod="glance-kuttl-tests/keystone-bootstrap-sv5w6" Dec 01 08:56:46 crc kubenswrapper[4813]: I1201 08:56:46.628994 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240btqv7m" Dec 01 08:56:46 crc kubenswrapper[4813]: I1201 08:56:46.631745 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368cx94s5" Dec 01 08:56:46 crc kubenswrapper[4813]: I1201 08:56:46.642759 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xdkhm\" (UniqueName: \"kubernetes.io/projected/323401ed-8df8-4341-8d07-ab27d5162fbd-kube-api-access-xdkhm\") pod \"323401ed-8df8-4341-8d07-ab27d5162fbd\" (UID: \"323401ed-8df8-4341-8d07-ab27d5162fbd\") " Dec 01 08:56:46 crc kubenswrapper[4813]: I1201 08:56:46.645889 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xn5qf\" (UniqueName: \"kubernetes.io/projected/071eb401-2f12-41e4-b52e-312ca4152da5-kube-api-access-xn5qf\") pod \"071eb401-2f12-41e4-b52e-312ca4152da5\" (UID: \"071eb401-2f12-41e4-b52e-312ca4152da5\") " Dec 01 08:56:46 crc kubenswrapper[4813]: I1201 08:56:46.646079 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/323401ed-8df8-4341-8d07-ab27d5162fbd-bundle\") pod \"323401ed-8df8-4341-8d07-ab27d5162fbd\" (UID: \"323401ed-8df8-4341-8d07-ab27d5162fbd\") " Dec 01 08:56:46 crc kubenswrapper[4813]: I1201 08:56:46.646209 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/323401ed-8df8-4341-8d07-ab27d5162fbd-util\") pod \"323401ed-8df8-4341-8d07-ab27d5162fbd\" (UID: \"323401ed-8df8-4341-8d07-ab27d5162fbd\") " Dec 01 08:56:46 crc kubenswrapper[4813]: I1201 08:56:46.646546 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/071eb401-2f12-41e4-b52e-312ca4152da5-bundle\") pod \"071eb401-2f12-41e4-b52e-312ca4152da5\" (UID: \"071eb401-2f12-41e4-b52e-312ca4152da5\") " Dec 01 08:56:46 crc kubenswrapper[4813]: I1201 08:56:46.646711 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/071eb401-2f12-41e4-b52e-312ca4152da5-util\") pod \"071eb401-2f12-41e4-b52e-312ca4152da5\" (UID: \"071eb401-2f12-41e4-b52e-312ca4152da5\") " Dec 01 08:56:46 crc kubenswrapper[4813]: I1201 08:56:46.646950 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/56b5c9f0-c484-4d62-88d9-67c039dd7bca-config-data\") pod \"keystone-bootstrap-sv5w6\" (UID: \"56b5c9f0-c484-4d62-88d9-67c039dd7bca\") " pod="glance-kuttl-tests/keystone-bootstrap-sv5w6" Dec 01 08:56:46 crc kubenswrapper[4813]: I1201 08:56:46.647139 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/56b5c9f0-c484-4d62-88d9-67c039dd7bca-credential-keys\") pod \"keystone-bootstrap-sv5w6\" (UID: \"56b5c9f0-c484-4d62-88d9-67c039dd7bca\") " pod="glance-kuttl-tests/keystone-bootstrap-sv5w6" Dec 01 08:56:46 crc kubenswrapper[4813]: I1201 08:56:46.648426 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/56b5c9f0-c484-4d62-88d9-67c039dd7bca-scripts\") pod \"keystone-bootstrap-sv5w6\" (UID: \"56b5c9f0-c484-4d62-88d9-67c039dd7bca\") " pod="glance-kuttl-tests/keystone-bootstrap-sv5w6" Dec 01 08:56:46 crc kubenswrapper[4813]: I1201 08:56:46.648538 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/56b5c9f0-c484-4d62-88d9-67c039dd7bca-fernet-keys\") pod \"keystone-bootstrap-sv5w6\" (UID: \"56b5c9f0-c484-4d62-88d9-67c039dd7bca\") " pod="glance-kuttl-tests/keystone-bootstrap-sv5w6" Dec 01 08:56:46 crc kubenswrapper[4813]: I1201 08:56:46.648684 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rlkg2\" (UniqueName: \"kubernetes.io/projected/56b5c9f0-c484-4d62-88d9-67c039dd7bca-kube-api-access-rlkg2\") pod \"keystone-bootstrap-sv5w6\" (UID: \"56b5c9f0-c484-4d62-88d9-67c039dd7bca\") " pod="glance-kuttl-tests/keystone-bootstrap-sv5w6" Dec 01 08:56:46 crc kubenswrapper[4813]: I1201 08:56:46.649166 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/071eb401-2f12-41e4-b52e-312ca4152da5-bundle" (OuterVolumeSpecName: "bundle") pod "071eb401-2f12-41e4-b52e-312ca4152da5" (UID: "071eb401-2f12-41e4-b52e-312ca4152da5"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 08:56:46 crc kubenswrapper[4813]: I1201 08:56:46.650323 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/323401ed-8df8-4341-8d07-ab27d5162fbd-bundle" (OuterVolumeSpecName: "bundle") pod "323401ed-8df8-4341-8d07-ab27d5162fbd" (UID: "323401ed-8df8-4341-8d07-ab27d5162fbd"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 08:56:46 crc kubenswrapper[4813]: I1201 08:56:46.650498 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/323401ed-8df8-4341-8d07-ab27d5162fbd-kube-api-access-xdkhm" (OuterVolumeSpecName: "kube-api-access-xdkhm") pod "323401ed-8df8-4341-8d07-ab27d5162fbd" (UID: "323401ed-8df8-4341-8d07-ab27d5162fbd"). InnerVolumeSpecName "kube-api-access-xdkhm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:56:46 crc kubenswrapper[4813]: I1201 08:56:46.654277 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/56b5c9f0-c484-4d62-88d9-67c039dd7bca-fernet-keys\") pod \"keystone-bootstrap-sv5w6\" (UID: \"56b5c9f0-c484-4d62-88d9-67c039dd7bca\") " pod="glance-kuttl-tests/keystone-bootstrap-sv5w6" Dec 01 08:56:46 crc kubenswrapper[4813]: I1201 08:56:46.663243 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/323401ed-8df8-4341-8d07-ab27d5162fbd-util" (OuterVolumeSpecName: "util") pod "323401ed-8df8-4341-8d07-ab27d5162fbd" (UID: "323401ed-8df8-4341-8d07-ab27d5162fbd"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 08:56:46 crc kubenswrapper[4813]: I1201 08:56:46.666700 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/56b5c9f0-c484-4d62-88d9-67c039dd7bca-credential-keys\") pod \"keystone-bootstrap-sv5w6\" (UID: \"56b5c9f0-c484-4d62-88d9-67c039dd7bca\") " pod="glance-kuttl-tests/keystone-bootstrap-sv5w6" Dec 01 08:56:46 crc kubenswrapper[4813]: I1201 08:56:46.667014 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/56b5c9f0-c484-4d62-88d9-67c039dd7bca-scripts\") pod \"keystone-bootstrap-sv5w6\" (UID: \"56b5c9f0-c484-4d62-88d9-67c039dd7bca\") " pod="glance-kuttl-tests/keystone-bootstrap-sv5w6" Dec 01 08:56:46 crc kubenswrapper[4813]: I1201 08:56:46.667617 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/071eb401-2f12-41e4-b52e-312ca4152da5-util" (OuterVolumeSpecName: "util") pod "071eb401-2f12-41e4-b52e-312ca4152da5" (UID: "071eb401-2f12-41e4-b52e-312ca4152da5"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 08:56:46 crc kubenswrapper[4813]: I1201 08:56:46.669233 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/56b5c9f0-c484-4d62-88d9-67c039dd7bca-config-data\") pod \"keystone-bootstrap-sv5w6\" (UID: \"56b5c9f0-c484-4d62-88d9-67c039dd7bca\") " pod="glance-kuttl-tests/keystone-bootstrap-sv5w6" Dec 01 08:56:46 crc kubenswrapper[4813]: I1201 08:56:46.673670 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/071eb401-2f12-41e4-b52e-312ca4152da5-kube-api-access-xn5qf" (OuterVolumeSpecName: "kube-api-access-xn5qf") pod "071eb401-2f12-41e4-b52e-312ca4152da5" (UID: "071eb401-2f12-41e4-b52e-312ca4152da5"). InnerVolumeSpecName "kube-api-access-xn5qf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:56:46 crc kubenswrapper[4813]: I1201 08:56:46.677432 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rlkg2\" (UniqueName: \"kubernetes.io/projected/56b5c9f0-c484-4d62-88d9-67c039dd7bca-kube-api-access-rlkg2\") pod \"keystone-bootstrap-sv5w6\" (UID: \"56b5c9f0-c484-4d62-88d9-67c039dd7bca\") " pod="glance-kuttl-tests/keystone-bootstrap-sv5w6" Dec 01 08:56:46 crc kubenswrapper[4813]: I1201 08:56:46.749744 4813 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/323401ed-8df8-4341-8d07-ab27d5162fbd-util\") on node \"crc\" DevicePath \"\"" Dec 01 08:56:46 crc kubenswrapper[4813]: I1201 08:56:46.749783 4813 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/071eb401-2f12-41e4-b52e-312ca4152da5-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 08:56:46 crc kubenswrapper[4813]: I1201 08:56:46.749792 4813 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/071eb401-2f12-41e4-b52e-312ca4152da5-util\") on node \"crc\" DevicePath \"\"" Dec 01 08:56:46 crc kubenswrapper[4813]: I1201 08:56:46.749801 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xdkhm\" (UniqueName: \"kubernetes.io/projected/323401ed-8df8-4341-8d07-ab27d5162fbd-kube-api-access-xdkhm\") on node \"crc\" DevicePath \"\"" Dec 01 08:56:46 crc kubenswrapper[4813]: I1201 08:56:46.749814 4813 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/323401ed-8df8-4341-8d07-ab27d5162fbd-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 08:56:46 crc kubenswrapper[4813]: I1201 08:56:46.749822 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xn5qf\" (UniqueName: \"kubernetes.io/projected/071eb401-2f12-41e4-b52e-312ca4152da5-kube-api-access-xn5qf\") on node \"crc\" DevicePath \"\"" Dec 01 08:56:46 crc kubenswrapper[4813]: I1201 08:56:46.803791 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-bootstrap-sv5w6" Dec 01 08:56:47 crc kubenswrapper[4813]: I1201 08:56:47.216370 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368cx94s5" event={"ID":"071eb401-2f12-41e4-b52e-312ca4152da5","Type":"ContainerDied","Data":"ec1f5ef5c5f0ac817fb3b72de3ef1e1835929b884abb19cc7b49e30f81b1d134"} Dec 01 08:56:47 crc kubenswrapper[4813]: I1201 08:56:47.216687 4813 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ec1f5ef5c5f0ac817fb3b72de3ef1e1835929b884abb19cc7b49e30f81b1d134" Dec 01 08:56:47 crc kubenswrapper[4813]: I1201 08:56:47.216555 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368cx94s5" Dec 01 08:56:47 crc kubenswrapper[4813]: I1201 08:56:47.223286 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240btqv7m" event={"ID":"323401ed-8df8-4341-8d07-ab27d5162fbd","Type":"ContainerDied","Data":"28c89dd0520e03bbcc3600cdf623348d5412ab3265c099940a9337bfe57c85bf"} Dec 01 08:56:47 crc kubenswrapper[4813]: I1201 08:56:47.223339 4813 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="28c89dd0520e03bbcc3600cdf623348d5412ab3265c099940a9337bfe57c85bf" Dec 01 08:56:47 crc kubenswrapper[4813]: I1201 08:56:47.223431 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240btqv7m" Dec 01 08:56:47 crc kubenswrapper[4813]: I1201 08:56:47.262308 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/keystone-bootstrap-sv5w6"] Dec 01 08:56:47 crc kubenswrapper[4813]: W1201 08:56:47.273216 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod56b5c9f0_c484_4d62_88d9_67c039dd7bca.slice/crio-a86270b50cd8b39af48cbf52c48186602601f4ddb560adaefae6b9e9f7d9443b WatchSource:0}: Error finding container a86270b50cd8b39af48cbf52c48186602601f4ddb560adaefae6b9e9f7d9443b: Status 404 returned error can't find the container with id a86270b50cd8b39af48cbf52c48186602601f4ddb560adaefae6b9e9f7d9443b Dec 01 08:56:48 crc kubenswrapper[4813]: I1201 08:56:48.237688 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-bootstrap-sv5w6" event={"ID":"56b5c9f0-c484-4d62-88d9-67c039dd7bca","Type":"ContainerStarted","Data":"4e11dedd29ab57c4ae5dcca4bc4950418b6887646e4a1a2186fa20a3abb3c157"} Dec 01 08:56:48 crc kubenswrapper[4813]: I1201 08:56:48.238107 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-bootstrap-sv5w6" event={"ID":"56b5c9f0-c484-4d62-88d9-67c039dd7bca","Type":"ContainerStarted","Data":"a86270b50cd8b39af48cbf52c48186602601f4ddb560adaefae6b9e9f7d9443b"} Dec 01 08:56:48 crc kubenswrapper[4813]: I1201 08:56:48.267621 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/keystone-bootstrap-sv5w6" podStartSLOduration=2.2675857280000002 podStartE2EDuration="2.267585728s" podCreationTimestamp="2025-12-01 08:56:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:56:48.262796664 +0000 UTC m=+992.885718290" watchObservedRunningTime="2025-12-01 08:56:48.267585728 +0000 UTC m=+992.890507354" Dec 01 08:56:50 crc kubenswrapper[4813]: I1201 08:56:50.262244 4813 generic.go:334] "Generic (PLEG): container finished" podID="56b5c9f0-c484-4d62-88d9-67c039dd7bca" containerID="4e11dedd29ab57c4ae5dcca4bc4950418b6887646e4a1a2186fa20a3abb3c157" exitCode=0 Dec 01 08:56:50 crc kubenswrapper[4813]: I1201 08:56:50.262800 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-bootstrap-sv5w6" event={"ID":"56b5c9f0-c484-4d62-88d9-67c039dd7bca","Type":"ContainerDied","Data":"4e11dedd29ab57c4ae5dcca4bc4950418b6887646e4a1a2186fa20a3abb3c157"} Dec 01 08:56:51 crc kubenswrapper[4813]: I1201 08:56:51.593842 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-bootstrap-sv5w6" Dec 01 08:56:51 crc kubenswrapper[4813]: I1201 08:56:51.779437 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/56b5c9f0-c484-4d62-88d9-67c039dd7bca-scripts\") pod \"56b5c9f0-c484-4d62-88d9-67c039dd7bca\" (UID: \"56b5c9f0-c484-4d62-88d9-67c039dd7bca\") " Dec 01 08:56:51 crc kubenswrapper[4813]: I1201 08:56:51.779515 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/56b5c9f0-c484-4d62-88d9-67c039dd7bca-config-data\") pod \"56b5c9f0-c484-4d62-88d9-67c039dd7bca\" (UID: \"56b5c9f0-c484-4d62-88d9-67c039dd7bca\") " Dec 01 08:56:51 crc kubenswrapper[4813]: I1201 08:56:51.779548 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/56b5c9f0-c484-4d62-88d9-67c039dd7bca-credential-keys\") pod \"56b5c9f0-c484-4d62-88d9-67c039dd7bca\" (UID: \"56b5c9f0-c484-4d62-88d9-67c039dd7bca\") " Dec 01 08:56:51 crc kubenswrapper[4813]: I1201 08:56:51.779621 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rlkg2\" (UniqueName: \"kubernetes.io/projected/56b5c9f0-c484-4d62-88d9-67c039dd7bca-kube-api-access-rlkg2\") pod \"56b5c9f0-c484-4d62-88d9-67c039dd7bca\" (UID: \"56b5c9f0-c484-4d62-88d9-67c039dd7bca\") " Dec 01 08:56:51 crc kubenswrapper[4813]: I1201 08:56:51.779758 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/56b5c9f0-c484-4d62-88d9-67c039dd7bca-fernet-keys\") pod \"56b5c9f0-c484-4d62-88d9-67c039dd7bca\" (UID: \"56b5c9f0-c484-4d62-88d9-67c039dd7bca\") " Dec 01 08:56:51 crc kubenswrapper[4813]: I1201 08:56:51.784482 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/56b5c9f0-c484-4d62-88d9-67c039dd7bca-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "56b5c9f0-c484-4d62-88d9-67c039dd7bca" (UID: "56b5c9f0-c484-4d62-88d9-67c039dd7bca"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:56:51 crc kubenswrapper[4813]: I1201 08:56:51.784740 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/56b5c9f0-c484-4d62-88d9-67c039dd7bca-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "56b5c9f0-c484-4d62-88d9-67c039dd7bca" (UID: "56b5c9f0-c484-4d62-88d9-67c039dd7bca"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:56:51 crc kubenswrapper[4813]: I1201 08:56:51.784745 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/56b5c9f0-c484-4d62-88d9-67c039dd7bca-kube-api-access-rlkg2" (OuterVolumeSpecName: "kube-api-access-rlkg2") pod "56b5c9f0-c484-4d62-88d9-67c039dd7bca" (UID: "56b5c9f0-c484-4d62-88d9-67c039dd7bca"). InnerVolumeSpecName "kube-api-access-rlkg2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:56:51 crc kubenswrapper[4813]: I1201 08:56:51.794256 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/56b5c9f0-c484-4d62-88d9-67c039dd7bca-scripts" (OuterVolumeSpecName: "scripts") pod "56b5c9f0-c484-4d62-88d9-67c039dd7bca" (UID: "56b5c9f0-c484-4d62-88d9-67c039dd7bca"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:56:51 crc kubenswrapper[4813]: I1201 08:56:51.801924 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/56b5c9f0-c484-4d62-88d9-67c039dd7bca-config-data" (OuterVolumeSpecName: "config-data") pod "56b5c9f0-c484-4d62-88d9-67c039dd7bca" (UID: "56b5c9f0-c484-4d62-88d9-67c039dd7bca"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:56:51 crc kubenswrapper[4813]: I1201 08:56:51.881018 4813 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/56b5c9f0-c484-4d62-88d9-67c039dd7bca-fernet-keys\") on node \"crc\" DevicePath \"\"" Dec 01 08:56:51 crc kubenswrapper[4813]: I1201 08:56:51.881072 4813 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/56b5c9f0-c484-4d62-88d9-67c039dd7bca-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 08:56:51 crc kubenswrapper[4813]: I1201 08:56:51.881082 4813 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/56b5c9f0-c484-4d62-88d9-67c039dd7bca-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 08:56:51 crc kubenswrapper[4813]: I1201 08:56:51.881090 4813 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/56b5c9f0-c484-4d62-88d9-67c039dd7bca-credential-keys\") on node \"crc\" DevicePath \"\"" Dec 01 08:56:51 crc kubenswrapper[4813]: I1201 08:56:51.881107 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rlkg2\" (UniqueName: \"kubernetes.io/projected/56b5c9f0-c484-4d62-88d9-67c039dd7bca-kube-api-access-rlkg2\") on node \"crc\" DevicePath \"\"" Dec 01 08:56:52 crc kubenswrapper[4813]: I1201 08:56:52.283530 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-bootstrap-sv5w6" event={"ID":"56b5c9f0-c484-4d62-88d9-67c039dd7bca","Type":"ContainerDied","Data":"a86270b50cd8b39af48cbf52c48186602601f4ddb560adaefae6b9e9f7d9443b"} Dec 01 08:56:52 crc kubenswrapper[4813]: I1201 08:56:52.283603 4813 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a86270b50cd8b39af48cbf52c48186602601f4ddb560adaefae6b9e9f7d9443b" Dec 01 08:56:52 crc kubenswrapper[4813]: I1201 08:56:52.283654 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-bootstrap-sv5w6" Dec 01 08:56:52 crc kubenswrapper[4813]: I1201 08:56:52.475060 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/keystone-86b5f6997f-8d5h9"] Dec 01 08:56:52 crc kubenswrapper[4813]: E1201 08:56:52.475567 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="323401ed-8df8-4341-8d07-ab27d5162fbd" containerName="extract" Dec 01 08:56:52 crc kubenswrapper[4813]: I1201 08:56:52.475601 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="323401ed-8df8-4341-8d07-ab27d5162fbd" containerName="extract" Dec 01 08:56:52 crc kubenswrapper[4813]: E1201 08:56:52.475625 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="071eb401-2f12-41e4-b52e-312ca4152da5" containerName="pull" Dec 01 08:56:52 crc kubenswrapper[4813]: I1201 08:56:52.475634 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="071eb401-2f12-41e4-b52e-312ca4152da5" containerName="pull" Dec 01 08:56:52 crc kubenswrapper[4813]: E1201 08:56:52.475652 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="071eb401-2f12-41e4-b52e-312ca4152da5" containerName="extract" Dec 01 08:56:52 crc kubenswrapper[4813]: I1201 08:56:52.475661 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="071eb401-2f12-41e4-b52e-312ca4152da5" containerName="extract" Dec 01 08:56:52 crc kubenswrapper[4813]: E1201 08:56:52.475675 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="323401ed-8df8-4341-8d07-ab27d5162fbd" containerName="pull" Dec 01 08:56:52 crc kubenswrapper[4813]: I1201 08:56:52.475684 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="323401ed-8df8-4341-8d07-ab27d5162fbd" containerName="pull" Dec 01 08:56:52 crc kubenswrapper[4813]: E1201 08:56:52.475701 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="071eb401-2f12-41e4-b52e-312ca4152da5" containerName="util" Dec 01 08:56:52 crc kubenswrapper[4813]: I1201 08:56:52.475709 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="071eb401-2f12-41e4-b52e-312ca4152da5" containerName="util" Dec 01 08:56:52 crc kubenswrapper[4813]: E1201 08:56:52.475725 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="323401ed-8df8-4341-8d07-ab27d5162fbd" containerName="util" Dec 01 08:56:52 crc kubenswrapper[4813]: I1201 08:56:52.475733 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="323401ed-8df8-4341-8d07-ab27d5162fbd" containerName="util" Dec 01 08:56:52 crc kubenswrapper[4813]: E1201 08:56:52.475747 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="56b5c9f0-c484-4d62-88d9-67c039dd7bca" containerName="keystone-bootstrap" Dec 01 08:56:52 crc kubenswrapper[4813]: I1201 08:56:52.475756 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="56b5c9f0-c484-4d62-88d9-67c039dd7bca" containerName="keystone-bootstrap" Dec 01 08:56:52 crc kubenswrapper[4813]: I1201 08:56:52.475932 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="56b5c9f0-c484-4d62-88d9-67c039dd7bca" containerName="keystone-bootstrap" Dec 01 08:56:52 crc kubenswrapper[4813]: I1201 08:56:52.475961 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="323401ed-8df8-4341-8d07-ab27d5162fbd" containerName="extract" Dec 01 08:56:52 crc kubenswrapper[4813]: I1201 08:56:52.476001 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="071eb401-2f12-41e4-b52e-312ca4152da5" containerName="extract" Dec 01 08:56:52 crc kubenswrapper[4813]: I1201 08:56:52.476753 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-86b5f6997f-8d5h9" Dec 01 08:56:52 crc kubenswrapper[4813]: I1201 08:56:52.479717 4813 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"keystone-scripts" Dec 01 08:56:52 crc kubenswrapper[4813]: I1201 08:56:52.479717 4813 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"keystone" Dec 01 08:56:52 crc kubenswrapper[4813]: I1201 08:56:52.479837 4813 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"keystone-config-data" Dec 01 08:56:52 crc kubenswrapper[4813]: I1201 08:56:52.479853 4813 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"keystone-keystone-dockercfg-cwrjn" Dec 01 08:56:52 crc kubenswrapper[4813]: I1201 08:56:52.484288 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/keystone-86b5f6997f-8d5h9"] Dec 01 08:56:52 crc kubenswrapper[4813]: I1201 08:56:52.489043 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a80abda8-962d-4b0a-bc3a-2e6d693d0393-scripts\") pod \"keystone-86b5f6997f-8d5h9\" (UID: \"a80abda8-962d-4b0a-bc3a-2e6d693d0393\") " pod="glance-kuttl-tests/keystone-86b5f6997f-8d5h9" Dec 01 08:56:52 crc kubenswrapper[4813]: I1201 08:56:52.489303 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/a80abda8-962d-4b0a-bc3a-2e6d693d0393-fernet-keys\") pod \"keystone-86b5f6997f-8d5h9\" (UID: \"a80abda8-962d-4b0a-bc3a-2e6d693d0393\") " pod="glance-kuttl-tests/keystone-86b5f6997f-8d5h9" Dec 01 08:56:52 crc kubenswrapper[4813]: I1201 08:56:52.489449 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fjj2b\" (UniqueName: \"kubernetes.io/projected/a80abda8-962d-4b0a-bc3a-2e6d693d0393-kube-api-access-fjj2b\") pod \"keystone-86b5f6997f-8d5h9\" (UID: \"a80abda8-962d-4b0a-bc3a-2e6d693d0393\") " pod="glance-kuttl-tests/keystone-86b5f6997f-8d5h9" Dec 01 08:56:52 crc kubenswrapper[4813]: I1201 08:56:52.489637 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/a80abda8-962d-4b0a-bc3a-2e6d693d0393-credential-keys\") pod \"keystone-86b5f6997f-8d5h9\" (UID: \"a80abda8-962d-4b0a-bc3a-2e6d693d0393\") " pod="glance-kuttl-tests/keystone-86b5f6997f-8d5h9" Dec 01 08:56:52 crc kubenswrapper[4813]: I1201 08:56:52.489736 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a80abda8-962d-4b0a-bc3a-2e6d693d0393-config-data\") pod \"keystone-86b5f6997f-8d5h9\" (UID: \"a80abda8-962d-4b0a-bc3a-2e6d693d0393\") " pod="glance-kuttl-tests/keystone-86b5f6997f-8d5h9" Dec 01 08:56:52 crc kubenswrapper[4813]: I1201 08:56:52.591603 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/a80abda8-962d-4b0a-bc3a-2e6d693d0393-credential-keys\") pod \"keystone-86b5f6997f-8d5h9\" (UID: \"a80abda8-962d-4b0a-bc3a-2e6d693d0393\") " pod="glance-kuttl-tests/keystone-86b5f6997f-8d5h9" Dec 01 08:56:52 crc kubenswrapper[4813]: I1201 08:56:52.591691 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a80abda8-962d-4b0a-bc3a-2e6d693d0393-config-data\") pod \"keystone-86b5f6997f-8d5h9\" (UID: \"a80abda8-962d-4b0a-bc3a-2e6d693d0393\") " pod="glance-kuttl-tests/keystone-86b5f6997f-8d5h9" Dec 01 08:56:52 crc kubenswrapper[4813]: I1201 08:56:52.591743 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a80abda8-962d-4b0a-bc3a-2e6d693d0393-scripts\") pod \"keystone-86b5f6997f-8d5h9\" (UID: \"a80abda8-962d-4b0a-bc3a-2e6d693d0393\") " pod="glance-kuttl-tests/keystone-86b5f6997f-8d5h9" Dec 01 08:56:52 crc kubenswrapper[4813]: I1201 08:56:52.591827 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/a80abda8-962d-4b0a-bc3a-2e6d693d0393-fernet-keys\") pod \"keystone-86b5f6997f-8d5h9\" (UID: \"a80abda8-962d-4b0a-bc3a-2e6d693d0393\") " pod="glance-kuttl-tests/keystone-86b5f6997f-8d5h9" Dec 01 08:56:52 crc kubenswrapper[4813]: I1201 08:56:52.591871 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fjj2b\" (UniqueName: \"kubernetes.io/projected/a80abda8-962d-4b0a-bc3a-2e6d693d0393-kube-api-access-fjj2b\") pod \"keystone-86b5f6997f-8d5h9\" (UID: \"a80abda8-962d-4b0a-bc3a-2e6d693d0393\") " pod="glance-kuttl-tests/keystone-86b5f6997f-8d5h9" Dec 01 08:56:52 crc kubenswrapper[4813]: I1201 08:56:52.598064 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/a80abda8-962d-4b0a-bc3a-2e6d693d0393-credential-keys\") pod \"keystone-86b5f6997f-8d5h9\" (UID: \"a80abda8-962d-4b0a-bc3a-2e6d693d0393\") " pod="glance-kuttl-tests/keystone-86b5f6997f-8d5h9" Dec 01 08:56:52 crc kubenswrapper[4813]: I1201 08:56:52.598224 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a80abda8-962d-4b0a-bc3a-2e6d693d0393-scripts\") pod \"keystone-86b5f6997f-8d5h9\" (UID: \"a80abda8-962d-4b0a-bc3a-2e6d693d0393\") " pod="glance-kuttl-tests/keystone-86b5f6997f-8d5h9" Dec 01 08:56:52 crc kubenswrapper[4813]: I1201 08:56:52.598847 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a80abda8-962d-4b0a-bc3a-2e6d693d0393-config-data\") pod \"keystone-86b5f6997f-8d5h9\" (UID: \"a80abda8-962d-4b0a-bc3a-2e6d693d0393\") " pod="glance-kuttl-tests/keystone-86b5f6997f-8d5h9" Dec 01 08:56:52 crc kubenswrapper[4813]: I1201 08:56:52.599481 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/a80abda8-962d-4b0a-bc3a-2e6d693d0393-fernet-keys\") pod \"keystone-86b5f6997f-8d5h9\" (UID: \"a80abda8-962d-4b0a-bc3a-2e6d693d0393\") " pod="glance-kuttl-tests/keystone-86b5f6997f-8d5h9" Dec 01 08:56:52 crc kubenswrapper[4813]: I1201 08:56:52.611420 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fjj2b\" (UniqueName: \"kubernetes.io/projected/a80abda8-962d-4b0a-bc3a-2e6d693d0393-kube-api-access-fjj2b\") pod \"keystone-86b5f6997f-8d5h9\" (UID: \"a80abda8-962d-4b0a-bc3a-2e6d693d0393\") " pod="glance-kuttl-tests/keystone-86b5f6997f-8d5h9" Dec 01 08:56:52 crc kubenswrapper[4813]: I1201 08:56:52.796044 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-86b5f6997f-8d5h9" Dec 01 08:56:53 crc kubenswrapper[4813]: I1201 08:56:53.336286 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/keystone-86b5f6997f-8d5h9"] Dec 01 08:56:54 crc kubenswrapper[4813]: I1201 08:56:54.310838 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-86b5f6997f-8d5h9" event={"ID":"a80abda8-962d-4b0a-bc3a-2e6d693d0393","Type":"ContainerStarted","Data":"54095cf0e86899dcdcc5b94f8681013eb42cc52529b80cec4da277170498a551"} Dec 01 08:56:54 crc kubenswrapper[4813]: I1201 08:56:54.311590 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-86b5f6997f-8d5h9" event={"ID":"a80abda8-962d-4b0a-bc3a-2e6d693d0393","Type":"ContainerStarted","Data":"5700b50bae4953f211699772cee4993299ebab0fbba90fc0e8fd84d8f4f66a25"} Dec 01 08:56:54 crc kubenswrapper[4813]: I1201 08:56:54.311686 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/keystone-86b5f6997f-8d5h9" Dec 01 08:56:58 crc kubenswrapper[4813]: I1201 08:56:58.628553 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/keystone-86b5f6997f-8d5h9" podStartSLOduration=6.628506782 podStartE2EDuration="6.628506782s" podCreationTimestamp="2025-12-01 08:56:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:56:54.350835581 +0000 UTC m=+998.973757167" watchObservedRunningTime="2025-12-01 08:56:58.628506782 +0000 UTC m=+1003.251428368" Dec 01 08:56:58 crc kubenswrapper[4813]: I1201 08:56:58.635501 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-controller-manager-5998794db6-9lz55"] Dec 01 08:56:58 crc kubenswrapper[4813]: I1201 08:56:58.637029 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-5998794db6-9lz55" Dec 01 08:56:58 crc kubenswrapper[4813]: I1201 08:56:58.641557 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-service-cert" Dec 01 08:56:58 crc kubenswrapper[4813]: I1201 08:56:58.645554 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-dockercfg-skcw9" Dec 01 08:56:58 crc kubenswrapper[4813]: I1201 08:56:58.647161 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-5998794db6-9lz55"] Dec 01 08:56:58 crc kubenswrapper[4813]: I1201 08:56:58.701047 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fmwgh\" (UniqueName: \"kubernetes.io/projected/d9af9c23-0392-4caf-b10f-afc96b109c49-kube-api-access-fmwgh\") pod \"horizon-operator-controller-manager-5998794db6-9lz55\" (UID: \"d9af9c23-0392-4caf-b10f-afc96b109c49\") " pod="openstack-operators/horizon-operator-controller-manager-5998794db6-9lz55" Dec 01 08:56:58 crc kubenswrapper[4813]: I1201 08:56:58.701233 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/d9af9c23-0392-4caf-b10f-afc96b109c49-webhook-cert\") pod \"horizon-operator-controller-manager-5998794db6-9lz55\" (UID: \"d9af9c23-0392-4caf-b10f-afc96b109c49\") " pod="openstack-operators/horizon-operator-controller-manager-5998794db6-9lz55" Dec 01 08:56:58 crc kubenswrapper[4813]: I1201 08:56:58.701498 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/d9af9c23-0392-4caf-b10f-afc96b109c49-apiservice-cert\") pod \"horizon-operator-controller-manager-5998794db6-9lz55\" (UID: \"d9af9c23-0392-4caf-b10f-afc96b109c49\") " pod="openstack-operators/horizon-operator-controller-manager-5998794db6-9lz55" Dec 01 08:56:58 crc kubenswrapper[4813]: I1201 08:56:58.802632 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/d9af9c23-0392-4caf-b10f-afc96b109c49-apiservice-cert\") pod \"horizon-operator-controller-manager-5998794db6-9lz55\" (UID: \"d9af9c23-0392-4caf-b10f-afc96b109c49\") " pod="openstack-operators/horizon-operator-controller-manager-5998794db6-9lz55" Dec 01 08:56:58 crc kubenswrapper[4813]: I1201 08:56:58.804054 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fmwgh\" (UniqueName: \"kubernetes.io/projected/d9af9c23-0392-4caf-b10f-afc96b109c49-kube-api-access-fmwgh\") pod \"horizon-operator-controller-manager-5998794db6-9lz55\" (UID: \"d9af9c23-0392-4caf-b10f-afc96b109c49\") " pod="openstack-operators/horizon-operator-controller-manager-5998794db6-9lz55" Dec 01 08:56:58 crc kubenswrapper[4813]: I1201 08:56:58.804121 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/d9af9c23-0392-4caf-b10f-afc96b109c49-webhook-cert\") pod \"horizon-operator-controller-manager-5998794db6-9lz55\" (UID: \"d9af9c23-0392-4caf-b10f-afc96b109c49\") " pod="openstack-operators/horizon-operator-controller-manager-5998794db6-9lz55" Dec 01 08:56:58 crc kubenswrapper[4813]: I1201 08:56:58.814789 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/d9af9c23-0392-4caf-b10f-afc96b109c49-apiservice-cert\") pod \"horizon-operator-controller-manager-5998794db6-9lz55\" (UID: \"d9af9c23-0392-4caf-b10f-afc96b109c49\") " pod="openstack-operators/horizon-operator-controller-manager-5998794db6-9lz55" Dec 01 08:56:58 crc kubenswrapper[4813]: I1201 08:56:58.822169 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fmwgh\" (UniqueName: \"kubernetes.io/projected/d9af9c23-0392-4caf-b10f-afc96b109c49-kube-api-access-fmwgh\") pod \"horizon-operator-controller-manager-5998794db6-9lz55\" (UID: \"d9af9c23-0392-4caf-b10f-afc96b109c49\") " pod="openstack-operators/horizon-operator-controller-manager-5998794db6-9lz55" Dec 01 08:56:58 crc kubenswrapper[4813]: I1201 08:56:58.828485 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/d9af9c23-0392-4caf-b10f-afc96b109c49-webhook-cert\") pod \"horizon-operator-controller-manager-5998794db6-9lz55\" (UID: \"d9af9c23-0392-4caf-b10f-afc96b109c49\") " pod="openstack-operators/horizon-operator-controller-manager-5998794db6-9lz55" Dec 01 08:56:58 crc kubenswrapper[4813]: I1201 08:56:58.970595 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-5998794db6-9lz55" Dec 01 08:56:59 crc kubenswrapper[4813]: I1201 08:56:59.291631 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-5998794db6-9lz55"] Dec 01 08:56:59 crc kubenswrapper[4813]: I1201 08:56:59.344487 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-5998794db6-9lz55" event={"ID":"d9af9c23-0392-4caf-b10f-afc96b109c49","Type":"ContainerStarted","Data":"bc0178df7b92ad4bf97f769662915b5cbd97481c819a6c7f9e0ce19ea5c03959"} Dec 01 08:57:02 crc kubenswrapper[4813]: I1201 08:57:02.371430 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-5998794db6-9lz55" event={"ID":"d9af9c23-0392-4caf-b10f-afc96b109c49","Type":"ContainerStarted","Data":"320abd634adbde8c1781f18d5c0e5bab3ed58625a20ff32b9d6c0f2498c80094"} Dec 01 08:57:02 crc kubenswrapper[4813]: I1201 08:57:02.372132 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-5998794db6-9lz55" Dec 01 08:57:02 crc kubenswrapper[4813]: I1201 08:57:02.397879 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-controller-manager-5998794db6-9lz55" podStartSLOduration=1.6796778639999999 podStartE2EDuration="4.397844691s" podCreationTimestamp="2025-12-01 08:56:58 +0000 UTC" firstStartedPulling="2025-12-01 08:56:59.302050416 +0000 UTC m=+1003.924972002" lastFinishedPulling="2025-12-01 08:57:02.020217243 +0000 UTC m=+1006.643138829" observedRunningTime="2025-12-01 08:57:02.395164935 +0000 UTC m=+1007.018086601" watchObservedRunningTime="2025-12-01 08:57:02.397844691 +0000 UTC m=+1007.020766317" Dec 01 08:57:07 crc kubenswrapper[4813]: I1201 08:57:07.998998 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-controller-manager-6b4988f668-x5frt"] Dec 01 08:57:08 crc kubenswrapper[4813]: I1201 08:57:08.001564 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-6b4988f668-x5frt" Dec 01 08:57:08 crc kubenswrapper[4813]: I1201 08:57:08.003692 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-service-cert" Dec 01 08:57:08 crc kubenswrapper[4813]: I1201 08:57:08.004240 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-dockercfg-n4759" Dec 01 08:57:08 crc kubenswrapper[4813]: I1201 08:57:08.017509 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-6b4988f668-x5frt"] Dec 01 08:57:08 crc kubenswrapper[4813]: I1201 08:57:08.203486 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6qdh5\" (UniqueName: \"kubernetes.io/projected/abb50b08-d82d-4e6a-af37-5a99d2624339-kube-api-access-6qdh5\") pod \"swift-operator-controller-manager-6b4988f668-x5frt\" (UID: \"abb50b08-d82d-4e6a-af37-5a99d2624339\") " pod="openstack-operators/swift-operator-controller-manager-6b4988f668-x5frt" Dec 01 08:57:08 crc kubenswrapper[4813]: I1201 08:57:08.203585 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/abb50b08-d82d-4e6a-af37-5a99d2624339-apiservice-cert\") pod \"swift-operator-controller-manager-6b4988f668-x5frt\" (UID: \"abb50b08-d82d-4e6a-af37-5a99d2624339\") " pod="openstack-operators/swift-operator-controller-manager-6b4988f668-x5frt" Dec 01 08:57:08 crc kubenswrapper[4813]: I1201 08:57:08.203730 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/abb50b08-d82d-4e6a-af37-5a99d2624339-webhook-cert\") pod \"swift-operator-controller-manager-6b4988f668-x5frt\" (UID: \"abb50b08-d82d-4e6a-af37-5a99d2624339\") " pod="openstack-operators/swift-operator-controller-manager-6b4988f668-x5frt" Dec 01 08:57:08 crc kubenswrapper[4813]: I1201 08:57:08.304655 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/abb50b08-d82d-4e6a-af37-5a99d2624339-webhook-cert\") pod \"swift-operator-controller-manager-6b4988f668-x5frt\" (UID: \"abb50b08-d82d-4e6a-af37-5a99d2624339\") " pod="openstack-operators/swift-operator-controller-manager-6b4988f668-x5frt" Dec 01 08:57:08 crc kubenswrapper[4813]: I1201 08:57:08.304754 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6qdh5\" (UniqueName: \"kubernetes.io/projected/abb50b08-d82d-4e6a-af37-5a99d2624339-kube-api-access-6qdh5\") pod \"swift-operator-controller-manager-6b4988f668-x5frt\" (UID: \"abb50b08-d82d-4e6a-af37-5a99d2624339\") " pod="openstack-operators/swift-operator-controller-manager-6b4988f668-x5frt" Dec 01 08:57:08 crc kubenswrapper[4813]: I1201 08:57:08.304836 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/abb50b08-d82d-4e6a-af37-5a99d2624339-apiservice-cert\") pod \"swift-operator-controller-manager-6b4988f668-x5frt\" (UID: \"abb50b08-d82d-4e6a-af37-5a99d2624339\") " pod="openstack-operators/swift-operator-controller-manager-6b4988f668-x5frt" Dec 01 08:57:08 crc kubenswrapper[4813]: I1201 08:57:08.311164 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/abb50b08-d82d-4e6a-af37-5a99d2624339-apiservice-cert\") pod \"swift-operator-controller-manager-6b4988f668-x5frt\" (UID: \"abb50b08-d82d-4e6a-af37-5a99d2624339\") " pod="openstack-operators/swift-operator-controller-manager-6b4988f668-x5frt" Dec 01 08:57:08 crc kubenswrapper[4813]: I1201 08:57:08.311348 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/abb50b08-d82d-4e6a-af37-5a99d2624339-webhook-cert\") pod \"swift-operator-controller-manager-6b4988f668-x5frt\" (UID: \"abb50b08-d82d-4e6a-af37-5a99d2624339\") " pod="openstack-operators/swift-operator-controller-manager-6b4988f668-x5frt" Dec 01 08:57:08 crc kubenswrapper[4813]: I1201 08:57:08.319892 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6qdh5\" (UniqueName: \"kubernetes.io/projected/abb50b08-d82d-4e6a-af37-5a99d2624339-kube-api-access-6qdh5\") pod \"swift-operator-controller-manager-6b4988f668-x5frt\" (UID: \"abb50b08-d82d-4e6a-af37-5a99d2624339\") " pod="openstack-operators/swift-operator-controller-manager-6b4988f668-x5frt" Dec 01 08:57:08 crc kubenswrapper[4813]: I1201 08:57:08.617612 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-6b4988f668-x5frt" Dec 01 08:57:08 crc kubenswrapper[4813]: I1201 08:57:08.976683 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-controller-manager-5998794db6-9lz55" Dec 01 08:57:09 crc kubenswrapper[4813]: I1201 08:57:09.222990 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-6b4988f668-x5frt"] Dec 01 08:57:09 crc kubenswrapper[4813]: I1201 08:57:09.426954 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-6b4988f668-x5frt" event={"ID":"abb50b08-d82d-4e6a-af37-5a99d2624339","Type":"ContainerStarted","Data":"6247e4da180a31a9b17d5988bcae4b950154328f189ef91fb77c80b4f1023053"} Dec 01 08:57:13 crc kubenswrapper[4813]: I1201 08:57:13.453530 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-6b4988f668-x5frt" event={"ID":"abb50b08-d82d-4e6a-af37-5a99d2624339","Type":"ContainerStarted","Data":"f4982383ff10a95aef6eb6240a6c60c5ba0aec1585bd468ca90f8641c3bfad63"} Dec 01 08:57:13 crc kubenswrapper[4813]: I1201 08:57:13.455241 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-6b4988f668-x5frt" Dec 01 08:57:13 crc kubenswrapper[4813]: I1201 08:57:13.481834 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-controller-manager-6b4988f668-x5frt" podStartSLOduration=2.905061237 podStartE2EDuration="6.481803037s" podCreationTimestamp="2025-12-01 08:57:07 +0000 UTC" firstStartedPulling="2025-12-01 08:57:09.229094615 +0000 UTC m=+1013.852016201" lastFinishedPulling="2025-12-01 08:57:12.805836415 +0000 UTC m=+1017.428758001" observedRunningTime="2025-12-01 08:57:13.472156157 +0000 UTC m=+1018.095077753" watchObservedRunningTime="2025-12-01 08:57:13.481803037 +0000 UTC m=+1018.104724633" Dec 01 08:57:18 crc kubenswrapper[4813]: I1201 08:57:18.624838 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-controller-manager-6b4988f668-x5frt" Dec 01 08:57:22 crc kubenswrapper[4813]: I1201 08:57:22.250070 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/swift-storage-0"] Dec 01 08:57:22 crc kubenswrapper[4813]: I1201 08:57:22.260513 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/swift-storage-0" Dec 01 08:57:22 crc kubenswrapper[4813]: I1201 08:57:22.264431 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"swift-ring-files" Dec 01 08:57:22 crc kubenswrapper[4813]: I1201 08:57:22.264430 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"swift-storage-config-data" Dec 01 08:57:22 crc kubenswrapper[4813]: I1201 08:57:22.264521 4813 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"swift-conf" Dec 01 08:57:22 crc kubenswrapper[4813]: I1201 08:57:22.264956 4813 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"swift-swift-dockercfg-w5f9s" Dec 01 08:57:22 crc kubenswrapper[4813]: I1201 08:57:22.288668 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/swift-storage-0"] Dec 01 08:57:22 crc kubenswrapper[4813]: I1201 08:57:22.312725 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/995100a7-4c5a-409a-81f2-9dbaf2fa1d47-lock\") pod \"swift-storage-0\" (UID: \"995100a7-4c5a-409a-81f2-9dbaf2fa1d47\") " pod="glance-kuttl-tests/swift-storage-0" Dec 01 08:57:22 crc kubenswrapper[4813]: I1201 08:57:22.312872 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/995100a7-4c5a-409a-81f2-9dbaf2fa1d47-cache\") pod \"swift-storage-0\" (UID: \"995100a7-4c5a-409a-81f2-9dbaf2fa1d47\") " pod="glance-kuttl-tests/swift-storage-0" Dec 01 08:57:22 crc kubenswrapper[4813]: I1201 08:57:22.312902 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/995100a7-4c5a-409a-81f2-9dbaf2fa1d47-etc-swift\") pod \"swift-storage-0\" (UID: \"995100a7-4c5a-409a-81f2-9dbaf2fa1d47\") " pod="glance-kuttl-tests/swift-storage-0" Dec 01 08:57:22 crc kubenswrapper[4813]: I1201 08:57:22.312955 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"swift-storage-0\" (UID: \"995100a7-4c5a-409a-81f2-9dbaf2fa1d47\") " pod="glance-kuttl-tests/swift-storage-0" Dec 01 08:57:22 crc kubenswrapper[4813]: I1201 08:57:22.313032 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6jgvz\" (UniqueName: \"kubernetes.io/projected/995100a7-4c5a-409a-81f2-9dbaf2fa1d47-kube-api-access-6jgvz\") pod \"swift-storage-0\" (UID: \"995100a7-4c5a-409a-81f2-9dbaf2fa1d47\") " pod="glance-kuttl-tests/swift-storage-0" Dec 01 08:57:22 crc kubenswrapper[4813]: I1201 08:57:22.413864 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/995100a7-4c5a-409a-81f2-9dbaf2fa1d47-cache\") pod \"swift-storage-0\" (UID: \"995100a7-4c5a-409a-81f2-9dbaf2fa1d47\") " pod="glance-kuttl-tests/swift-storage-0" Dec 01 08:57:22 crc kubenswrapper[4813]: I1201 08:57:22.413919 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/995100a7-4c5a-409a-81f2-9dbaf2fa1d47-etc-swift\") pod \"swift-storage-0\" (UID: \"995100a7-4c5a-409a-81f2-9dbaf2fa1d47\") " pod="glance-kuttl-tests/swift-storage-0" Dec 01 08:57:22 crc kubenswrapper[4813]: I1201 08:57:22.413978 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"swift-storage-0\" (UID: \"995100a7-4c5a-409a-81f2-9dbaf2fa1d47\") " pod="glance-kuttl-tests/swift-storage-0" Dec 01 08:57:22 crc kubenswrapper[4813]: I1201 08:57:22.414029 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6jgvz\" (UniqueName: \"kubernetes.io/projected/995100a7-4c5a-409a-81f2-9dbaf2fa1d47-kube-api-access-6jgvz\") pod \"swift-storage-0\" (UID: \"995100a7-4c5a-409a-81f2-9dbaf2fa1d47\") " pod="glance-kuttl-tests/swift-storage-0" Dec 01 08:57:22 crc kubenswrapper[4813]: I1201 08:57:22.414090 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/995100a7-4c5a-409a-81f2-9dbaf2fa1d47-lock\") pod \"swift-storage-0\" (UID: \"995100a7-4c5a-409a-81f2-9dbaf2fa1d47\") " pod="glance-kuttl-tests/swift-storage-0" Dec 01 08:57:22 crc kubenswrapper[4813]: E1201 08:57:22.414632 4813 projected.go:288] Couldn't get configMap glance-kuttl-tests/swift-ring-files: configmap "swift-ring-files" not found Dec 01 08:57:22 crc kubenswrapper[4813]: E1201 08:57:22.414687 4813 projected.go:194] Error preparing data for projected volume etc-swift for pod glance-kuttl-tests/swift-storage-0: configmap "swift-ring-files" not found Dec 01 08:57:22 crc kubenswrapper[4813]: I1201 08:57:22.414777 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/995100a7-4c5a-409a-81f2-9dbaf2fa1d47-cache\") pod \"swift-storage-0\" (UID: \"995100a7-4c5a-409a-81f2-9dbaf2fa1d47\") " pod="glance-kuttl-tests/swift-storage-0" Dec 01 08:57:22 crc kubenswrapper[4813]: E1201 08:57:22.414834 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/995100a7-4c5a-409a-81f2-9dbaf2fa1d47-etc-swift podName:995100a7-4c5a-409a-81f2-9dbaf2fa1d47 nodeName:}" failed. No retries permitted until 2025-12-01 08:57:22.914775897 +0000 UTC m=+1027.537697553 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/995100a7-4c5a-409a-81f2-9dbaf2fa1d47-etc-swift") pod "swift-storage-0" (UID: "995100a7-4c5a-409a-81f2-9dbaf2fa1d47") : configmap "swift-ring-files" not found Dec 01 08:57:22 crc kubenswrapper[4813]: I1201 08:57:22.414839 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/995100a7-4c5a-409a-81f2-9dbaf2fa1d47-lock\") pod \"swift-storage-0\" (UID: \"995100a7-4c5a-409a-81f2-9dbaf2fa1d47\") " pod="glance-kuttl-tests/swift-storage-0" Dec 01 08:57:22 crc kubenswrapper[4813]: I1201 08:57:22.415039 4813 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"swift-storage-0\" (UID: \"995100a7-4c5a-409a-81f2-9dbaf2fa1d47\") device mount path \"/mnt/openstack/pv02\"" pod="glance-kuttl-tests/swift-storage-0" Dec 01 08:57:22 crc kubenswrapper[4813]: I1201 08:57:22.433400 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6jgvz\" (UniqueName: \"kubernetes.io/projected/995100a7-4c5a-409a-81f2-9dbaf2fa1d47-kube-api-access-6jgvz\") pod \"swift-storage-0\" (UID: \"995100a7-4c5a-409a-81f2-9dbaf2fa1d47\") " pod="glance-kuttl-tests/swift-storage-0" Dec 01 08:57:22 crc kubenswrapper[4813]: I1201 08:57:22.436496 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"swift-storage-0\" (UID: \"995100a7-4c5a-409a-81f2-9dbaf2fa1d47\") " pod="glance-kuttl-tests/swift-storage-0" Dec 01 08:57:22 crc kubenswrapper[4813]: I1201 08:57:22.890680 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/swift-ring-rebalance-bn9nq"] Dec 01 08:57:22 crc kubenswrapper[4813]: I1201 08:57:22.891787 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/swift-ring-rebalance-bn9nq" Dec 01 08:57:22 crc kubenswrapper[4813]: I1201 08:57:22.896266 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"swift-ring-scripts" Dec 01 08:57:22 crc kubenswrapper[4813]: I1201 08:57:22.896392 4813 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"swift-proxy-config-data" Dec 01 08:57:22 crc kubenswrapper[4813]: I1201 08:57:22.896444 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"swift-ring-config-data" Dec 01 08:57:22 crc kubenswrapper[4813]: I1201 08:57:22.915876 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/swift-ring-rebalance-bn9nq"] Dec 01 08:57:22 crc kubenswrapper[4813]: I1201 08:57:22.930874 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/swift-ring-rebalance-bn9nq"] Dec 01 08:57:22 crc kubenswrapper[4813]: E1201 08:57:22.941580 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[dispersionconf etc-swift kube-api-access-5nz24 ring-data-devices scripts swiftconf], unattached volumes=[], failed to process volumes=[dispersionconf etc-swift kube-api-access-5nz24 ring-data-devices scripts swiftconf]: context canceled" pod="glance-kuttl-tests/swift-ring-rebalance-bn9nq" podUID="40083b10-6d19-4a06-b286-27ce09863cc3" Dec 01 08:57:22 crc kubenswrapper[4813]: I1201 08:57:22.971663 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/swift-ring-rebalance-6bsvb"] Dec 01 08:57:22 crc kubenswrapper[4813]: I1201 08:57:22.972889 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/swift-ring-rebalance-6bsvb" Dec 01 08:57:22 crc kubenswrapper[4813]: I1201 08:57:22.978032 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/881622cc-f398-4ea0-8fef-a443848b664f-etc-swift\") pod \"swift-ring-rebalance-6bsvb\" (UID: \"881622cc-f398-4ea0-8fef-a443848b664f\") " pod="glance-kuttl-tests/swift-ring-rebalance-6bsvb" Dec 01 08:57:22 crc kubenswrapper[4813]: I1201 08:57:22.978124 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/881622cc-f398-4ea0-8fef-a443848b664f-dispersionconf\") pod \"swift-ring-rebalance-6bsvb\" (UID: \"881622cc-f398-4ea0-8fef-a443848b664f\") " pod="glance-kuttl-tests/swift-ring-rebalance-6bsvb" Dec 01 08:57:22 crc kubenswrapper[4813]: I1201 08:57:22.978164 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/881622cc-f398-4ea0-8fef-a443848b664f-swiftconf\") pod \"swift-ring-rebalance-6bsvb\" (UID: \"881622cc-f398-4ea0-8fef-a443848b664f\") " pod="glance-kuttl-tests/swift-ring-rebalance-6bsvb" Dec 01 08:57:22 crc kubenswrapper[4813]: I1201 08:57:22.978184 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/40083b10-6d19-4a06-b286-27ce09863cc3-swiftconf\") pod \"swift-ring-rebalance-bn9nq\" (UID: \"40083b10-6d19-4a06-b286-27ce09863cc3\") " pod="glance-kuttl-tests/swift-ring-rebalance-bn9nq" Dec 01 08:57:22 crc kubenswrapper[4813]: I1201 08:57:22.978209 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/881622cc-f398-4ea0-8fef-a443848b664f-scripts\") pod \"swift-ring-rebalance-6bsvb\" (UID: \"881622cc-f398-4ea0-8fef-a443848b664f\") " pod="glance-kuttl-tests/swift-ring-rebalance-6bsvb" Dec 01 08:57:22 crc kubenswrapper[4813]: I1201 08:57:22.978258 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/881622cc-f398-4ea0-8fef-a443848b664f-ring-data-devices\") pod \"swift-ring-rebalance-6bsvb\" (UID: \"881622cc-f398-4ea0-8fef-a443848b664f\") " pod="glance-kuttl-tests/swift-ring-rebalance-6bsvb" Dec 01 08:57:22 crc kubenswrapper[4813]: I1201 08:57:22.978282 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c5spg\" (UniqueName: \"kubernetes.io/projected/881622cc-f398-4ea0-8fef-a443848b664f-kube-api-access-c5spg\") pod \"swift-ring-rebalance-6bsvb\" (UID: \"881622cc-f398-4ea0-8fef-a443848b664f\") " pod="glance-kuttl-tests/swift-ring-rebalance-6bsvb" Dec 01 08:57:22 crc kubenswrapper[4813]: I1201 08:57:22.978305 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/40083b10-6d19-4a06-b286-27ce09863cc3-scripts\") pod \"swift-ring-rebalance-bn9nq\" (UID: \"40083b10-6d19-4a06-b286-27ce09863cc3\") " pod="glance-kuttl-tests/swift-ring-rebalance-bn9nq" Dec 01 08:57:22 crc kubenswrapper[4813]: I1201 08:57:22.978337 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5nz24\" (UniqueName: \"kubernetes.io/projected/40083b10-6d19-4a06-b286-27ce09863cc3-kube-api-access-5nz24\") pod \"swift-ring-rebalance-bn9nq\" (UID: \"40083b10-6d19-4a06-b286-27ce09863cc3\") " pod="glance-kuttl-tests/swift-ring-rebalance-bn9nq" Dec 01 08:57:22 crc kubenswrapper[4813]: I1201 08:57:22.978414 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/40083b10-6d19-4a06-b286-27ce09863cc3-ring-data-devices\") pod \"swift-ring-rebalance-bn9nq\" (UID: \"40083b10-6d19-4a06-b286-27ce09863cc3\") " pod="glance-kuttl-tests/swift-ring-rebalance-bn9nq" Dec 01 08:57:22 crc kubenswrapper[4813]: I1201 08:57:22.978471 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/40083b10-6d19-4a06-b286-27ce09863cc3-etc-swift\") pod \"swift-ring-rebalance-bn9nq\" (UID: \"40083b10-6d19-4a06-b286-27ce09863cc3\") " pod="glance-kuttl-tests/swift-ring-rebalance-bn9nq" Dec 01 08:57:22 crc kubenswrapper[4813]: I1201 08:57:22.978503 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/995100a7-4c5a-409a-81f2-9dbaf2fa1d47-etc-swift\") pod \"swift-storage-0\" (UID: \"995100a7-4c5a-409a-81f2-9dbaf2fa1d47\") " pod="glance-kuttl-tests/swift-storage-0" Dec 01 08:57:22 crc kubenswrapper[4813]: I1201 08:57:22.978525 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/40083b10-6d19-4a06-b286-27ce09863cc3-dispersionconf\") pod \"swift-ring-rebalance-bn9nq\" (UID: \"40083b10-6d19-4a06-b286-27ce09863cc3\") " pod="glance-kuttl-tests/swift-ring-rebalance-bn9nq" Dec 01 08:57:22 crc kubenswrapper[4813]: E1201 08:57:22.978717 4813 projected.go:288] Couldn't get configMap glance-kuttl-tests/swift-ring-files: configmap "swift-ring-files" not found Dec 01 08:57:22 crc kubenswrapper[4813]: E1201 08:57:22.978739 4813 projected.go:194] Error preparing data for projected volume etc-swift for pod glance-kuttl-tests/swift-storage-0: configmap "swift-ring-files" not found Dec 01 08:57:22 crc kubenswrapper[4813]: E1201 08:57:22.978772 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/995100a7-4c5a-409a-81f2-9dbaf2fa1d47-etc-swift podName:995100a7-4c5a-409a-81f2-9dbaf2fa1d47 nodeName:}" failed. No retries permitted until 2025-12-01 08:57:23.978759849 +0000 UTC m=+1028.601681435 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/995100a7-4c5a-409a-81f2-9dbaf2fa1d47-etc-swift") pod "swift-storage-0" (UID: "995100a7-4c5a-409a-81f2-9dbaf2fa1d47") : configmap "swift-ring-files" not found Dec 01 08:57:22 crc kubenswrapper[4813]: I1201 08:57:22.983953 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/swift-ring-rebalance-6bsvb"] Dec 01 08:57:23 crc kubenswrapper[4813]: I1201 08:57:23.079610 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/40083b10-6d19-4a06-b286-27ce09863cc3-swiftconf\") pod \"swift-ring-rebalance-bn9nq\" (UID: \"40083b10-6d19-4a06-b286-27ce09863cc3\") " pod="glance-kuttl-tests/swift-ring-rebalance-bn9nq" Dec 01 08:57:23 crc kubenswrapper[4813]: I1201 08:57:23.079655 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/881622cc-f398-4ea0-8fef-a443848b664f-scripts\") pod \"swift-ring-rebalance-6bsvb\" (UID: \"881622cc-f398-4ea0-8fef-a443848b664f\") " pod="glance-kuttl-tests/swift-ring-rebalance-6bsvb" Dec 01 08:57:23 crc kubenswrapper[4813]: I1201 08:57:23.079681 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/881622cc-f398-4ea0-8fef-a443848b664f-ring-data-devices\") pod \"swift-ring-rebalance-6bsvb\" (UID: \"881622cc-f398-4ea0-8fef-a443848b664f\") " pod="glance-kuttl-tests/swift-ring-rebalance-6bsvb" Dec 01 08:57:23 crc kubenswrapper[4813]: I1201 08:57:23.079709 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c5spg\" (UniqueName: \"kubernetes.io/projected/881622cc-f398-4ea0-8fef-a443848b664f-kube-api-access-c5spg\") pod \"swift-ring-rebalance-6bsvb\" (UID: \"881622cc-f398-4ea0-8fef-a443848b664f\") " pod="glance-kuttl-tests/swift-ring-rebalance-6bsvb" Dec 01 08:57:23 crc kubenswrapper[4813]: I1201 08:57:23.079736 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/40083b10-6d19-4a06-b286-27ce09863cc3-scripts\") pod \"swift-ring-rebalance-bn9nq\" (UID: \"40083b10-6d19-4a06-b286-27ce09863cc3\") " pod="glance-kuttl-tests/swift-ring-rebalance-bn9nq" Dec 01 08:57:23 crc kubenswrapper[4813]: I1201 08:57:23.079774 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5nz24\" (UniqueName: \"kubernetes.io/projected/40083b10-6d19-4a06-b286-27ce09863cc3-kube-api-access-5nz24\") pod \"swift-ring-rebalance-bn9nq\" (UID: \"40083b10-6d19-4a06-b286-27ce09863cc3\") " pod="glance-kuttl-tests/swift-ring-rebalance-bn9nq" Dec 01 08:57:23 crc kubenswrapper[4813]: I1201 08:57:23.079806 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/40083b10-6d19-4a06-b286-27ce09863cc3-ring-data-devices\") pod \"swift-ring-rebalance-bn9nq\" (UID: \"40083b10-6d19-4a06-b286-27ce09863cc3\") " pod="glance-kuttl-tests/swift-ring-rebalance-bn9nq" Dec 01 08:57:23 crc kubenswrapper[4813]: I1201 08:57:23.079835 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/40083b10-6d19-4a06-b286-27ce09863cc3-etc-swift\") pod \"swift-ring-rebalance-bn9nq\" (UID: \"40083b10-6d19-4a06-b286-27ce09863cc3\") " pod="glance-kuttl-tests/swift-ring-rebalance-bn9nq" Dec 01 08:57:23 crc kubenswrapper[4813]: I1201 08:57:23.079882 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/40083b10-6d19-4a06-b286-27ce09863cc3-dispersionconf\") pod \"swift-ring-rebalance-bn9nq\" (UID: \"40083b10-6d19-4a06-b286-27ce09863cc3\") " pod="glance-kuttl-tests/swift-ring-rebalance-bn9nq" Dec 01 08:57:23 crc kubenswrapper[4813]: I1201 08:57:23.079926 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/881622cc-f398-4ea0-8fef-a443848b664f-etc-swift\") pod \"swift-ring-rebalance-6bsvb\" (UID: \"881622cc-f398-4ea0-8fef-a443848b664f\") " pod="glance-kuttl-tests/swift-ring-rebalance-6bsvb" Dec 01 08:57:23 crc kubenswrapper[4813]: I1201 08:57:23.079951 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/881622cc-f398-4ea0-8fef-a443848b664f-dispersionconf\") pod \"swift-ring-rebalance-6bsvb\" (UID: \"881622cc-f398-4ea0-8fef-a443848b664f\") " pod="glance-kuttl-tests/swift-ring-rebalance-6bsvb" Dec 01 08:57:23 crc kubenswrapper[4813]: I1201 08:57:23.079988 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/881622cc-f398-4ea0-8fef-a443848b664f-swiftconf\") pod \"swift-ring-rebalance-6bsvb\" (UID: \"881622cc-f398-4ea0-8fef-a443848b664f\") " pod="glance-kuttl-tests/swift-ring-rebalance-6bsvb" Dec 01 08:57:23 crc kubenswrapper[4813]: I1201 08:57:23.080810 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/881622cc-f398-4ea0-8fef-a443848b664f-scripts\") pod \"swift-ring-rebalance-6bsvb\" (UID: \"881622cc-f398-4ea0-8fef-a443848b664f\") " pod="glance-kuttl-tests/swift-ring-rebalance-6bsvb" Dec 01 08:57:23 crc kubenswrapper[4813]: I1201 08:57:23.081284 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/40083b10-6d19-4a06-b286-27ce09863cc3-ring-data-devices\") pod \"swift-ring-rebalance-bn9nq\" (UID: \"40083b10-6d19-4a06-b286-27ce09863cc3\") " pod="glance-kuttl-tests/swift-ring-rebalance-bn9nq" Dec 01 08:57:23 crc kubenswrapper[4813]: I1201 08:57:23.081722 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/881622cc-f398-4ea0-8fef-a443848b664f-ring-data-devices\") pod \"swift-ring-rebalance-6bsvb\" (UID: \"881622cc-f398-4ea0-8fef-a443848b664f\") " pod="glance-kuttl-tests/swift-ring-rebalance-6bsvb" Dec 01 08:57:23 crc kubenswrapper[4813]: I1201 08:57:23.082097 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/881622cc-f398-4ea0-8fef-a443848b664f-etc-swift\") pod \"swift-ring-rebalance-6bsvb\" (UID: \"881622cc-f398-4ea0-8fef-a443848b664f\") " pod="glance-kuttl-tests/swift-ring-rebalance-6bsvb" Dec 01 08:57:23 crc kubenswrapper[4813]: I1201 08:57:23.082102 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/40083b10-6d19-4a06-b286-27ce09863cc3-etc-swift\") pod \"swift-ring-rebalance-bn9nq\" (UID: \"40083b10-6d19-4a06-b286-27ce09863cc3\") " pod="glance-kuttl-tests/swift-ring-rebalance-bn9nq" Dec 01 08:57:23 crc kubenswrapper[4813]: I1201 08:57:23.083132 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/40083b10-6d19-4a06-b286-27ce09863cc3-scripts\") pod \"swift-ring-rebalance-bn9nq\" (UID: \"40083b10-6d19-4a06-b286-27ce09863cc3\") " pod="glance-kuttl-tests/swift-ring-rebalance-bn9nq" Dec 01 08:57:23 crc kubenswrapper[4813]: I1201 08:57:23.083952 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/881622cc-f398-4ea0-8fef-a443848b664f-swiftconf\") pod \"swift-ring-rebalance-6bsvb\" (UID: \"881622cc-f398-4ea0-8fef-a443848b664f\") " pod="glance-kuttl-tests/swift-ring-rebalance-6bsvb" Dec 01 08:57:23 crc kubenswrapper[4813]: I1201 08:57:23.090614 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/40083b10-6d19-4a06-b286-27ce09863cc3-swiftconf\") pod \"swift-ring-rebalance-bn9nq\" (UID: \"40083b10-6d19-4a06-b286-27ce09863cc3\") " pod="glance-kuttl-tests/swift-ring-rebalance-bn9nq" Dec 01 08:57:23 crc kubenswrapper[4813]: I1201 08:57:23.090990 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/881622cc-f398-4ea0-8fef-a443848b664f-dispersionconf\") pod \"swift-ring-rebalance-6bsvb\" (UID: \"881622cc-f398-4ea0-8fef-a443848b664f\") " pod="glance-kuttl-tests/swift-ring-rebalance-6bsvb" Dec 01 08:57:23 crc kubenswrapper[4813]: I1201 08:57:23.096399 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/40083b10-6d19-4a06-b286-27ce09863cc3-dispersionconf\") pod \"swift-ring-rebalance-bn9nq\" (UID: \"40083b10-6d19-4a06-b286-27ce09863cc3\") " pod="glance-kuttl-tests/swift-ring-rebalance-bn9nq" Dec 01 08:57:23 crc kubenswrapper[4813]: I1201 08:57:23.101770 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c5spg\" (UniqueName: \"kubernetes.io/projected/881622cc-f398-4ea0-8fef-a443848b664f-kube-api-access-c5spg\") pod \"swift-ring-rebalance-6bsvb\" (UID: \"881622cc-f398-4ea0-8fef-a443848b664f\") " pod="glance-kuttl-tests/swift-ring-rebalance-6bsvb" Dec 01 08:57:23 crc kubenswrapper[4813]: I1201 08:57:23.111435 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5nz24\" (UniqueName: \"kubernetes.io/projected/40083b10-6d19-4a06-b286-27ce09863cc3-kube-api-access-5nz24\") pod \"swift-ring-rebalance-bn9nq\" (UID: \"40083b10-6d19-4a06-b286-27ce09863cc3\") " pod="glance-kuttl-tests/swift-ring-rebalance-bn9nq" Dec 01 08:57:23 crc kubenswrapper[4813]: I1201 08:57:23.293291 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/swift-ring-rebalance-6bsvb" Dec 01 08:57:23 crc kubenswrapper[4813]: I1201 08:57:23.525693 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/swift-ring-rebalance-bn9nq" Dec 01 08:57:23 crc kubenswrapper[4813]: I1201 08:57:23.539141 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/swift-ring-rebalance-bn9nq" Dec 01 08:57:23 crc kubenswrapper[4813]: I1201 08:57:23.573420 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-index-vlgfp"] Dec 01 08:57:23 crc kubenswrapper[4813]: I1201 08:57:23.574737 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-index-vlgfp" Dec 01 08:57:23 crc kubenswrapper[4813]: I1201 08:57:23.577024 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-index-dockercfg-npv4m" Dec 01 08:57:23 crc kubenswrapper[4813]: I1201 08:57:23.582749 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-index-vlgfp"] Dec 01 08:57:23 crc kubenswrapper[4813]: I1201 08:57:23.686246 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/40083b10-6d19-4a06-b286-27ce09863cc3-etc-swift\") pod \"40083b10-6d19-4a06-b286-27ce09863cc3\" (UID: \"40083b10-6d19-4a06-b286-27ce09863cc3\") " Dec 01 08:57:23 crc kubenswrapper[4813]: I1201 08:57:23.686308 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/40083b10-6d19-4a06-b286-27ce09863cc3-swiftconf\") pod \"40083b10-6d19-4a06-b286-27ce09863cc3\" (UID: \"40083b10-6d19-4a06-b286-27ce09863cc3\") " Dec 01 08:57:23 crc kubenswrapper[4813]: I1201 08:57:23.686554 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/40083b10-6d19-4a06-b286-27ce09863cc3-dispersionconf\") pod \"40083b10-6d19-4a06-b286-27ce09863cc3\" (UID: \"40083b10-6d19-4a06-b286-27ce09863cc3\") " Dec 01 08:57:23 crc kubenswrapper[4813]: I1201 08:57:23.686633 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/40083b10-6d19-4a06-b286-27ce09863cc3-ring-data-devices\") pod \"40083b10-6d19-4a06-b286-27ce09863cc3\" (UID: \"40083b10-6d19-4a06-b286-27ce09863cc3\") " Dec 01 08:57:23 crc kubenswrapper[4813]: I1201 08:57:23.686685 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/40083b10-6d19-4a06-b286-27ce09863cc3-scripts\") pod \"40083b10-6d19-4a06-b286-27ce09863cc3\" (UID: \"40083b10-6d19-4a06-b286-27ce09863cc3\") " Dec 01 08:57:23 crc kubenswrapper[4813]: I1201 08:57:23.686717 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5nz24\" (UniqueName: \"kubernetes.io/projected/40083b10-6d19-4a06-b286-27ce09863cc3-kube-api-access-5nz24\") pod \"40083b10-6d19-4a06-b286-27ce09863cc3\" (UID: \"40083b10-6d19-4a06-b286-27ce09863cc3\") " Dec 01 08:57:23 crc kubenswrapper[4813]: I1201 08:57:23.687028 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ftnrl\" (UniqueName: \"kubernetes.io/projected/b95238b9-4f2a-4ee8-8b18-44857550a8d1-kube-api-access-ftnrl\") pod \"glance-operator-index-vlgfp\" (UID: \"b95238b9-4f2a-4ee8-8b18-44857550a8d1\") " pod="openstack-operators/glance-operator-index-vlgfp" Dec 01 08:57:23 crc kubenswrapper[4813]: I1201 08:57:23.686828 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/40083b10-6d19-4a06-b286-27ce09863cc3-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "40083b10-6d19-4a06-b286-27ce09863cc3" (UID: "40083b10-6d19-4a06-b286-27ce09863cc3"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 08:57:23 crc kubenswrapper[4813]: I1201 08:57:23.687143 4813 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/40083b10-6d19-4a06-b286-27ce09863cc3-etc-swift\") on node \"crc\" DevicePath \"\"" Dec 01 08:57:23 crc kubenswrapper[4813]: I1201 08:57:23.687183 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/40083b10-6d19-4a06-b286-27ce09863cc3-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "40083b10-6d19-4a06-b286-27ce09863cc3" (UID: "40083b10-6d19-4a06-b286-27ce09863cc3"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:57:23 crc kubenswrapper[4813]: I1201 08:57:23.687199 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/40083b10-6d19-4a06-b286-27ce09863cc3-scripts" (OuterVolumeSpecName: "scripts") pod "40083b10-6d19-4a06-b286-27ce09863cc3" (UID: "40083b10-6d19-4a06-b286-27ce09863cc3"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:57:23 crc kubenswrapper[4813]: I1201 08:57:23.690253 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/40083b10-6d19-4a06-b286-27ce09863cc3-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "40083b10-6d19-4a06-b286-27ce09863cc3" (UID: "40083b10-6d19-4a06-b286-27ce09863cc3"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:57:23 crc kubenswrapper[4813]: I1201 08:57:23.690626 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/40083b10-6d19-4a06-b286-27ce09863cc3-kube-api-access-5nz24" (OuterVolumeSpecName: "kube-api-access-5nz24") pod "40083b10-6d19-4a06-b286-27ce09863cc3" (UID: "40083b10-6d19-4a06-b286-27ce09863cc3"). InnerVolumeSpecName "kube-api-access-5nz24". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:57:23 crc kubenswrapper[4813]: I1201 08:57:23.691188 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/40083b10-6d19-4a06-b286-27ce09863cc3-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "40083b10-6d19-4a06-b286-27ce09863cc3" (UID: "40083b10-6d19-4a06-b286-27ce09863cc3"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:57:23 crc kubenswrapper[4813]: I1201 08:57:23.760996 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/swift-ring-rebalance-6bsvb"] Dec 01 08:57:23 crc kubenswrapper[4813]: W1201 08:57:23.765536 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod881622cc_f398_4ea0_8fef_a443848b664f.slice/crio-4a18090a04f4f00a90d4d7a9c79f4d0a0978cd11763577c4075bf2599606c786 WatchSource:0}: Error finding container 4a18090a04f4f00a90d4d7a9c79f4d0a0978cd11763577c4075bf2599606c786: Status 404 returned error can't find the container with id 4a18090a04f4f00a90d4d7a9c79f4d0a0978cd11763577c4075bf2599606c786 Dec 01 08:57:23 crc kubenswrapper[4813]: I1201 08:57:23.788028 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ftnrl\" (UniqueName: \"kubernetes.io/projected/b95238b9-4f2a-4ee8-8b18-44857550a8d1-kube-api-access-ftnrl\") pod \"glance-operator-index-vlgfp\" (UID: \"b95238b9-4f2a-4ee8-8b18-44857550a8d1\") " pod="openstack-operators/glance-operator-index-vlgfp" Dec 01 08:57:23 crc kubenswrapper[4813]: I1201 08:57:23.788153 4813 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/40083b10-6d19-4a06-b286-27ce09863cc3-swiftconf\") on node \"crc\" DevicePath \"\"" Dec 01 08:57:23 crc kubenswrapper[4813]: I1201 08:57:23.788170 4813 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/40083b10-6d19-4a06-b286-27ce09863cc3-dispersionconf\") on node \"crc\" DevicePath \"\"" Dec 01 08:57:23 crc kubenswrapper[4813]: I1201 08:57:23.788183 4813 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/40083b10-6d19-4a06-b286-27ce09863cc3-ring-data-devices\") on node \"crc\" DevicePath \"\"" Dec 01 08:57:23 crc kubenswrapper[4813]: I1201 08:57:23.788194 4813 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/40083b10-6d19-4a06-b286-27ce09863cc3-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 08:57:23 crc kubenswrapper[4813]: I1201 08:57:23.788214 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5nz24\" (UniqueName: \"kubernetes.io/projected/40083b10-6d19-4a06-b286-27ce09863cc3-kube-api-access-5nz24\") on node \"crc\" DevicePath \"\"" Dec 01 08:57:23 crc kubenswrapper[4813]: I1201 08:57:23.821773 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ftnrl\" (UniqueName: \"kubernetes.io/projected/b95238b9-4f2a-4ee8-8b18-44857550a8d1-kube-api-access-ftnrl\") pod \"glance-operator-index-vlgfp\" (UID: \"b95238b9-4f2a-4ee8-8b18-44857550a8d1\") " pod="openstack-operators/glance-operator-index-vlgfp" Dec 01 08:57:23 crc kubenswrapper[4813]: I1201 08:57:23.889503 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-index-vlgfp" Dec 01 08:57:23 crc kubenswrapper[4813]: I1201 08:57:23.991431 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/995100a7-4c5a-409a-81f2-9dbaf2fa1d47-etc-swift\") pod \"swift-storage-0\" (UID: \"995100a7-4c5a-409a-81f2-9dbaf2fa1d47\") " pod="glance-kuttl-tests/swift-storage-0" Dec 01 08:57:23 crc kubenswrapper[4813]: E1201 08:57:23.991611 4813 projected.go:288] Couldn't get configMap glance-kuttl-tests/swift-ring-files: configmap "swift-ring-files" not found Dec 01 08:57:23 crc kubenswrapper[4813]: E1201 08:57:23.991629 4813 projected.go:194] Error preparing data for projected volume etc-swift for pod glance-kuttl-tests/swift-storage-0: configmap "swift-ring-files" not found Dec 01 08:57:23 crc kubenswrapper[4813]: E1201 08:57:23.991707 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/995100a7-4c5a-409a-81f2-9dbaf2fa1d47-etc-swift podName:995100a7-4c5a-409a-81f2-9dbaf2fa1d47 nodeName:}" failed. No retries permitted until 2025-12-01 08:57:25.991667058 +0000 UTC m=+1030.614588634 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/995100a7-4c5a-409a-81f2-9dbaf2fa1d47-etc-swift") pod "swift-storage-0" (UID: "995100a7-4c5a-409a-81f2-9dbaf2fa1d47") : configmap "swift-ring-files" not found Dec 01 08:57:24 crc kubenswrapper[4813]: I1201 08:57:24.293536 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/keystone-86b5f6997f-8d5h9" Dec 01 08:57:24 crc kubenswrapper[4813]: I1201 08:57:24.311367 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-index-vlgfp"] Dec 01 08:57:24 crc kubenswrapper[4813]: W1201 08:57:24.312052 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb95238b9_4f2a_4ee8_8b18_44857550a8d1.slice/crio-5af7b5b24658c333787f5403d3cccd1a018f85e48e1fb76dff3d2b0a66974019 WatchSource:0}: Error finding container 5af7b5b24658c333787f5403d3cccd1a018f85e48e1fb76dff3d2b0a66974019: Status 404 returned error can't find the container with id 5af7b5b24658c333787f5403d3cccd1a018f85e48e1fb76dff3d2b0a66974019 Dec 01 08:57:24 crc kubenswrapper[4813]: I1201 08:57:24.534003 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-index-vlgfp" event={"ID":"b95238b9-4f2a-4ee8-8b18-44857550a8d1","Type":"ContainerStarted","Data":"5af7b5b24658c333787f5403d3cccd1a018f85e48e1fb76dff3d2b0a66974019"} Dec 01 08:57:24 crc kubenswrapper[4813]: I1201 08:57:24.535469 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-ring-rebalance-6bsvb" event={"ID":"881622cc-f398-4ea0-8fef-a443848b664f","Type":"ContainerStarted","Data":"4a18090a04f4f00a90d4d7a9c79f4d0a0978cd11763577c4075bf2599606c786"} Dec 01 08:57:24 crc kubenswrapper[4813]: I1201 08:57:24.535504 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/swift-ring-rebalance-bn9nq" Dec 01 08:57:24 crc kubenswrapper[4813]: I1201 08:57:24.600122 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/swift-ring-rebalance-bn9nq"] Dec 01 08:57:24 crc kubenswrapper[4813]: I1201 08:57:24.605215 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/swift-ring-rebalance-bn9nq"] Dec 01 08:57:26 crc kubenswrapper[4813]: I1201 08:57:26.053806 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/995100a7-4c5a-409a-81f2-9dbaf2fa1d47-etc-swift\") pod \"swift-storage-0\" (UID: \"995100a7-4c5a-409a-81f2-9dbaf2fa1d47\") " pod="glance-kuttl-tests/swift-storage-0" Dec 01 08:57:26 crc kubenswrapper[4813]: E1201 08:57:26.054034 4813 projected.go:288] Couldn't get configMap glance-kuttl-tests/swift-ring-files: configmap "swift-ring-files" not found Dec 01 08:57:26 crc kubenswrapper[4813]: E1201 08:57:26.054443 4813 projected.go:194] Error preparing data for projected volume etc-swift for pod glance-kuttl-tests/swift-storage-0: configmap "swift-ring-files" not found Dec 01 08:57:26 crc kubenswrapper[4813]: E1201 08:57:26.054514 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/995100a7-4c5a-409a-81f2-9dbaf2fa1d47-etc-swift podName:995100a7-4c5a-409a-81f2-9dbaf2fa1d47 nodeName:}" failed. No retries permitted until 2025-12-01 08:57:30.054492792 +0000 UTC m=+1034.677414378 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/995100a7-4c5a-409a-81f2-9dbaf2fa1d47-etc-swift") pod "swift-storage-0" (UID: "995100a7-4c5a-409a-81f2-9dbaf2fa1d47") : configmap "swift-ring-files" not found Dec 01 08:57:26 crc kubenswrapper[4813]: I1201 08:57:26.416012 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="40083b10-6d19-4a06-b286-27ce09863cc3" path="/var/lib/kubelet/pods/40083b10-6d19-4a06-b286-27ce09863cc3/volumes" Dec 01 08:57:29 crc kubenswrapper[4813]: I1201 08:57:29.520045 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/swift-proxy-6bd58cfcf7-vhtgp"] Dec 01 08:57:29 crc kubenswrapper[4813]: I1201 08:57:29.521561 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-vhtgp" Dec 01 08:57:29 crc kubenswrapper[4813]: I1201 08:57:29.539073 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/swift-proxy-6bd58cfcf7-vhtgp"] Dec 01 08:57:29 crc kubenswrapper[4813]: I1201 08:57:29.722948 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5504ea2a-5701-4d37-8f4f-613caab65413-log-httpd\") pod \"swift-proxy-6bd58cfcf7-vhtgp\" (UID: \"5504ea2a-5701-4d37-8f4f-613caab65413\") " pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-vhtgp" Dec 01 08:57:29 crc kubenswrapper[4813]: I1201 08:57:29.723043 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5504ea2a-5701-4d37-8f4f-613caab65413-run-httpd\") pod \"swift-proxy-6bd58cfcf7-vhtgp\" (UID: \"5504ea2a-5701-4d37-8f4f-613caab65413\") " pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-vhtgp" Dec 01 08:57:29 crc kubenswrapper[4813]: I1201 08:57:29.723112 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/5504ea2a-5701-4d37-8f4f-613caab65413-etc-swift\") pod \"swift-proxy-6bd58cfcf7-vhtgp\" (UID: \"5504ea2a-5701-4d37-8f4f-613caab65413\") " pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-vhtgp" Dec 01 08:57:29 crc kubenswrapper[4813]: I1201 08:57:29.723319 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gbncc\" (UniqueName: \"kubernetes.io/projected/5504ea2a-5701-4d37-8f4f-613caab65413-kube-api-access-gbncc\") pod \"swift-proxy-6bd58cfcf7-vhtgp\" (UID: \"5504ea2a-5701-4d37-8f4f-613caab65413\") " pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-vhtgp" Dec 01 08:57:29 crc kubenswrapper[4813]: I1201 08:57:29.723405 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5504ea2a-5701-4d37-8f4f-613caab65413-config-data\") pod \"swift-proxy-6bd58cfcf7-vhtgp\" (UID: \"5504ea2a-5701-4d37-8f4f-613caab65413\") " pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-vhtgp" Dec 01 08:57:29 crc kubenswrapper[4813]: I1201 08:57:29.824433 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5504ea2a-5701-4d37-8f4f-613caab65413-log-httpd\") pod \"swift-proxy-6bd58cfcf7-vhtgp\" (UID: \"5504ea2a-5701-4d37-8f4f-613caab65413\") " pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-vhtgp" Dec 01 08:57:29 crc kubenswrapper[4813]: I1201 08:57:29.824486 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5504ea2a-5701-4d37-8f4f-613caab65413-run-httpd\") pod \"swift-proxy-6bd58cfcf7-vhtgp\" (UID: \"5504ea2a-5701-4d37-8f4f-613caab65413\") " pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-vhtgp" Dec 01 08:57:29 crc kubenswrapper[4813]: I1201 08:57:29.824554 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/5504ea2a-5701-4d37-8f4f-613caab65413-etc-swift\") pod \"swift-proxy-6bd58cfcf7-vhtgp\" (UID: \"5504ea2a-5701-4d37-8f4f-613caab65413\") " pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-vhtgp" Dec 01 08:57:29 crc kubenswrapper[4813]: I1201 08:57:29.824636 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gbncc\" (UniqueName: \"kubernetes.io/projected/5504ea2a-5701-4d37-8f4f-613caab65413-kube-api-access-gbncc\") pod \"swift-proxy-6bd58cfcf7-vhtgp\" (UID: \"5504ea2a-5701-4d37-8f4f-613caab65413\") " pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-vhtgp" Dec 01 08:57:29 crc kubenswrapper[4813]: I1201 08:57:29.824677 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5504ea2a-5701-4d37-8f4f-613caab65413-config-data\") pod \"swift-proxy-6bd58cfcf7-vhtgp\" (UID: \"5504ea2a-5701-4d37-8f4f-613caab65413\") " pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-vhtgp" Dec 01 08:57:29 crc kubenswrapper[4813]: E1201 08:57:29.824762 4813 projected.go:288] Couldn't get configMap glance-kuttl-tests/swift-ring-files: configmap "swift-ring-files" not found Dec 01 08:57:29 crc kubenswrapper[4813]: E1201 08:57:29.824787 4813 projected.go:194] Error preparing data for projected volume etc-swift for pod glance-kuttl-tests/swift-proxy-6bd58cfcf7-vhtgp: configmap "swift-ring-files" not found Dec 01 08:57:29 crc kubenswrapper[4813]: E1201 08:57:29.824857 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/5504ea2a-5701-4d37-8f4f-613caab65413-etc-swift podName:5504ea2a-5701-4d37-8f4f-613caab65413 nodeName:}" failed. No retries permitted until 2025-12-01 08:57:30.324831019 +0000 UTC m=+1034.947752625 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/5504ea2a-5701-4d37-8f4f-613caab65413-etc-swift") pod "swift-proxy-6bd58cfcf7-vhtgp" (UID: "5504ea2a-5701-4d37-8f4f-613caab65413") : configmap "swift-ring-files" not found Dec 01 08:57:29 crc kubenswrapper[4813]: I1201 08:57:29.825092 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5504ea2a-5701-4d37-8f4f-613caab65413-log-httpd\") pod \"swift-proxy-6bd58cfcf7-vhtgp\" (UID: \"5504ea2a-5701-4d37-8f4f-613caab65413\") " pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-vhtgp" Dec 01 08:57:29 crc kubenswrapper[4813]: I1201 08:57:29.825329 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5504ea2a-5701-4d37-8f4f-613caab65413-run-httpd\") pod \"swift-proxy-6bd58cfcf7-vhtgp\" (UID: \"5504ea2a-5701-4d37-8f4f-613caab65413\") " pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-vhtgp" Dec 01 08:57:29 crc kubenswrapper[4813]: I1201 08:57:29.830001 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5504ea2a-5701-4d37-8f4f-613caab65413-config-data\") pod \"swift-proxy-6bd58cfcf7-vhtgp\" (UID: \"5504ea2a-5701-4d37-8f4f-613caab65413\") " pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-vhtgp" Dec 01 08:57:29 crc kubenswrapper[4813]: I1201 08:57:29.861384 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gbncc\" (UniqueName: \"kubernetes.io/projected/5504ea2a-5701-4d37-8f4f-613caab65413-kube-api-access-gbncc\") pod \"swift-proxy-6bd58cfcf7-vhtgp\" (UID: \"5504ea2a-5701-4d37-8f4f-613caab65413\") " pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-vhtgp" Dec 01 08:57:30 crc kubenswrapper[4813]: I1201 08:57:30.129037 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/995100a7-4c5a-409a-81f2-9dbaf2fa1d47-etc-swift\") pod \"swift-storage-0\" (UID: \"995100a7-4c5a-409a-81f2-9dbaf2fa1d47\") " pod="glance-kuttl-tests/swift-storage-0" Dec 01 08:57:30 crc kubenswrapper[4813]: E1201 08:57:30.129354 4813 projected.go:288] Couldn't get configMap glance-kuttl-tests/swift-ring-files: configmap "swift-ring-files" not found Dec 01 08:57:30 crc kubenswrapper[4813]: E1201 08:57:30.129412 4813 projected.go:194] Error preparing data for projected volume etc-swift for pod glance-kuttl-tests/swift-storage-0: configmap "swift-ring-files" not found Dec 01 08:57:30 crc kubenswrapper[4813]: E1201 08:57:30.129527 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/995100a7-4c5a-409a-81f2-9dbaf2fa1d47-etc-swift podName:995100a7-4c5a-409a-81f2-9dbaf2fa1d47 nodeName:}" failed. No retries permitted until 2025-12-01 08:57:38.129496081 +0000 UTC m=+1042.752417697 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/995100a7-4c5a-409a-81f2-9dbaf2fa1d47-etc-swift") pod "swift-storage-0" (UID: "995100a7-4c5a-409a-81f2-9dbaf2fa1d47") : configmap "swift-ring-files" not found Dec 01 08:57:30 crc kubenswrapper[4813]: I1201 08:57:30.331758 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/5504ea2a-5701-4d37-8f4f-613caab65413-etc-swift\") pod \"swift-proxy-6bd58cfcf7-vhtgp\" (UID: \"5504ea2a-5701-4d37-8f4f-613caab65413\") " pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-vhtgp" Dec 01 08:57:30 crc kubenswrapper[4813]: E1201 08:57:30.331999 4813 projected.go:288] Couldn't get configMap glance-kuttl-tests/swift-ring-files: configmap "swift-ring-files" not found Dec 01 08:57:30 crc kubenswrapper[4813]: E1201 08:57:30.332032 4813 projected.go:194] Error preparing data for projected volume etc-swift for pod glance-kuttl-tests/swift-proxy-6bd58cfcf7-vhtgp: configmap "swift-ring-files" not found Dec 01 08:57:30 crc kubenswrapper[4813]: E1201 08:57:30.332112 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/5504ea2a-5701-4d37-8f4f-613caab65413-etc-swift podName:5504ea2a-5701-4d37-8f4f-613caab65413 nodeName:}" failed. No retries permitted until 2025-12-01 08:57:31.332085981 +0000 UTC m=+1035.955007567 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/5504ea2a-5701-4d37-8f4f-613caab65413-etc-swift") pod "swift-proxy-6bd58cfcf7-vhtgp" (UID: "5504ea2a-5701-4d37-8f4f-613caab65413") : configmap "swift-ring-files" not found Dec 01 08:57:31 crc kubenswrapper[4813]: I1201 08:57:31.346864 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/5504ea2a-5701-4d37-8f4f-613caab65413-etc-swift\") pod \"swift-proxy-6bd58cfcf7-vhtgp\" (UID: \"5504ea2a-5701-4d37-8f4f-613caab65413\") " pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-vhtgp" Dec 01 08:57:31 crc kubenswrapper[4813]: E1201 08:57:31.347057 4813 projected.go:288] Couldn't get configMap glance-kuttl-tests/swift-ring-files: configmap "swift-ring-files" not found Dec 01 08:57:31 crc kubenswrapper[4813]: E1201 08:57:31.347350 4813 projected.go:194] Error preparing data for projected volume etc-swift for pod glance-kuttl-tests/swift-proxy-6bd58cfcf7-vhtgp: configmap "swift-ring-files" not found Dec 01 08:57:31 crc kubenswrapper[4813]: E1201 08:57:31.347420 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/5504ea2a-5701-4d37-8f4f-613caab65413-etc-swift podName:5504ea2a-5701-4d37-8f4f-613caab65413 nodeName:}" failed. No retries permitted until 2025-12-01 08:57:33.347399291 +0000 UTC m=+1037.970320877 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/5504ea2a-5701-4d37-8f4f-613caab65413-etc-swift") pod "swift-proxy-6bd58cfcf7-vhtgp" (UID: "5504ea2a-5701-4d37-8f4f-613caab65413") : configmap "swift-ring-files" not found Dec 01 08:57:31 crc kubenswrapper[4813]: I1201 08:57:31.629824 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-index-vlgfp" event={"ID":"b95238b9-4f2a-4ee8-8b18-44857550a8d1","Type":"ContainerStarted","Data":"d5551be8c8531c0a6392a5c6b07fcad3f7680e8c8d36c88e535facac73b021db"} Dec 01 08:57:31 crc kubenswrapper[4813]: I1201 08:57:31.652898 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-index-vlgfp" podStartSLOduration=5.608803196 podStartE2EDuration="8.65284814s" podCreationTimestamp="2025-12-01 08:57:23 +0000 UTC" firstStartedPulling="2025-12-01 08:57:24.314822008 +0000 UTC m=+1028.937743614" lastFinishedPulling="2025-12-01 08:57:27.358866972 +0000 UTC m=+1031.981788558" observedRunningTime="2025-12-01 08:57:31.647308044 +0000 UTC m=+1036.270229630" watchObservedRunningTime="2025-12-01 08:57:31.65284814 +0000 UTC m=+1036.275769736" Dec 01 08:57:32 crc kubenswrapper[4813]: I1201 08:57:32.639696 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-ring-rebalance-6bsvb" event={"ID":"881622cc-f398-4ea0-8fef-a443848b664f","Type":"ContainerStarted","Data":"290ea1b79b6f4b663007a194b70dbae1e7de36d13ce1e6540f824831f816f892"} Dec 01 08:57:32 crc kubenswrapper[4813]: I1201 08:57:32.665415 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/swift-ring-rebalance-6bsvb" podStartSLOduration=2.6535457190000002 podStartE2EDuration="10.665387737s" podCreationTimestamp="2025-12-01 08:57:22 +0000 UTC" firstStartedPulling="2025-12-01 08:57:23.768305716 +0000 UTC m=+1028.391227302" lastFinishedPulling="2025-12-01 08:57:31.780147734 +0000 UTC m=+1036.403069320" observedRunningTime="2025-12-01 08:57:32.661656252 +0000 UTC m=+1037.284577838" watchObservedRunningTime="2025-12-01 08:57:32.665387737 +0000 UTC m=+1037.288309333" Dec 01 08:57:33 crc kubenswrapper[4813]: I1201 08:57:33.379193 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/5504ea2a-5701-4d37-8f4f-613caab65413-etc-swift\") pod \"swift-proxy-6bd58cfcf7-vhtgp\" (UID: \"5504ea2a-5701-4d37-8f4f-613caab65413\") " pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-vhtgp" Dec 01 08:57:33 crc kubenswrapper[4813]: E1201 08:57:33.379425 4813 projected.go:288] Couldn't get configMap glance-kuttl-tests/swift-ring-files: configmap "swift-ring-files" not found Dec 01 08:57:33 crc kubenswrapper[4813]: E1201 08:57:33.379723 4813 projected.go:194] Error preparing data for projected volume etc-swift for pod glance-kuttl-tests/swift-proxy-6bd58cfcf7-vhtgp: configmap "swift-ring-files" not found Dec 01 08:57:33 crc kubenswrapper[4813]: E1201 08:57:33.379808 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/5504ea2a-5701-4d37-8f4f-613caab65413-etc-swift podName:5504ea2a-5701-4d37-8f4f-613caab65413 nodeName:}" failed. No retries permitted until 2025-12-01 08:57:37.379781098 +0000 UTC m=+1042.002702704 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/5504ea2a-5701-4d37-8f4f-613caab65413-etc-swift") pod "swift-proxy-6bd58cfcf7-vhtgp" (UID: "5504ea2a-5701-4d37-8f4f-613caab65413") : configmap "swift-ring-files" not found Dec 01 08:57:33 crc kubenswrapper[4813]: I1201 08:57:33.890896 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/glance-operator-index-vlgfp" Dec 01 08:57:33 crc kubenswrapper[4813]: I1201 08:57:33.891065 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-index-vlgfp" Dec 01 08:57:33 crc kubenswrapper[4813]: I1201 08:57:33.927826 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/glance-operator-index-vlgfp" Dec 01 08:57:35 crc kubenswrapper[4813]: I1201 08:57:35.705352 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-index-vlgfp" Dec 01 08:57:37 crc kubenswrapper[4813]: I1201 08:57:37.442149 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/5504ea2a-5701-4d37-8f4f-613caab65413-etc-swift\") pod \"swift-proxy-6bd58cfcf7-vhtgp\" (UID: \"5504ea2a-5701-4d37-8f4f-613caab65413\") " pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-vhtgp" Dec 01 08:57:37 crc kubenswrapper[4813]: E1201 08:57:37.442466 4813 projected.go:288] Couldn't get configMap glance-kuttl-tests/swift-ring-files: configmap "swift-ring-files" not found Dec 01 08:57:37 crc kubenswrapper[4813]: E1201 08:57:37.443171 4813 projected.go:194] Error preparing data for projected volume etc-swift for pod glance-kuttl-tests/swift-proxy-6bd58cfcf7-vhtgp: configmap "swift-ring-files" not found Dec 01 08:57:37 crc kubenswrapper[4813]: E1201 08:57:37.443267 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/5504ea2a-5701-4d37-8f4f-613caab65413-etc-swift podName:5504ea2a-5701-4d37-8f4f-613caab65413 nodeName:}" failed. No retries permitted until 2025-12-01 08:57:45.443235727 +0000 UTC m=+1050.066157343 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/5504ea2a-5701-4d37-8f4f-613caab65413-etc-swift") pod "swift-proxy-6bd58cfcf7-vhtgp" (UID: "5504ea2a-5701-4d37-8f4f-613caab65413") : configmap "swift-ring-files" not found Dec 01 08:57:38 crc kubenswrapper[4813]: I1201 08:57:38.154614 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/995100a7-4c5a-409a-81f2-9dbaf2fa1d47-etc-swift\") pod \"swift-storage-0\" (UID: \"995100a7-4c5a-409a-81f2-9dbaf2fa1d47\") " pod="glance-kuttl-tests/swift-storage-0" Dec 01 08:57:38 crc kubenswrapper[4813]: E1201 08:57:38.154813 4813 projected.go:288] Couldn't get configMap glance-kuttl-tests/swift-ring-files: configmap "swift-ring-files" not found Dec 01 08:57:38 crc kubenswrapper[4813]: E1201 08:57:38.154843 4813 projected.go:194] Error preparing data for projected volume etc-swift for pod glance-kuttl-tests/swift-storage-0: configmap "swift-ring-files" not found Dec 01 08:57:38 crc kubenswrapper[4813]: E1201 08:57:38.154906 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/995100a7-4c5a-409a-81f2-9dbaf2fa1d47-etc-swift podName:995100a7-4c5a-409a-81f2-9dbaf2fa1d47 nodeName:}" failed. No retries permitted until 2025-12-01 08:57:54.154883232 +0000 UTC m=+1058.777804818 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/995100a7-4c5a-409a-81f2-9dbaf2fa1d47-etc-swift") pod "swift-storage-0" (UID: "995100a7-4c5a-409a-81f2-9dbaf2fa1d47") : configmap "swift-ring-files" not found Dec 01 08:57:38 crc kubenswrapper[4813]: I1201 08:57:38.221755 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/254139c32b21a718739d6f558ac0bb26827dad1d88305bb3f8b52462e1xbkjm"] Dec 01 08:57:38 crc kubenswrapper[4813]: I1201 08:57:38.223343 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/254139c32b21a718739d6f558ac0bb26827dad1d88305bb3f8b52462e1xbkjm" Dec 01 08:57:38 crc kubenswrapper[4813]: I1201 08:57:38.226194 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-fml9r" Dec 01 08:57:38 crc kubenswrapper[4813]: I1201 08:57:38.235376 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/254139c32b21a718739d6f558ac0bb26827dad1d88305bb3f8b52462e1xbkjm"] Dec 01 08:57:38 crc kubenswrapper[4813]: I1201 08:57:38.261816 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hqnt8\" (UniqueName: \"kubernetes.io/projected/f9042460-3e84-44c7-a99d-2939f19d78d2-kube-api-access-hqnt8\") pod \"254139c32b21a718739d6f558ac0bb26827dad1d88305bb3f8b52462e1xbkjm\" (UID: \"f9042460-3e84-44c7-a99d-2939f19d78d2\") " pod="openstack-operators/254139c32b21a718739d6f558ac0bb26827dad1d88305bb3f8b52462e1xbkjm" Dec 01 08:57:38 crc kubenswrapper[4813]: I1201 08:57:38.262001 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/f9042460-3e84-44c7-a99d-2939f19d78d2-bundle\") pod \"254139c32b21a718739d6f558ac0bb26827dad1d88305bb3f8b52462e1xbkjm\" (UID: \"f9042460-3e84-44c7-a99d-2939f19d78d2\") " pod="openstack-operators/254139c32b21a718739d6f558ac0bb26827dad1d88305bb3f8b52462e1xbkjm" Dec 01 08:57:38 crc kubenswrapper[4813]: I1201 08:57:38.262087 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/f9042460-3e84-44c7-a99d-2939f19d78d2-util\") pod \"254139c32b21a718739d6f558ac0bb26827dad1d88305bb3f8b52462e1xbkjm\" (UID: \"f9042460-3e84-44c7-a99d-2939f19d78d2\") " pod="openstack-operators/254139c32b21a718739d6f558ac0bb26827dad1d88305bb3f8b52462e1xbkjm" Dec 01 08:57:38 crc kubenswrapper[4813]: I1201 08:57:38.363035 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/f9042460-3e84-44c7-a99d-2939f19d78d2-bundle\") pod \"254139c32b21a718739d6f558ac0bb26827dad1d88305bb3f8b52462e1xbkjm\" (UID: \"f9042460-3e84-44c7-a99d-2939f19d78d2\") " pod="openstack-operators/254139c32b21a718739d6f558ac0bb26827dad1d88305bb3f8b52462e1xbkjm" Dec 01 08:57:38 crc kubenswrapper[4813]: I1201 08:57:38.363121 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/f9042460-3e84-44c7-a99d-2939f19d78d2-util\") pod \"254139c32b21a718739d6f558ac0bb26827dad1d88305bb3f8b52462e1xbkjm\" (UID: \"f9042460-3e84-44c7-a99d-2939f19d78d2\") " pod="openstack-operators/254139c32b21a718739d6f558ac0bb26827dad1d88305bb3f8b52462e1xbkjm" Dec 01 08:57:38 crc kubenswrapper[4813]: I1201 08:57:38.363155 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hqnt8\" (UniqueName: \"kubernetes.io/projected/f9042460-3e84-44c7-a99d-2939f19d78d2-kube-api-access-hqnt8\") pod \"254139c32b21a718739d6f558ac0bb26827dad1d88305bb3f8b52462e1xbkjm\" (UID: \"f9042460-3e84-44c7-a99d-2939f19d78d2\") " pod="openstack-operators/254139c32b21a718739d6f558ac0bb26827dad1d88305bb3f8b52462e1xbkjm" Dec 01 08:57:38 crc kubenswrapper[4813]: I1201 08:57:38.363721 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/f9042460-3e84-44c7-a99d-2939f19d78d2-bundle\") pod \"254139c32b21a718739d6f558ac0bb26827dad1d88305bb3f8b52462e1xbkjm\" (UID: \"f9042460-3e84-44c7-a99d-2939f19d78d2\") " pod="openstack-operators/254139c32b21a718739d6f558ac0bb26827dad1d88305bb3f8b52462e1xbkjm" Dec 01 08:57:38 crc kubenswrapper[4813]: I1201 08:57:38.363738 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/f9042460-3e84-44c7-a99d-2939f19d78d2-util\") pod \"254139c32b21a718739d6f558ac0bb26827dad1d88305bb3f8b52462e1xbkjm\" (UID: \"f9042460-3e84-44c7-a99d-2939f19d78d2\") " pod="openstack-operators/254139c32b21a718739d6f558ac0bb26827dad1d88305bb3f8b52462e1xbkjm" Dec 01 08:57:38 crc kubenswrapper[4813]: I1201 08:57:38.385586 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hqnt8\" (UniqueName: \"kubernetes.io/projected/f9042460-3e84-44c7-a99d-2939f19d78d2-kube-api-access-hqnt8\") pod \"254139c32b21a718739d6f558ac0bb26827dad1d88305bb3f8b52462e1xbkjm\" (UID: \"f9042460-3e84-44c7-a99d-2939f19d78d2\") " pod="openstack-operators/254139c32b21a718739d6f558ac0bb26827dad1d88305bb3f8b52462e1xbkjm" Dec 01 08:57:38 crc kubenswrapper[4813]: I1201 08:57:38.542023 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/254139c32b21a718739d6f558ac0bb26827dad1d88305bb3f8b52462e1xbkjm" Dec 01 08:57:38 crc kubenswrapper[4813]: I1201 08:57:38.694790 4813 generic.go:334] "Generic (PLEG): container finished" podID="881622cc-f398-4ea0-8fef-a443848b664f" containerID="290ea1b79b6f4b663007a194b70dbae1e7de36d13ce1e6540f824831f816f892" exitCode=0 Dec 01 08:57:38 crc kubenswrapper[4813]: I1201 08:57:38.695079 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-ring-rebalance-6bsvb" event={"ID":"881622cc-f398-4ea0-8fef-a443848b664f","Type":"ContainerDied","Data":"290ea1b79b6f4b663007a194b70dbae1e7de36d13ce1e6540f824831f816f892"} Dec 01 08:57:38 crc kubenswrapper[4813]: I1201 08:57:38.831994 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/254139c32b21a718739d6f558ac0bb26827dad1d88305bb3f8b52462e1xbkjm"] Dec 01 08:57:39 crc kubenswrapper[4813]: I1201 08:57:39.705216 4813 generic.go:334] "Generic (PLEG): container finished" podID="f9042460-3e84-44c7-a99d-2939f19d78d2" containerID="6ded26056932a862dac56217fb260faaea05eba2290dfb9fffd9893408dd638c" exitCode=0 Dec 01 08:57:39 crc kubenswrapper[4813]: I1201 08:57:39.705359 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/254139c32b21a718739d6f558ac0bb26827dad1d88305bb3f8b52462e1xbkjm" event={"ID":"f9042460-3e84-44c7-a99d-2939f19d78d2","Type":"ContainerDied","Data":"6ded26056932a862dac56217fb260faaea05eba2290dfb9fffd9893408dd638c"} Dec 01 08:57:39 crc kubenswrapper[4813]: I1201 08:57:39.705626 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/254139c32b21a718739d6f558ac0bb26827dad1d88305bb3f8b52462e1xbkjm" event={"ID":"f9042460-3e84-44c7-a99d-2939f19d78d2","Type":"ContainerStarted","Data":"eaa508f6b4a3e08e871b8beeb4b9c18f24248a994212c89d645575a6e3eaf0f0"} Dec 01 08:57:40 crc kubenswrapper[4813]: I1201 08:57:40.075024 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/swift-ring-rebalance-6bsvb" Dec 01 08:57:40 crc kubenswrapper[4813]: I1201 08:57:40.102650 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/881622cc-f398-4ea0-8fef-a443848b664f-scripts\") pod \"881622cc-f398-4ea0-8fef-a443848b664f\" (UID: \"881622cc-f398-4ea0-8fef-a443848b664f\") " Dec 01 08:57:40 crc kubenswrapper[4813]: I1201 08:57:40.102742 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/881622cc-f398-4ea0-8fef-a443848b664f-etc-swift\") pod \"881622cc-f398-4ea0-8fef-a443848b664f\" (UID: \"881622cc-f398-4ea0-8fef-a443848b664f\") " Dec 01 08:57:40 crc kubenswrapper[4813]: I1201 08:57:40.102815 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/881622cc-f398-4ea0-8fef-a443848b664f-swiftconf\") pod \"881622cc-f398-4ea0-8fef-a443848b664f\" (UID: \"881622cc-f398-4ea0-8fef-a443848b664f\") " Dec 01 08:57:40 crc kubenswrapper[4813]: I1201 08:57:40.102843 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c5spg\" (UniqueName: \"kubernetes.io/projected/881622cc-f398-4ea0-8fef-a443848b664f-kube-api-access-c5spg\") pod \"881622cc-f398-4ea0-8fef-a443848b664f\" (UID: \"881622cc-f398-4ea0-8fef-a443848b664f\") " Dec 01 08:57:40 crc kubenswrapper[4813]: I1201 08:57:40.102934 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/881622cc-f398-4ea0-8fef-a443848b664f-ring-data-devices\") pod \"881622cc-f398-4ea0-8fef-a443848b664f\" (UID: \"881622cc-f398-4ea0-8fef-a443848b664f\") " Dec 01 08:57:40 crc kubenswrapper[4813]: I1201 08:57:40.102979 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/881622cc-f398-4ea0-8fef-a443848b664f-dispersionconf\") pod \"881622cc-f398-4ea0-8fef-a443848b664f\" (UID: \"881622cc-f398-4ea0-8fef-a443848b664f\") " Dec 01 08:57:40 crc kubenswrapper[4813]: I1201 08:57:40.103876 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/881622cc-f398-4ea0-8fef-a443848b664f-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "881622cc-f398-4ea0-8fef-a443848b664f" (UID: "881622cc-f398-4ea0-8fef-a443848b664f"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:57:40 crc kubenswrapper[4813]: I1201 08:57:40.104726 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/881622cc-f398-4ea0-8fef-a443848b664f-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "881622cc-f398-4ea0-8fef-a443848b664f" (UID: "881622cc-f398-4ea0-8fef-a443848b664f"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 08:57:40 crc kubenswrapper[4813]: I1201 08:57:40.109837 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/881622cc-f398-4ea0-8fef-a443848b664f-kube-api-access-c5spg" (OuterVolumeSpecName: "kube-api-access-c5spg") pod "881622cc-f398-4ea0-8fef-a443848b664f" (UID: "881622cc-f398-4ea0-8fef-a443848b664f"). InnerVolumeSpecName "kube-api-access-c5spg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:57:40 crc kubenswrapper[4813]: I1201 08:57:40.112406 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/881622cc-f398-4ea0-8fef-a443848b664f-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "881622cc-f398-4ea0-8fef-a443848b664f" (UID: "881622cc-f398-4ea0-8fef-a443848b664f"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:57:40 crc kubenswrapper[4813]: I1201 08:57:40.127729 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/881622cc-f398-4ea0-8fef-a443848b664f-scripts" (OuterVolumeSpecName: "scripts") pod "881622cc-f398-4ea0-8fef-a443848b664f" (UID: "881622cc-f398-4ea0-8fef-a443848b664f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:57:40 crc kubenswrapper[4813]: I1201 08:57:40.127748 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/881622cc-f398-4ea0-8fef-a443848b664f-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "881622cc-f398-4ea0-8fef-a443848b664f" (UID: "881622cc-f398-4ea0-8fef-a443848b664f"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:57:40 crc kubenswrapper[4813]: I1201 08:57:40.204334 4813 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/881622cc-f398-4ea0-8fef-a443848b664f-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 08:57:40 crc kubenswrapper[4813]: I1201 08:57:40.204367 4813 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/881622cc-f398-4ea0-8fef-a443848b664f-etc-swift\") on node \"crc\" DevicePath \"\"" Dec 01 08:57:40 crc kubenswrapper[4813]: I1201 08:57:40.204377 4813 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/881622cc-f398-4ea0-8fef-a443848b664f-swiftconf\") on node \"crc\" DevicePath \"\"" Dec 01 08:57:40 crc kubenswrapper[4813]: I1201 08:57:40.204389 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c5spg\" (UniqueName: \"kubernetes.io/projected/881622cc-f398-4ea0-8fef-a443848b664f-kube-api-access-c5spg\") on node \"crc\" DevicePath \"\"" Dec 01 08:57:40 crc kubenswrapper[4813]: I1201 08:57:40.204403 4813 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/881622cc-f398-4ea0-8fef-a443848b664f-ring-data-devices\") on node \"crc\" DevicePath \"\"" Dec 01 08:57:40 crc kubenswrapper[4813]: I1201 08:57:40.204412 4813 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/881622cc-f398-4ea0-8fef-a443848b664f-dispersionconf\") on node \"crc\" DevicePath \"\"" Dec 01 08:57:40 crc kubenswrapper[4813]: I1201 08:57:40.713905 4813 generic.go:334] "Generic (PLEG): container finished" podID="f9042460-3e84-44c7-a99d-2939f19d78d2" containerID="6c9d4fd47e98ed28fd8994c1016abb1bd81eb6ea18f494e4bf22aeeb2ec24c2d" exitCode=0 Dec 01 08:57:40 crc kubenswrapper[4813]: I1201 08:57:40.714011 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/254139c32b21a718739d6f558ac0bb26827dad1d88305bb3f8b52462e1xbkjm" event={"ID":"f9042460-3e84-44c7-a99d-2939f19d78d2","Type":"ContainerDied","Data":"6c9d4fd47e98ed28fd8994c1016abb1bd81eb6ea18f494e4bf22aeeb2ec24c2d"} Dec 01 08:57:40 crc kubenswrapper[4813]: I1201 08:57:40.716324 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-ring-rebalance-6bsvb" event={"ID":"881622cc-f398-4ea0-8fef-a443848b664f","Type":"ContainerDied","Data":"4a18090a04f4f00a90d4d7a9c79f4d0a0978cd11763577c4075bf2599606c786"} Dec 01 08:57:40 crc kubenswrapper[4813]: I1201 08:57:40.716374 4813 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4a18090a04f4f00a90d4d7a9c79f4d0a0978cd11763577c4075bf2599606c786" Dec 01 08:57:40 crc kubenswrapper[4813]: I1201 08:57:40.716390 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/swift-ring-rebalance-6bsvb" Dec 01 08:57:41 crc kubenswrapper[4813]: I1201 08:57:41.744448 4813 generic.go:334] "Generic (PLEG): container finished" podID="f9042460-3e84-44c7-a99d-2939f19d78d2" containerID="f134099a523e27c706e1cb041a4e91b3cdd27b3f9d19bf30f9407a8691909167" exitCode=0 Dec 01 08:57:41 crc kubenswrapper[4813]: I1201 08:57:41.744551 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/254139c32b21a718739d6f558ac0bb26827dad1d88305bb3f8b52462e1xbkjm" event={"ID":"f9042460-3e84-44c7-a99d-2939f19d78d2","Type":"ContainerDied","Data":"f134099a523e27c706e1cb041a4e91b3cdd27b3f9d19bf30f9407a8691909167"} Dec 01 08:57:43 crc kubenswrapper[4813]: I1201 08:57:43.131632 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/254139c32b21a718739d6f558ac0bb26827dad1d88305bb3f8b52462e1xbkjm" Dec 01 08:57:43 crc kubenswrapper[4813]: I1201 08:57:43.251409 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/f9042460-3e84-44c7-a99d-2939f19d78d2-bundle\") pod \"f9042460-3e84-44c7-a99d-2939f19d78d2\" (UID: \"f9042460-3e84-44c7-a99d-2939f19d78d2\") " Dec 01 08:57:43 crc kubenswrapper[4813]: I1201 08:57:43.251591 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hqnt8\" (UniqueName: \"kubernetes.io/projected/f9042460-3e84-44c7-a99d-2939f19d78d2-kube-api-access-hqnt8\") pod \"f9042460-3e84-44c7-a99d-2939f19d78d2\" (UID: \"f9042460-3e84-44c7-a99d-2939f19d78d2\") " Dec 01 08:57:43 crc kubenswrapper[4813]: I1201 08:57:43.251650 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/f9042460-3e84-44c7-a99d-2939f19d78d2-util\") pod \"f9042460-3e84-44c7-a99d-2939f19d78d2\" (UID: \"f9042460-3e84-44c7-a99d-2939f19d78d2\") " Dec 01 08:57:43 crc kubenswrapper[4813]: I1201 08:57:43.252610 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f9042460-3e84-44c7-a99d-2939f19d78d2-bundle" (OuterVolumeSpecName: "bundle") pod "f9042460-3e84-44c7-a99d-2939f19d78d2" (UID: "f9042460-3e84-44c7-a99d-2939f19d78d2"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 08:57:43 crc kubenswrapper[4813]: I1201 08:57:43.260186 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f9042460-3e84-44c7-a99d-2939f19d78d2-kube-api-access-hqnt8" (OuterVolumeSpecName: "kube-api-access-hqnt8") pod "f9042460-3e84-44c7-a99d-2939f19d78d2" (UID: "f9042460-3e84-44c7-a99d-2939f19d78d2"). InnerVolumeSpecName "kube-api-access-hqnt8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:57:43 crc kubenswrapper[4813]: I1201 08:57:43.275467 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f9042460-3e84-44c7-a99d-2939f19d78d2-util" (OuterVolumeSpecName: "util") pod "f9042460-3e84-44c7-a99d-2939f19d78d2" (UID: "f9042460-3e84-44c7-a99d-2939f19d78d2"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 08:57:43 crc kubenswrapper[4813]: I1201 08:57:43.353104 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hqnt8\" (UniqueName: \"kubernetes.io/projected/f9042460-3e84-44c7-a99d-2939f19d78d2-kube-api-access-hqnt8\") on node \"crc\" DevicePath \"\"" Dec 01 08:57:43 crc kubenswrapper[4813]: I1201 08:57:43.353368 4813 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/f9042460-3e84-44c7-a99d-2939f19d78d2-util\") on node \"crc\" DevicePath \"\"" Dec 01 08:57:43 crc kubenswrapper[4813]: I1201 08:57:43.353376 4813 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/f9042460-3e84-44c7-a99d-2939f19d78d2-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 08:57:43 crc kubenswrapper[4813]: I1201 08:57:43.770248 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/254139c32b21a718739d6f558ac0bb26827dad1d88305bb3f8b52462e1xbkjm" event={"ID":"f9042460-3e84-44c7-a99d-2939f19d78d2","Type":"ContainerDied","Data":"eaa508f6b4a3e08e871b8beeb4b9c18f24248a994212c89d645575a6e3eaf0f0"} Dec 01 08:57:43 crc kubenswrapper[4813]: I1201 08:57:43.770297 4813 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="eaa508f6b4a3e08e871b8beeb4b9c18f24248a994212c89d645575a6e3eaf0f0" Dec 01 08:57:43 crc kubenswrapper[4813]: I1201 08:57:43.770366 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/254139c32b21a718739d6f558ac0bb26827dad1d88305bb3f8b52462e1xbkjm" Dec 01 08:57:45 crc kubenswrapper[4813]: I1201 08:57:45.487369 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/5504ea2a-5701-4d37-8f4f-613caab65413-etc-swift\") pod \"swift-proxy-6bd58cfcf7-vhtgp\" (UID: \"5504ea2a-5701-4d37-8f4f-613caab65413\") " pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-vhtgp" Dec 01 08:57:45 crc kubenswrapper[4813]: I1201 08:57:45.497324 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/5504ea2a-5701-4d37-8f4f-613caab65413-etc-swift\") pod \"swift-proxy-6bd58cfcf7-vhtgp\" (UID: \"5504ea2a-5701-4d37-8f4f-613caab65413\") " pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-vhtgp" Dec 01 08:57:45 crc kubenswrapper[4813]: I1201 08:57:45.745646 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-vhtgp" Dec 01 08:57:46 crc kubenswrapper[4813]: I1201 08:57:46.245424 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/swift-proxy-6bd58cfcf7-vhtgp"] Dec 01 08:57:46 crc kubenswrapper[4813]: W1201 08:57:46.249008 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5504ea2a_5701_4d37_8f4f_613caab65413.slice/crio-c12ece2b54c95f80ea4622f335f60a10f8a8d8daed227503d298eff7feff8e5a WatchSource:0}: Error finding container c12ece2b54c95f80ea4622f335f60a10f8a8d8daed227503d298eff7feff8e5a: Status 404 returned error can't find the container with id c12ece2b54c95f80ea4622f335f60a10f8a8d8daed227503d298eff7feff8e5a Dec 01 08:57:46 crc kubenswrapper[4813]: I1201 08:57:46.797162 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-vhtgp" event={"ID":"5504ea2a-5701-4d37-8f4f-613caab65413","Type":"ContainerStarted","Data":"d57b6a9a7684f7aceb73a0c233d97f348bb0a09a608815152313412c1a8f0cdc"} Dec 01 08:57:46 crc kubenswrapper[4813]: I1201 08:57:46.797551 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-vhtgp" event={"ID":"5504ea2a-5701-4d37-8f4f-613caab65413","Type":"ContainerStarted","Data":"c12ece2b54c95f80ea4622f335f60a10f8a8d8daed227503d298eff7feff8e5a"} Dec 01 08:57:47 crc kubenswrapper[4813]: I1201 08:57:47.221945 4813 patch_prober.go:28] interesting pod/machine-config-daemon-9wn4q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 08:57:47 crc kubenswrapper[4813]: I1201 08:57:47.222092 4813 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9wn4q" podUID="3d423402-e08f-42e9-b60c-6eb9bf2f6d48" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 08:57:47 crc kubenswrapper[4813]: I1201 08:57:47.809558 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-vhtgp" event={"ID":"5504ea2a-5701-4d37-8f4f-613caab65413","Type":"ContainerStarted","Data":"af790186878c3a0dfbe4b88e5964cbbe3ef87f73830bc59c1d00f90328cc6e30"} Dec 01 08:57:47 crc kubenswrapper[4813]: I1201 08:57:47.809737 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-vhtgp" Dec 01 08:57:47 crc kubenswrapper[4813]: I1201 08:57:47.809753 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-vhtgp" Dec 01 08:57:47 crc kubenswrapper[4813]: I1201 08:57:47.835229 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-vhtgp" podStartSLOduration=18.835201583 podStartE2EDuration="18.835201583s" podCreationTimestamp="2025-12-01 08:57:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:57:47.82871234 +0000 UTC m=+1052.451633926" watchObservedRunningTime="2025-12-01 08:57:47.835201583 +0000 UTC m=+1052.458123169" Dec 01 08:57:54 crc kubenswrapper[4813]: I1201 08:57:54.181096 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/995100a7-4c5a-409a-81f2-9dbaf2fa1d47-etc-swift\") pod \"swift-storage-0\" (UID: \"995100a7-4c5a-409a-81f2-9dbaf2fa1d47\") " pod="glance-kuttl-tests/swift-storage-0" Dec 01 08:57:54 crc kubenswrapper[4813]: I1201 08:57:54.191473 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/995100a7-4c5a-409a-81f2-9dbaf2fa1d47-etc-swift\") pod \"swift-storage-0\" (UID: \"995100a7-4c5a-409a-81f2-9dbaf2fa1d47\") " pod="glance-kuttl-tests/swift-storage-0" Dec 01 08:57:54 crc kubenswrapper[4813]: I1201 08:57:54.377983 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/swift-storage-0" Dec 01 08:57:54 crc kubenswrapper[4813]: I1201 08:57:54.857728 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/swift-storage-0"] Dec 01 08:57:54 crc kubenswrapper[4813]: I1201 08:57:54.878814 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"995100a7-4c5a-409a-81f2-9dbaf2fa1d47","Type":"ContainerStarted","Data":"28acebd64f7f122629afcca41dc4293b68b2b5d7a6a7fb39197f92cfceb2c7e1"} Dec 01 08:57:55 crc kubenswrapper[4813]: I1201 08:57:55.749152 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-vhtgp" Dec 01 08:57:55 crc kubenswrapper[4813]: I1201 08:57:55.754844 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-vhtgp" Dec 01 08:57:56 crc kubenswrapper[4813]: I1201 08:57:56.898745 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"995100a7-4c5a-409a-81f2-9dbaf2fa1d47","Type":"ContainerStarted","Data":"0b01d236c0fd4b7e6153fcbb9c688913b9adc82fea28ce6cfa31368fb7072a52"} Dec 01 08:57:56 crc kubenswrapper[4813]: I1201 08:57:56.899400 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"995100a7-4c5a-409a-81f2-9dbaf2fa1d47","Type":"ContainerStarted","Data":"532f0c05bdb2cee021285ec588aac0effe66a65c6f37749e25369d978d0515e4"} Dec 01 08:57:56 crc kubenswrapper[4813]: I1201 08:57:56.899419 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"995100a7-4c5a-409a-81f2-9dbaf2fa1d47","Type":"ContainerStarted","Data":"c9f5cb19a620cfc7b9df0bad296f8f90075cf217fda34e7c9818acd5277ecc4b"} Dec 01 08:57:57 crc kubenswrapper[4813]: I1201 08:57:57.914474 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"995100a7-4c5a-409a-81f2-9dbaf2fa1d47","Type":"ContainerStarted","Data":"2b9455a05903cfc3c8f0fe13912f2aeb0af2368d2f612165ece50ff474a9e935"} Dec 01 08:57:59 crc kubenswrapper[4813]: I1201 08:57:59.936436 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"995100a7-4c5a-409a-81f2-9dbaf2fa1d47","Type":"ContainerStarted","Data":"71ea72b9275533c0074b122f73f7beebac12af351e19c3c9c0e6c0b137e24f6f"} Dec 01 08:57:59 crc kubenswrapper[4813]: I1201 08:57:59.937066 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"995100a7-4c5a-409a-81f2-9dbaf2fa1d47","Type":"ContainerStarted","Data":"6e3df495999a12b94a6ca78f0149ccd17c13cf04f3fc31edaef7a77490b259d1"} Dec 01 08:57:59 crc kubenswrapper[4813]: I1201 08:57:59.937087 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"995100a7-4c5a-409a-81f2-9dbaf2fa1d47","Type":"ContainerStarted","Data":"d2d428af1b1d67ed1e1389a7fb3e3e1ea61d4ffd00a4526c2b45c9655f43d7ca"} Dec 01 08:57:59 crc kubenswrapper[4813]: I1201 08:57:59.937098 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"995100a7-4c5a-409a-81f2-9dbaf2fa1d47","Type":"ContainerStarted","Data":"4a635b3fb091ce09c36d9181c2aa094b2a84e366b1bb19e73b4059e4af3205fa"} Dec 01 08:58:01 crc kubenswrapper[4813]: I1201 08:58:01.955540 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"995100a7-4c5a-409a-81f2-9dbaf2fa1d47","Type":"ContainerStarted","Data":"cfcbba573ff80c783a56d8b194b0bfee5c5bf61df702623a4752329024c48781"} Dec 01 08:58:01 crc kubenswrapper[4813]: I1201 08:58:01.955982 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"995100a7-4c5a-409a-81f2-9dbaf2fa1d47","Type":"ContainerStarted","Data":"72d0f4ae012d4d6cd410f0e6e9f0041810afd974e12e68981d6f390c1ce1cb8a"} Dec 01 08:58:01 crc kubenswrapper[4813]: I1201 08:58:01.955994 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"995100a7-4c5a-409a-81f2-9dbaf2fa1d47","Type":"ContainerStarted","Data":"4cd563e572aa4851bd7a2a6fdce31ea0962d58c89ab9c06f628d341281b7a914"} Dec 01 08:58:01 crc kubenswrapper[4813]: I1201 08:58:01.956004 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"995100a7-4c5a-409a-81f2-9dbaf2fa1d47","Type":"ContainerStarted","Data":"90152b553c2c0696cb360f3d2b45d4e4f1d069d88a813084717ae44e3d7f007e"} Dec 01 08:58:02 crc kubenswrapper[4813]: I1201 08:58:02.975174 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"995100a7-4c5a-409a-81f2-9dbaf2fa1d47","Type":"ContainerStarted","Data":"08b235b36c9b32c3a32345cc647c1bd488a16d47403f387ba92c8337ae7120e1"} Dec 01 08:58:02 crc kubenswrapper[4813]: I1201 08:58:02.975546 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"995100a7-4c5a-409a-81f2-9dbaf2fa1d47","Type":"ContainerStarted","Data":"8a6edcb51ab7c3bb702c42f67b2432c8ace1cb2fa02e6a3db3fe9519ecf7f7f3"} Dec 01 08:58:02 crc kubenswrapper[4813]: I1201 08:58:02.975563 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"995100a7-4c5a-409a-81f2-9dbaf2fa1d47","Type":"ContainerStarted","Data":"b5f856516db2a807cb93878232098a03c9c22a32b0221eba5c835097c972aca5"} Dec 01 08:58:03 crc kubenswrapper[4813]: I1201 08:58:03.039821 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/swift-storage-0" podStartSLOduration=35.896967899 podStartE2EDuration="42.039759708s" podCreationTimestamp="2025-12-01 08:57:21 +0000 UTC" firstStartedPulling="2025-12-01 08:57:54.87089982 +0000 UTC m=+1059.493821406" lastFinishedPulling="2025-12-01 08:58:01.013691619 +0000 UTC m=+1065.636613215" observedRunningTime="2025-12-01 08:58:03.030096286 +0000 UTC m=+1067.653017872" watchObservedRunningTime="2025-12-01 08:58:03.039759708 +0000 UTC m=+1067.662681294" Dec 01 08:58:05 crc kubenswrapper[4813]: I1201 08:58:05.224756 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-controller-manager-59d87c8d7c-vcbsn"] Dec 01 08:58:05 crc kubenswrapper[4813]: E1201 08:58:05.226873 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f9042460-3e84-44c7-a99d-2939f19d78d2" containerName="util" Dec 01 08:58:05 crc kubenswrapper[4813]: I1201 08:58:05.227059 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="f9042460-3e84-44c7-a99d-2939f19d78d2" containerName="util" Dec 01 08:58:05 crc kubenswrapper[4813]: E1201 08:58:05.227218 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f9042460-3e84-44c7-a99d-2939f19d78d2" containerName="pull" Dec 01 08:58:05 crc kubenswrapper[4813]: I1201 08:58:05.227316 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="f9042460-3e84-44c7-a99d-2939f19d78d2" containerName="pull" Dec 01 08:58:05 crc kubenswrapper[4813]: E1201 08:58:05.227435 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="881622cc-f398-4ea0-8fef-a443848b664f" containerName="swift-ring-rebalance" Dec 01 08:58:05 crc kubenswrapper[4813]: I1201 08:58:05.227528 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="881622cc-f398-4ea0-8fef-a443848b664f" containerName="swift-ring-rebalance" Dec 01 08:58:05 crc kubenswrapper[4813]: E1201 08:58:05.227613 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f9042460-3e84-44c7-a99d-2939f19d78d2" containerName="extract" Dec 01 08:58:05 crc kubenswrapper[4813]: I1201 08:58:05.227698 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="f9042460-3e84-44c7-a99d-2939f19d78d2" containerName="extract" Dec 01 08:58:05 crc kubenswrapper[4813]: I1201 08:58:05.228042 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="f9042460-3e84-44c7-a99d-2939f19d78d2" containerName="extract" Dec 01 08:58:05 crc kubenswrapper[4813]: I1201 08:58:05.228168 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="881622cc-f398-4ea0-8fef-a443848b664f" containerName="swift-ring-rebalance" Dec 01 08:58:05 crc kubenswrapper[4813]: I1201 08:58:05.229012 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-59d87c8d7c-vcbsn" Dec 01 08:58:05 crc kubenswrapper[4813]: I1201 08:58:05.231072 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-service-cert" Dec 01 08:58:05 crc kubenswrapper[4813]: I1201 08:58:05.231072 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-dockercfg-bczvs" Dec 01 08:58:05 crc kubenswrapper[4813]: I1201 08:58:05.235195 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-59d87c8d7c-vcbsn"] Dec 01 08:58:05 crc kubenswrapper[4813]: I1201 08:58:05.391638 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-58dbm\" (UniqueName: \"kubernetes.io/projected/3e52d3b1-2c61-4e2b-8222-35745063e838-kube-api-access-58dbm\") pod \"glance-operator-controller-manager-59d87c8d7c-vcbsn\" (UID: \"3e52d3b1-2c61-4e2b-8222-35745063e838\") " pod="openstack-operators/glance-operator-controller-manager-59d87c8d7c-vcbsn" Dec 01 08:58:05 crc kubenswrapper[4813]: I1201 08:58:05.391743 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/3e52d3b1-2c61-4e2b-8222-35745063e838-webhook-cert\") pod \"glance-operator-controller-manager-59d87c8d7c-vcbsn\" (UID: \"3e52d3b1-2c61-4e2b-8222-35745063e838\") " pod="openstack-operators/glance-operator-controller-manager-59d87c8d7c-vcbsn" Dec 01 08:58:05 crc kubenswrapper[4813]: I1201 08:58:05.391809 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/3e52d3b1-2c61-4e2b-8222-35745063e838-apiservice-cert\") pod \"glance-operator-controller-manager-59d87c8d7c-vcbsn\" (UID: \"3e52d3b1-2c61-4e2b-8222-35745063e838\") " pod="openstack-operators/glance-operator-controller-manager-59d87c8d7c-vcbsn" Dec 01 08:58:05 crc kubenswrapper[4813]: I1201 08:58:05.493387 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/3e52d3b1-2c61-4e2b-8222-35745063e838-webhook-cert\") pod \"glance-operator-controller-manager-59d87c8d7c-vcbsn\" (UID: \"3e52d3b1-2c61-4e2b-8222-35745063e838\") " pod="openstack-operators/glance-operator-controller-manager-59d87c8d7c-vcbsn" Dec 01 08:58:05 crc kubenswrapper[4813]: I1201 08:58:05.493526 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/3e52d3b1-2c61-4e2b-8222-35745063e838-apiservice-cert\") pod \"glance-operator-controller-manager-59d87c8d7c-vcbsn\" (UID: \"3e52d3b1-2c61-4e2b-8222-35745063e838\") " pod="openstack-operators/glance-operator-controller-manager-59d87c8d7c-vcbsn" Dec 01 08:58:05 crc kubenswrapper[4813]: I1201 08:58:05.493620 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-58dbm\" (UniqueName: \"kubernetes.io/projected/3e52d3b1-2c61-4e2b-8222-35745063e838-kube-api-access-58dbm\") pod \"glance-operator-controller-manager-59d87c8d7c-vcbsn\" (UID: \"3e52d3b1-2c61-4e2b-8222-35745063e838\") " pod="openstack-operators/glance-operator-controller-manager-59d87c8d7c-vcbsn" Dec 01 08:58:05 crc kubenswrapper[4813]: I1201 08:58:05.503189 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/3e52d3b1-2c61-4e2b-8222-35745063e838-webhook-cert\") pod \"glance-operator-controller-manager-59d87c8d7c-vcbsn\" (UID: \"3e52d3b1-2c61-4e2b-8222-35745063e838\") " pod="openstack-operators/glance-operator-controller-manager-59d87c8d7c-vcbsn" Dec 01 08:58:05 crc kubenswrapper[4813]: I1201 08:58:05.503475 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/3e52d3b1-2c61-4e2b-8222-35745063e838-apiservice-cert\") pod \"glance-operator-controller-manager-59d87c8d7c-vcbsn\" (UID: \"3e52d3b1-2c61-4e2b-8222-35745063e838\") " pod="openstack-operators/glance-operator-controller-manager-59d87c8d7c-vcbsn" Dec 01 08:58:05 crc kubenswrapper[4813]: I1201 08:58:05.513737 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-58dbm\" (UniqueName: \"kubernetes.io/projected/3e52d3b1-2c61-4e2b-8222-35745063e838-kube-api-access-58dbm\") pod \"glance-operator-controller-manager-59d87c8d7c-vcbsn\" (UID: \"3e52d3b1-2c61-4e2b-8222-35745063e838\") " pod="openstack-operators/glance-operator-controller-manager-59d87c8d7c-vcbsn" Dec 01 08:58:05 crc kubenswrapper[4813]: I1201 08:58:05.550649 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-59d87c8d7c-vcbsn" Dec 01 08:58:06 crc kubenswrapper[4813]: I1201 08:58:06.034638 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-59d87c8d7c-vcbsn"] Dec 01 08:58:06 crc kubenswrapper[4813]: W1201 08:58:06.043333 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3e52d3b1_2c61_4e2b_8222_35745063e838.slice/crio-712ef1886d55f92e2e2b344468bfbf42d6fb5ade1f15d3b2e6b3cf9b6eeeb647 WatchSource:0}: Error finding container 712ef1886d55f92e2e2b344468bfbf42d6fb5ade1f15d3b2e6b3cf9b6eeeb647: Status 404 returned error can't find the container with id 712ef1886d55f92e2e2b344468bfbf42d6fb5ade1f15d3b2e6b3cf9b6eeeb647 Dec 01 08:58:07 crc kubenswrapper[4813]: I1201 08:58:07.016495 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-59d87c8d7c-vcbsn" event={"ID":"3e52d3b1-2c61-4e2b-8222-35745063e838","Type":"ContainerStarted","Data":"712ef1886d55f92e2e2b344468bfbf42d6fb5ade1f15d3b2e6b3cf9b6eeeb647"} Dec 01 08:58:08 crc kubenswrapper[4813]: I1201 08:58:08.026740 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-59d87c8d7c-vcbsn" event={"ID":"3e52d3b1-2c61-4e2b-8222-35745063e838","Type":"ContainerStarted","Data":"46fb724958d5e9eced4e641498c92ea22e8ac57e844321ca8c097a007224858b"} Dec 01 08:58:08 crc kubenswrapper[4813]: I1201 08:58:08.027110 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-59d87c8d7c-vcbsn" Dec 01 08:58:08 crc kubenswrapper[4813]: I1201 08:58:08.053186 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-controller-manager-59d87c8d7c-vcbsn" podStartSLOduration=1.276721488 podStartE2EDuration="3.05312432s" podCreationTimestamp="2025-12-01 08:58:05 +0000 UTC" firstStartedPulling="2025-12-01 08:58:06.04661298 +0000 UTC m=+1070.669534566" lastFinishedPulling="2025-12-01 08:58:07.823015792 +0000 UTC m=+1072.445937398" observedRunningTime="2025-12-01 08:58:08.044647171 +0000 UTC m=+1072.667568767" watchObservedRunningTime="2025-12-01 08:58:08.05312432 +0000 UTC m=+1072.676045926" Dec 01 08:58:15 crc kubenswrapper[4813]: I1201 08:58:15.555220 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-controller-manager-59d87c8d7c-vcbsn" Dec 01 08:58:17 crc kubenswrapper[4813]: I1201 08:58:17.221557 4813 patch_prober.go:28] interesting pod/machine-config-daemon-9wn4q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 08:58:17 crc kubenswrapper[4813]: I1201 08:58:17.221681 4813 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9wn4q" podUID="3d423402-e08f-42e9-b60c-6eb9bf2f6d48" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 08:58:21 crc kubenswrapper[4813]: I1201 08:58:21.780302 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-d205-account-create-update-mgb49"] Dec 01 08:58:21 crc kubenswrapper[4813]: I1201 08:58:21.782804 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-d205-account-create-update-mgb49" Dec 01 08:58:21 crc kubenswrapper[4813]: I1201 08:58:21.785454 4813 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-db-secret" Dec 01 08:58:21 crc kubenswrapper[4813]: I1201 08:58:21.788403 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-db-create-fb2hf"] Dec 01 08:58:21 crc kubenswrapper[4813]: I1201 08:58:21.790125 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-fb2hf" Dec 01 08:58:21 crc kubenswrapper[4813]: I1201 08:58:21.798161 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-create-fb2hf"] Dec 01 08:58:21 crc kubenswrapper[4813]: I1201 08:58:21.810399 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-d205-account-create-update-mgb49"] Dec 01 08:58:21 crc kubenswrapper[4813]: I1201 08:58:21.833982 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/openstackclient"] Dec 01 08:58:21 crc kubenswrapper[4813]: I1201 08:58:21.835073 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/openstackclient" Dec 01 08:58:21 crc kubenswrapper[4813]: I1201 08:58:21.838059 4813 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"default-dockercfg-r6mcj" Dec 01 08:58:21 crc kubenswrapper[4813]: I1201 08:58:21.838070 4813 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"openstack-config-secret" Dec 01 08:58:21 crc kubenswrapper[4813]: I1201 08:58:21.838890 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"openstack-scripts-9db6gc427h" Dec 01 08:58:21 crc kubenswrapper[4813]: I1201 08:58:21.845655 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"openstack-config" Dec 01 08:58:21 crc kubenswrapper[4813]: I1201 08:58:21.854999 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/openstackclient"] Dec 01 08:58:21 crc kubenswrapper[4813]: I1201 08:58:21.864978 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7q9h5\" (UniqueName: \"kubernetes.io/projected/7dc297d9-6c3f-49e7-bb11-25697b36da14-kube-api-access-7q9h5\") pod \"glance-d205-account-create-update-mgb49\" (UID: \"7dc297d9-6c3f-49e7-bb11-25697b36da14\") " pod="glance-kuttl-tests/glance-d205-account-create-update-mgb49" Dec 01 08:58:21 crc kubenswrapper[4813]: I1201 08:58:21.865040 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/905a377a-0608-4830-abc3-10b9404846b7-operator-scripts\") pod \"glance-db-create-fb2hf\" (UID: \"905a377a-0608-4830-abc3-10b9404846b7\") " pod="glance-kuttl-tests/glance-db-create-fb2hf" Dec 01 08:58:21 crc kubenswrapper[4813]: I1201 08:58:21.865090 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7dc297d9-6c3f-49e7-bb11-25697b36da14-operator-scripts\") pod \"glance-d205-account-create-update-mgb49\" (UID: \"7dc297d9-6c3f-49e7-bb11-25697b36da14\") " pod="glance-kuttl-tests/glance-d205-account-create-update-mgb49" Dec 01 08:58:21 crc kubenswrapper[4813]: I1201 08:58:21.865305 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lrxjb\" (UniqueName: \"kubernetes.io/projected/905a377a-0608-4830-abc3-10b9404846b7-kube-api-access-lrxjb\") pod \"glance-db-create-fb2hf\" (UID: \"905a377a-0608-4830-abc3-10b9404846b7\") " pod="glance-kuttl-tests/glance-db-create-fb2hf" Dec 01 08:58:21 crc kubenswrapper[4813]: I1201 08:58:21.967383 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/79ccaf0c-6d3c-4539-b716-72d794c0eb64-openstack-config\") pod \"openstackclient\" (UID: \"79ccaf0c-6d3c-4539-b716-72d794c0eb64\") " pod="glance-kuttl-tests/openstackclient" Dec 01 08:58:21 crc kubenswrapper[4813]: I1201 08:58:21.967527 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7q9h5\" (UniqueName: \"kubernetes.io/projected/7dc297d9-6c3f-49e7-bb11-25697b36da14-kube-api-access-7q9h5\") pod \"glance-d205-account-create-update-mgb49\" (UID: \"7dc297d9-6c3f-49e7-bb11-25697b36da14\") " pod="glance-kuttl-tests/glance-d205-account-create-update-mgb49" Dec 01 08:58:21 crc kubenswrapper[4813]: I1201 08:58:21.967590 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/905a377a-0608-4830-abc3-10b9404846b7-operator-scripts\") pod \"glance-db-create-fb2hf\" (UID: \"905a377a-0608-4830-abc3-10b9404846b7\") " pod="glance-kuttl-tests/glance-db-create-fb2hf" Dec 01 08:58:21 crc kubenswrapper[4813]: I1201 08:58:21.967637 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7dc297d9-6c3f-49e7-bb11-25697b36da14-operator-scripts\") pod \"glance-d205-account-create-update-mgb49\" (UID: \"7dc297d9-6c3f-49e7-bb11-25697b36da14\") " pod="glance-kuttl-tests/glance-d205-account-create-update-mgb49" Dec 01 08:58:21 crc kubenswrapper[4813]: I1201 08:58:21.967681 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-scripts\" (UniqueName: \"kubernetes.io/configmap/79ccaf0c-6d3c-4539-b716-72d794c0eb64-openstack-scripts\") pod \"openstackclient\" (UID: \"79ccaf0c-6d3c-4539-b716-72d794c0eb64\") " pod="glance-kuttl-tests/openstackclient" Dec 01 08:58:21 crc kubenswrapper[4813]: I1201 08:58:21.967708 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qh2qn\" (UniqueName: \"kubernetes.io/projected/79ccaf0c-6d3c-4539-b716-72d794c0eb64-kube-api-access-qh2qn\") pod \"openstackclient\" (UID: \"79ccaf0c-6d3c-4539-b716-72d794c0eb64\") " pod="glance-kuttl-tests/openstackclient" Dec 01 08:58:21 crc kubenswrapper[4813]: I1201 08:58:21.967749 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/79ccaf0c-6d3c-4539-b716-72d794c0eb64-openstack-config-secret\") pod \"openstackclient\" (UID: \"79ccaf0c-6d3c-4539-b716-72d794c0eb64\") " pod="glance-kuttl-tests/openstackclient" Dec 01 08:58:21 crc kubenswrapper[4813]: I1201 08:58:21.967954 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lrxjb\" (UniqueName: \"kubernetes.io/projected/905a377a-0608-4830-abc3-10b9404846b7-kube-api-access-lrxjb\") pod \"glance-db-create-fb2hf\" (UID: \"905a377a-0608-4830-abc3-10b9404846b7\") " pod="glance-kuttl-tests/glance-db-create-fb2hf" Dec 01 08:58:21 crc kubenswrapper[4813]: I1201 08:58:21.968635 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7dc297d9-6c3f-49e7-bb11-25697b36da14-operator-scripts\") pod \"glance-d205-account-create-update-mgb49\" (UID: \"7dc297d9-6c3f-49e7-bb11-25697b36da14\") " pod="glance-kuttl-tests/glance-d205-account-create-update-mgb49" Dec 01 08:58:21 crc kubenswrapper[4813]: I1201 08:58:21.968912 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/905a377a-0608-4830-abc3-10b9404846b7-operator-scripts\") pod \"glance-db-create-fb2hf\" (UID: \"905a377a-0608-4830-abc3-10b9404846b7\") " pod="glance-kuttl-tests/glance-db-create-fb2hf" Dec 01 08:58:21 crc kubenswrapper[4813]: I1201 08:58:21.988276 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lrxjb\" (UniqueName: \"kubernetes.io/projected/905a377a-0608-4830-abc3-10b9404846b7-kube-api-access-lrxjb\") pod \"glance-db-create-fb2hf\" (UID: \"905a377a-0608-4830-abc3-10b9404846b7\") " pod="glance-kuttl-tests/glance-db-create-fb2hf" Dec 01 08:58:22 crc kubenswrapper[4813]: I1201 08:58:22.001105 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7q9h5\" (UniqueName: \"kubernetes.io/projected/7dc297d9-6c3f-49e7-bb11-25697b36da14-kube-api-access-7q9h5\") pod \"glance-d205-account-create-update-mgb49\" (UID: \"7dc297d9-6c3f-49e7-bb11-25697b36da14\") " pod="glance-kuttl-tests/glance-d205-account-create-update-mgb49" Dec 01 08:58:22 crc kubenswrapper[4813]: I1201 08:58:22.070535 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-scripts\" (UniqueName: \"kubernetes.io/configmap/79ccaf0c-6d3c-4539-b716-72d794c0eb64-openstack-scripts\") pod \"openstackclient\" (UID: \"79ccaf0c-6d3c-4539-b716-72d794c0eb64\") " pod="glance-kuttl-tests/openstackclient" Dec 01 08:58:22 crc kubenswrapper[4813]: I1201 08:58:22.070626 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qh2qn\" (UniqueName: \"kubernetes.io/projected/79ccaf0c-6d3c-4539-b716-72d794c0eb64-kube-api-access-qh2qn\") pod \"openstackclient\" (UID: \"79ccaf0c-6d3c-4539-b716-72d794c0eb64\") " pod="glance-kuttl-tests/openstackclient" Dec 01 08:58:22 crc kubenswrapper[4813]: I1201 08:58:22.070680 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/79ccaf0c-6d3c-4539-b716-72d794c0eb64-openstack-config-secret\") pod \"openstackclient\" (UID: \"79ccaf0c-6d3c-4539-b716-72d794c0eb64\") " pod="glance-kuttl-tests/openstackclient" Dec 01 08:58:22 crc kubenswrapper[4813]: I1201 08:58:22.070827 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/79ccaf0c-6d3c-4539-b716-72d794c0eb64-openstack-config\") pod \"openstackclient\" (UID: \"79ccaf0c-6d3c-4539-b716-72d794c0eb64\") " pod="glance-kuttl-tests/openstackclient" Dec 01 08:58:22 crc kubenswrapper[4813]: I1201 08:58:22.071886 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/79ccaf0c-6d3c-4539-b716-72d794c0eb64-openstack-config\") pod \"openstackclient\" (UID: \"79ccaf0c-6d3c-4539-b716-72d794c0eb64\") " pod="glance-kuttl-tests/openstackclient" Dec 01 08:58:22 crc kubenswrapper[4813]: I1201 08:58:22.072032 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-scripts\" (UniqueName: \"kubernetes.io/configmap/79ccaf0c-6d3c-4539-b716-72d794c0eb64-openstack-scripts\") pod \"openstackclient\" (UID: \"79ccaf0c-6d3c-4539-b716-72d794c0eb64\") " pod="glance-kuttl-tests/openstackclient" Dec 01 08:58:22 crc kubenswrapper[4813]: I1201 08:58:22.074854 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/79ccaf0c-6d3c-4539-b716-72d794c0eb64-openstack-config-secret\") pod \"openstackclient\" (UID: \"79ccaf0c-6d3c-4539-b716-72d794c0eb64\") " pod="glance-kuttl-tests/openstackclient" Dec 01 08:58:22 crc kubenswrapper[4813]: I1201 08:58:22.087137 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qh2qn\" (UniqueName: \"kubernetes.io/projected/79ccaf0c-6d3c-4539-b716-72d794c0eb64-kube-api-access-qh2qn\") pod \"openstackclient\" (UID: \"79ccaf0c-6d3c-4539-b716-72d794c0eb64\") " pod="glance-kuttl-tests/openstackclient" Dec 01 08:58:22 crc kubenswrapper[4813]: I1201 08:58:22.119689 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-d205-account-create-update-mgb49" Dec 01 08:58:22 crc kubenswrapper[4813]: I1201 08:58:22.132875 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-fb2hf" Dec 01 08:58:22 crc kubenswrapper[4813]: I1201 08:58:22.154124 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/openstackclient" Dec 01 08:58:22 crc kubenswrapper[4813]: I1201 08:58:22.577545 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-create-fb2hf"] Dec 01 08:58:22 crc kubenswrapper[4813]: W1201 08:58:22.582079 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod905a377a_0608_4830_abc3_10b9404846b7.slice/crio-c707fa2e243b820eb6d522b31d4c3371797ddbe137fb7c6f5261ea04f410ffd1 WatchSource:0}: Error finding container c707fa2e243b820eb6d522b31d4c3371797ddbe137fb7c6f5261ea04f410ffd1: Status 404 returned error can't find the container with id c707fa2e243b820eb6d522b31d4c3371797ddbe137fb7c6f5261ea04f410ffd1 Dec 01 08:58:22 crc kubenswrapper[4813]: I1201 08:58:22.636996 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/openstackclient"] Dec 01 08:58:22 crc kubenswrapper[4813]: W1201 08:58:22.649299 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod79ccaf0c_6d3c_4539_b716_72d794c0eb64.slice/crio-60fcb86c74ec781407237cddbb211e6fdd1d12376ad6ed875cc559ca6b87f1ad WatchSource:0}: Error finding container 60fcb86c74ec781407237cddbb211e6fdd1d12376ad6ed875cc559ca6b87f1ad: Status 404 returned error can't find the container with id 60fcb86c74ec781407237cddbb211e6fdd1d12376ad6ed875cc559ca6b87f1ad Dec 01 08:58:22 crc kubenswrapper[4813]: I1201 08:58:22.667837 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-d205-account-create-update-mgb49"] Dec 01 08:58:23 crc kubenswrapper[4813]: E1201 08:58:23.097816 4813 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod905a377a_0608_4830_abc3_10b9404846b7.slice/crio-68ba99704890378fa2659521be07b13eef554ebf618621fd31d7efe2d3fe1049.scope\": RecentStats: unable to find data in memory cache]" Dec 01 08:58:23 crc kubenswrapper[4813]: I1201 08:58:23.193455 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstackclient" event={"ID":"79ccaf0c-6d3c-4539-b716-72d794c0eb64","Type":"ContainerStarted","Data":"60fcb86c74ec781407237cddbb211e6fdd1d12376ad6ed875cc559ca6b87f1ad"} Dec 01 08:58:23 crc kubenswrapper[4813]: I1201 08:58:23.195668 4813 generic.go:334] "Generic (PLEG): container finished" podID="905a377a-0608-4830-abc3-10b9404846b7" containerID="68ba99704890378fa2659521be07b13eef554ebf618621fd31d7efe2d3fe1049" exitCode=0 Dec 01 08:58:23 crc kubenswrapper[4813]: I1201 08:58:23.195753 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-create-fb2hf" event={"ID":"905a377a-0608-4830-abc3-10b9404846b7","Type":"ContainerDied","Data":"68ba99704890378fa2659521be07b13eef554ebf618621fd31d7efe2d3fe1049"} Dec 01 08:58:23 crc kubenswrapper[4813]: I1201 08:58:23.196053 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-create-fb2hf" event={"ID":"905a377a-0608-4830-abc3-10b9404846b7","Type":"ContainerStarted","Data":"c707fa2e243b820eb6d522b31d4c3371797ddbe137fb7c6f5261ea04f410ffd1"} Dec 01 08:58:23 crc kubenswrapper[4813]: I1201 08:58:23.197353 4813 generic.go:334] "Generic (PLEG): container finished" podID="7dc297d9-6c3f-49e7-bb11-25697b36da14" containerID="20f85026200b9a4f47c934240e34d5be7ca58f51c42699d023d34b52854312d9" exitCode=0 Dec 01 08:58:23 crc kubenswrapper[4813]: I1201 08:58:23.197405 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-d205-account-create-update-mgb49" event={"ID":"7dc297d9-6c3f-49e7-bb11-25697b36da14","Type":"ContainerDied","Data":"20f85026200b9a4f47c934240e34d5be7ca58f51c42699d023d34b52854312d9"} Dec 01 08:58:23 crc kubenswrapper[4813]: I1201 08:58:23.197452 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-d205-account-create-update-mgb49" event={"ID":"7dc297d9-6c3f-49e7-bb11-25697b36da14","Type":"ContainerStarted","Data":"9cebb6fbae65f6aaca9eb6dc6142c7e2e6d922c082c0cb7b89057a64547735a3"} Dec 01 08:58:24 crc kubenswrapper[4813]: I1201 08:58:24.634127 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-fb2hf" Dec 01 08:58:24 crc kubenswrapper[4813]: I1201 08:58:24.640151 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-d205-account-create-update-mgb49" Dec 01 08:58:24 crc kubenswrapper[4813]: I1201 08:58:24.722451 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lrxjb\" (UniqueName: \"kubernetes.io/projected/905a377a-0608-4830-abc3-10b9404846b7-kube-api-access-lrxjb\") pod \"905a377a-0608-4830-abc3-10b9404846b7\" (UID: \"905a377a-0608-4830-abc3-10b9404846b7\") " Dec 01 08:58:24 crc kubenswrapper[4813]: I1201 08:58:24.722573 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7dc297d9-6c3f-49e7-bb11-25697b36da14-operator-scripts\") pod \"7dc297d9-6c3f-49e7-bb11-25697b36da14\" (UID: \"7dc297d9-6c3f-49e7-bb11-25697b36da14\") " Dec 01 08:58:24 crc kubenswrapper[4813]: I1201 08:58:24.722610 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7q9h5\" (UniqueName: \"kubernetes.io/projected/7dc297d9-6c3f-49e7-bb11-25697b36da14-kube-api-access-7q9h5\") pod \"7dc297d9-6c3f-49e7-bb11-25697b36da14\" (UID: \"7dc297d9-6c3f-49e7-bb11-25697b36da14\") " Dec 01 08:58:24 crc kubenswrapper[4813]: I1201 08:58:24.722655 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/905a377a-0608-4830-abc3-10b9404846b7-operator-scripts\") pod \"905a377a-0608-4830-abc3-10b9404846b7\" (UID: \"905a377a-0608-4830-abc3-10b9404846b7\") " Dec 01 08:58:24 crc kubenswrapper[4813]: I1201 08:58:24.723654 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/905a377a-0608-4830-abc3-10b9404846b7-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "905a377a-0608-4830-abc3-10b9404846b7" (UID: "905a377a-0608-4830-abc3-10b9404846b7"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:58:24 crc kubenswrapper[4813]: I1201 08:58:24.724034 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7dc297d9-6c3f-49e7-bb11-25697b36da14-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "7dc297d9-6c3f-49e7-bb11-25697b36da14" (UID: "7dc297d9-6c3f-49e7-bb11-25697b36da14"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:58:24 crc kubenswrapper[4813]: I1201 08:58:24.743155 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/905a377a-0608-4830-abc3-10b9404846b7-kube-api-access-lrxjb" (OuterVolumeSpecName: "kube-api-access-lrxjb") pod "905a377a-0608-4830-abc3-10b9404846b7" (UID: "905a377a-0608-4830-abc3-10b9404846b7"). InnerVolumeSpecName "kube-api-access-lrxjb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:58:24 crc kubenswrapper[4813]: I1201 08:58:24.743275 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7dc297d9-6c3f-49e7-bb11-25697b36da14-kube-api-access-7q9h5" (OuterVolumeSpecName: "kube-api-access-7q9h5") pod "7dc297d9-6c3f-49e7-bb11-25697b36da14" (UID: "7dc297d9-6c3f-49e7-bb11-25697b36da14"). InnerVolumeSpecName "kube-api-access-7q9h5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:58:24 crc kubenswrapper[4813]: I1201 08:58:24.824021 4813 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7dc297d9-6c3f-49e7-bb11-25697b36da14-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 08:58:24 crc kubenswrapper[4813]: I1201 08:58:24.824087 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7q9h5\" (UniqueName: \"kubernetes.io/projected/7dc297d9-6c3f-49e7-bb11-25697b36da14-kube-api-access-7q9h5\") on node \"crc\" DevicePath \"\"" Dec 01 08:58:24 crc kubenswrapper[4813]: I1201 08:58:24.824099 4813 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/905a377a-0608-4830-abc3-10b9404846b7-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 08:58:24 crc kubenswrapper[4813]: I1201 08:58:24.824108 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lrxjb\" (UniqueName: \"kubernetes.io/projected/905a377a-0608-4830-abc3-10b9404846b7-kube-api-access-lrxjb\") on node \"crc\" DevicePath \"\"" Dec 01 08:58:25 crc kubenswrapper[4813]: I1201 08:58:25.218529 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-fb2hf" Dec 01 08:58:25 crc kubenswrapper[4813]: I1201 08:58:25.218792 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-create-fb2hf" event={"ID":"905a377a-0608-4830-abc3-10b9404846b7","Type":"ContainerDied","Data":"c707fa2e243b820eb6d522b31d4c3371797ddbe137fb7c6f5261ea04f410ffd1"} Dec 01 08:58:25 crc kubenswrapper[4813]: I1201 08:58:25.218841 4813 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c707fa2e243b820eb6d522b31d4c3371797ddbe137fb7c6f5261ea04f410ffd1" Dec 01 08:58:25 crc kubenswrapper[4813]: I1201 08:58:25.220408 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-d205-account-create-update-mgb49" event={"ID":"7dc297d9-6c3f-49e7-bb11-25697b36da14","Type":"ContainerDied","Data":"9cebb6fbae65f6aaca9eb6dc6142c7e2e6d922c082c0cb7b89057a64547735a3"} Dec 01 08:58:25 crc kubenswrapper[4813]: I1201 08:58:25.220431 4813 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9cebb6fbae65f6aaca9eb6dc6142c7e2e6d922c082c0cb7b89057a64547735a3" Dec 01 08:58:25 crc kubenswrapper[4813]: I1201 08:58:25.220490 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-d205-account-create-update-mgb49" Dec 01 08:58:26 crc kubenswrapper[4813]: I1201 08:58:26.885163 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-db-sync-lpj7j"] Dec 01 08:58:26 crc kubenswrapper[4813]: E1201 08:58:26.885908 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7dc297d9-6c3f-49e7-bb11-25697b36da14" containerName="mariadb-account-create-update" Dec 01 08:58:26 crc kubenswrapper[4813]: I1201 08:58:26.885944 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="7dc297d9-6c3f-49e7-bb11-25697b36da14" containerName="mariadb-account-create-update" Dec 01 08:58:26 crc kubenswrapper[4813]: E1201 08:58:26.885996 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="905a377a-0608-4830-abc3-10b9404846b7" containerName="mariadb-database-create" Dec 01 08:58:26 crc kubenswrapper[4813]: I1201 08:58:26.886005 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="905a377a-0608-4830-abc3-10b9404846b7" containerName="mariadb-database-create" Dec 01 08:58:26 crc kubenswrapper[4813]: I1201 08:58:26.886518 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="905a377a-0608-4830-abc3-10b9404846b7" containerName="mariadb-database-create" Dec 01 08:58:26 crc kubenswrapper[4813]: I1201 08:58:26.886543 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="7dc297d9-6c3f-49e7-bb11-25697b36da14" containerName="mariadb-account-create-update" Dec 01 08:58:26 crc kubenswrapper[4813]: I1201 08:58:26.888210 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-lpj7j" Dec 01 08:58:26 crc kubenswrapper[4813]: I1201 08:58:26.898927 4813 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-config-data" Dec 01 08:58:26 crc kubenswrapper[4813]: I1201 08:58:26.899216 4813 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-glance-dockercfg-hfrh7" Dec 01 08:58:26 crc kubenswrapper[4813]: I1201 08:58:26.917597 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-sync-lpj7j"] Dec 01 08:58:27 crc kubenswrapper[4813]: I1201 08:58:27.057074 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/51769471-fc41-44cc-8b88-cb2e913188b9-config-data\") pod \"glance-db-sync-lpj7j\" (UID: \"51769471-fc41-44cc-8b88-cb2e913188b9\") " pod="glance-kuttl-tests/glance-db-sync-lpj7j" Dec 01 08:58:27 crc kubenswrapper[4813]: I1201 08:58:27.057194 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wn4px\" (UniqueName: \"kubernetes.io/projected/51769471-fc41-44cc-8b88-cb2e913188b9-kube-api-access-wn4px\") pod \"glance-db-sync-lpj7j\" (UID: \"51769471-fc41-44cc-8b88-cb2e913188b9\") " pod="glance-kuttl-tests/glance-db-sync-lpj7j" Dec 01 08:58:27 crc kubenswrapper[4813]: I1201 08:58:27.057271 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/51769471-fc41-44cc-8b88-cb2e913188b9-db-sync-config-data\") pod \"glance-db-sync-lpj7j\" (UID: \"51769471-fc41-44cc-8b88-cb2e913188b9\") " pod="glance-kuttl-tests/glance-db-sync-lpj7j" Dec 01 08:58:27 crc kubenswrapper[4813]: I1201 08:58:27.159077 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/51769471-fc41-44cc-8b88-cb2e913188b9-config-data\") pod \"glance-db-sync-lpj7j\" (UID: \"51769471-fc41-44cc-8b88-cb2e913188b9\") " pod="glance-kuttl-tests/glance-db-sync-lpj7j" Dec 01 08:58:27 crc kubenswrapper[4813]: I1201 08:58:27.160541 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wn4px\" (UniqueName: \"kubernetes.io/projected/51769471-fc41-44cc-8b88-cb2e913188b9-kube-api-access-wn4px\") pod \"glance-db-sync-lpj7j\" (UID: \"51769471-fc41-44cc-8b88-cb2e913188b9\") " pod="glance-kuttl-tests/glance-db-sync-lpj7j" Dec 01 08:58:27 crc kubenswrapper[4813]: I1201 08:58:27.160657 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/51769471-fc41-44cc-8b88-cb2e913188b9-db-sync-config-data\") pod \"glance-db-sync-lpj7j\" (UID: \"51769471-fc41-44cc-8b88-cb2e913188b9\") " pod="glance-kuttl-tests/glance-db-sync-lpj7j" Dec 01 08:58:27 crc kubenswrapper[4813]: I1201 08:58:27.164432 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/51769471-fc41-44cc-8b88-cb2e913188b9-db-sync-config-data\") pod \"glance-db-sync-lpj7j\" (UID: \"51769471-fc41-44cc-8b88-cb2e913188b9\") " pod="glance-kuttl-tests/glance-db-sync-lpj7j" Dec 01 08:58:27 crc kubenswrapper[4813]: I1201 08:58:27.169466 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/51769471-fc41-44cc-8b88-cb2e913188b9-config-data\") pod \"glance-db-sync-lpj7j\" (UID: \"51769471-fc41-44cc-8b88-cb2e913188b9\") " pod="glance-kuttl-tests/glance-db-sync-lpj7j" Dec 01 08:58:27 crc kubenswrapper[4813]: I1201 08:58:27.185168 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wn4px\" (UniqueName: \"kubernetes.io/projected/51769471-fc41-44cc-8b88-cb2e913188b9-kube-api-access-wn4px\") pod \"glance-db-sync-lpj7j\" (UID: \"51769471-fc41-44cc-8b88-cb2e913188b9\") " pod="glance-kuttl-tests/glance-db-sync-lpj7j" Dec 01 08:58:27 crc kubenswrapper[4813]: I1201 08:58:27.225298 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-lpj7j" Dec 01 08:58:31 crc kubenswrapper[4813]: I1201 08:58:31.698807 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-sync-lpj7j"] Dec 01 08:58:31 crc kubenswrapper[4813]: W1201 08:58:31.703070 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod51769471_fc41_44cc_8b88_cb2e913188b9.slice/crio-0f5562575186142850b3d471bdadec1ddbedacba63c2e6464cbd41e51b94f7f4 WatchSource:0}: Error finding container 0f5562575186142850b3d471bdadec1ddbedacba63c2e6464cbd41e51b94f7f4: Status 404 returned error can't find the container with id 0f5562575186142850b3d471bdadec1ddbedacba63c2e6464cbd41e51b94f7f4 Dec 01 08:58:32 crc kubenswrapper[4813]: I1201 08:58:32.283871 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstackclient" event={"ID":"79ccaf0c-6d3c-4539-b716-72d794c0eb64","Type":"ContainerStarted","Data":"30a5931573955f0ccc85fb3be8db76491c5ee3689e487770b1c50a739162e8c3"} Dec 01 08:58:32 crc kubenswrapper[4813]: I1201 08:58:32.285016 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-lpj7j" event={"ID":"51769471-fc41-44cc-8b88-cb2e913188b9","Type":"ContainerStarted","Data":"0f5562575186142850b3d471bdadec1ddbedacba63c2e6464cbd41e51b94f7f4"} Dec 01 08:58:36 crc kubenswrapper[4813]: I1201 08:58:36.430592 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/openstackclient" podStartSLOduration=6.771718418 podStartE2EDuration="15.430510479s" podCreationTimestamp="2025-12-01 08:58:21 +0000 UTC" firstStartedPulling="2025-12-01 08:58:22.666074989 +0000 UTC m=+1087.288996575" lastFinishedPulling="2025-12-01 08:58:31.32486705 +0000 UTC m=+1095.947788636" observedRunningTime="2025-12-01 08:58:32.305358244 +0000 UTC m=+1096.928279830" watchObservedRunningTime="2025-12-01 08:58:36.430510479 +0000 UTC m=+1101.053432095" Dec 01 08:58:44 crc kubenswrapper[4813]: I1201 08:58:44.412743 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-lpj7j" event={"ID":"51769471-fc41-44cc-8b88-cb2e913188b9","Type":"ContainerStarted","Data":"47f3a60a2c5f83d711d86643c6edac1f3b9962268dc6d7aed6aa40747477311b"} Dec 01 08:58:44 crc kubenswrapper[4813]: I1201 08:58:44.447000 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-db-sync-lpj7j" podStartSLOduration=6.845647145 podStartE2EDuration="18.446895315s" podCreationTimestamp="2025-12-01 08:58:26 +0000 UTC" firstStartedPulling="2025-12-01 08:58:31.707834592 +0000 UTC m=+1096.330756178" lastFinishedPulling="2025-12-01 08:58:43.309082762 +0000 UTC m=+1107.932004348" observedRunningTime="2025-12-01 08:58:44.432225072 +0000 UTC m=+1109.055146708" watchObservedRunningTime="2025-12-01 08:58:44.446895315 +0000 UTC m=+1109.069816951" Dec 01 08:58:47 crc kubenswrapper[4813]: I1201 08:58:47.222373 4813 patch_prober.go:28] interesting pod/machine-config-daemon-9wn4q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 08:58:47 crc kubenswrapper[4813]: I1201 08:58:47.222537 4813 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9wn4q" podUID="3d423402-e08f-42e9-b60c-6eb9bf2f6d48" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 08:58:47 crc kubenswrapper[4813]: I1201 08:58:47.222635 4813 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-9wn4q" Dec 01 08:58:47 crc kubenswrapper[4813]: I1201 08:58:47.223714 4813 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"f46458d1e4ce790f7980bd9d4f88fad10c8bcf27bf165b2631ed78f8dc68e1b3"} pod="openshift-machine-config-operator/machine-config-daemon-9wn4q" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 01 08:58:47 crc kubenswrapper[4813]: I1201 08:58:47.224240 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-9wn4q" podUID="3d423402-e08f-42e9-b60c-6eb9bf2f6d48" containerName="machine-config-daemon" containerID="cri-o://f46458d1e4ce790f7980bd9d4f88fad10c8bcf27bf165b2631ed78f8dc68e1b3" gracePeriod=600 Dec 01 08:58:47 crc kubenswrapper[4813]: I1201 08:58:47.446542 4813 generic.go:334] "Generic (PLEG): container finished" podID="3d423402-e08f-42e9-b60c-6eb9bf2f6d48" containerID="f46458d1e4ce790f7980bd9d4f88fad10c8bcf27bf165b2631ed78f8dc68e1b3" exitCode=0 Dec 01 08:58:47 crc kubenswrapper[4813]: I1201 08:58:47.446627 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-9wn4q" event={"ID":"3d423402-e08f-42e9-b60c-6eb9bf2f6d48","Type":"ContainerDied","Data":"f46458d1e4ce790f7980bd9d4f88fad10c8bcf27bf165b2631ed78f8dc68e1b3"} Dec 01 08:58:47 crc kubenswrapper[4813]: I1201 08:58:47.446792 4813 scope.go:117] "RemoveContainer" containerID="bc619037491f7f2216e483229bcdbf7a96de1d588814d7688551705019230c50" Dec 01 08:58:48 crc kubenswrapper[4813]: I1201 08:58:48.459457 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-9wn4q" event={"ID":"3d423402-e08f-42e9-b60c-6eb9bf2f6d48","Type":"ContainerStarted","Data":"e5115101af56b8d932bd5a35078afaf20c64fdf9b230d78f43cbc4c834a68bf5"} Dec 01 08:58:50 crc kubenswrapper[4813]: I1201 08:58:50.478879 4813 generic.go:334] "Generic (PLEG): container finished" podID="51769471-fc41-44cc-8b88-cb2e913188b9" containerID="47f3a60a2c5f83d711d86643c6edac1f3b9962268dc6d7aed6aa40747477311b" exitCode=0 Dec 01 08:58:50 crc kubenswrapper[4813]: I1201 08:58:50.479025 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-lpj7j" event={"ID":"51769471-fc41-44cc-8b88-cb2e913188b9","Type":"ContainerDied","Data":"47f3a60a2c5f83d711d86643c6edac1f3b9962268dc6d7aed6aa40747477311b"} Dec 01 08:58:51 crc kubenswrapper[4813]: I1201 08:58:51.807195 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-lpj7j" Dec 01 08:58:51 crc kubenswrapper[4813]: I1201 08:58:51.910441 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/51769471-fc41-44cc-8b88-cb2e913188b9-db-sync-config-data\") pod \"51769471-fc41-44cc-8b88-cb2e913188b9\" (UID: \"51769471-fc41-44cc-8b88-cb2e913188b9\") " Dec 01 08:58:51 crc kubenswrapper[4813]: I1201 08:58:51.910532 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/51769471-fc41-44cc-8b88-cb2e913188b9-config-data\") pod \"51769471-fc41-44cc-8b88-cb2e913188b9\" (UID: \"51769471-fc41-44cc-8b88-cb2e913188b9\") " Dec 01 08:58:51 crc kubenswrapper[4813]: I1201 08:58:51.910574 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wn4px\" (UniqueName: \"kubernetes.io/projected/51769471-fc41-44cc-8b88-cb2e913188b9-kube-api-access-wn4px\") pod \"51769471-fc41-44cc-8b88-cb2e913188b9\" (UID: \"51769471-fc41-44cc-8b88-cb2e913188b9\") " Dec 01 08:58:51 crc kubenswrapper[4813]: I1201 08:58:51.917330 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/51769471-fc41-44cc-8b88-cb2e913188b9-kube-api-access-wn4px" (OuterVolumeSpecName: "kube-api-access-wn4px") pod "51769471-fc41-44cc-8b88-cb2e913188b9" (UID: "51769471-fc41-44cc-8b88-cb2e913188b9"). InnerVolumeSpecName "kube-api-access-wn4px". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:58:51 crc kubenswrapper[4813]: I1201 08:58:51.917604 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/51769471-fc41-44cc-8b88-cb2e913188b9-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "51769471-fc41-44cc-8b88-cb2e913188b9" (UID: "51769471-fc41-44cc-8b88-cb2e913188b9"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:58:51 crc kubenswrapper[4813]: I1201 08:58:51.950261 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/51769471-fc41-44cc-8b88-cb2e913188b9-config-data" (OuterVolumeSpecName: "config-data") pod "51769471-fc41-44cc-8b88-cb2e913188b9" (UID: "51769471-fc41-44cc-8b88-cb2e913188b9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:58:52 crc kubenswrapper[4813]: I1201 08:58:52.012576 4813 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/51769471-fc41-44cc-8b88-cb2e913188b9-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 08:58:52 crc kubenswrapper[4813]: I1201 08:58:52.012608 4813 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/51769471-fc41-44cc-8b88-cb2e913188b9-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 08:58:52 crc kubenswrapper[4813]: I1201 08:58:52.012618 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wn4px\" (UniqueName: \"kubernetes.io/projected/51769471-fc41-44cc-8b88-cb2e913188b9-kube-api-access-wn4px\") on node \"crc\" DevicePath \"\"" Dec 01 08:58:52 crc kubenswrapper[4813]: I1201 08:58:52.499600 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-lpj7j" event={"ID":"51769471-fc41-44cc-8b88-cb2e913188b9","Type":"ContainerDied","Data":"0f5562575186142850b3d471bdadec1ddbedacba63c2e6464cbd41e51b94f7f4"} Dec 01 08:58:52 crc kubenswrapper[4813]: I1201 08:58:52.499680 4813 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0f5562575186142850b3d471bdadec1ddbedacba63c2e6464cbd41e51b94f7f4" Dec 01 08:58:52 crc kubenswrapper[4813]: I1201 08:58:52.499745 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-lpj7j" Dec 01 08:58:53 crc kubenswrapper[4813]: I1201 08:58:53.018006 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Dec 01 08:58:53 crc kubenswrapper[4813]: E1201 08:58:53.018485 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="51769471-fc41-44cc-8b88-cb2e913188b9" containerName="glance-db-sync" Dec 01 08:58:53 crc kubenswrapper[4813]: I1201 08:58:53.018510 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="51769471-fc41-44cc-8b88-cb2e913188b9" containerName="glance-db-sync" Dec 01 08:58:53 crc kubenswrapper[4813]: I1201 08:58:53.018681 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="51769471-fc41-44cc-8b88-cb2e913188b9" containerName="glance-db-sync" Dec 01 08:58:53 crc kubenswrapper[4813]: I1201 08:58:53.019593 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Dec 01 08:58:53 crc kubenswrapper[4813]: I1201 08:58:53.021875 4813 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-default-single-config-data" Dec 01 08:58:53 crc kubenswrapper[4813]: I1201 08:58:53.023077 4813 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-scripts" Dec 01 08:58:53 crc kubenswrapper[4813]: I1201 08:58:53.023377 4813 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-glance-dockercfg-hfrh7" Dec 01 08:58:53 crc kubenswrapper[4813]: I1201 08:58:53.023857 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-single-1"] Dec 01 08:58:53 crc kubenswrapper[4813]: I1201 08:58:53.025364 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-1" Dec 01 08:58:53 crc kubenswrapper[4813]: I1201 08:58:53.043880 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Dec 01 08:58:53 crc kubenswrapper[4813]: I1201 08:58:53.066805 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-single-1"] Dec 01 08:58:53 crc kubenswrapper[4813]: I1201 08:58:53.132100 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/0d38a9cc-3031-42c1-a421-c7446e2ae2a3-run\") pod \"glance-default-single-0\" (UID: \"0d38a9cc-3031-42c1-a421-c7446e2ae2a3\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 01 08:58:53 crc kubenswrapper[4813]: I1201 08:58:53.132347 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/0d38a9cc-3031-42c1-a421-c7446e2ae2a3-etc-iscsi\") pod \"glance-default-single-0\" (UID: \"0d38a9cc-3031-42c1-a421-c7446e2ae2a3\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 01 08:58:53 crc kubenswrapper[4813]: I1201 08:58:53.132367 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/4464d6de-546d-4efe-b4bb-1e49b2b0f147-run\") pod \"glance-default-single-1\" (UID: \"4464d6de-546d-4efe-b4bb-1e49b2b0f147\") " pod="glance-kuttl-tests/glance-default-single-1" Dec 01 08:58:53 crc kubenswrapper[4813]: I1201 08:58:53.132394 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/4464d6de-546d-4efe-b4bb-1e49b2b0f147-var-locks-brick\") pod \"glance-default-single-1\" (UID: \"4464d6de-546d-4efe-b4bb-1e49b2b0f147\") " pod="glance-kuttl-tests/glance-default-single-1" Dec 01 08:58:53 crc kubenswrapper[4813]: I1201 08:58:53.132410 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/0d38a9cc-3031-42c1-a421-c7446e2ae2a3-dev\") pod \"glance-default-single-0\" (UID: \"0d38a9cc-3031-42c1-a421-c7446e2ae2a3\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 01 08:58:53 crc kubenswrapper[4813]: I1201 08:58:53.132442 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-single-1\" (UID: \"4464d6de-546d-4efe-b4bb-1e49b2b0f147\") " pod="glance-kuttl-tests/glance-default-single-1" Dec 01 08:58:53 crc kubenswrapper[4813]: I1201 08:58:53.132459 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4464d6de-546d-4efe-b4bb-1e49b2b0f147-logs\") pod \"glance-default-single-1\" (UID: \"4464d6de-546d-4efe-b4bb-1e49b2b0f147\") " pod="glance-kuttl-tests/glance-default-single-1" Dec 01 08:58:53 crc kubenswrapper[4813]: I1201 08:58:53.132483 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0d38a9cc-3031-42c1-a421-c7446e2ae2a3-scripts\") pod \"glance-default-single-0\" (UID: \"0d38a9cc-3031-42c1-a421-c7446e2ae2a3\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 01 08:58:53 crc kubenswrapper[4813]: I1201 08:58:53.132499 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/0d38a9cc-3031-42c1-a421-c7446e2ae2a3-var-locks-brick\") pod \"glance-default-single-0\" (UID: \"0d38a9cc-3031-42c1-a421-c7446e2ae2a3\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 01 08:58:53 crc kubenswrapper[4813]: I1201 08:58:53.132517 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4464d6de-546d-4efe-b4bb-1e49b2b0f147-scripts\") pod \"glance-default-single-1\" (UID: \"4464d6de-546d-4efe-b4bb-1e49b2b0f147\") " pod="glance-kuttl-tests/glance-default-single-1" Dec 01 08:58:53 crc kubenswrapper[4813]: I1201 08:58:53.132531 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/0d38a9cc-3031-42c1-a421-c7446e2ae2a3-lib-modules\") pod \"glance-default-single-0\" (UID: \"0d38a9cc-3031-42c1-a421-c7446e2ae2a3\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 01 08:58:53 crc kubenswrapper[4813]: I1201 08:58:53.132557 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0d38a9cc-3031-42c1-a421-c7446e2ae2a3-logs\") pod \"glance-default-single-0\" (UID: \"0d38a9cc-3031-42c1-a421-c7446e2ae2a3\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 01 08:58:53 crc kubenswrapper[4813]: I1201 08:58:53.132576 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7lzsr\" (UniqueName: \"kubernetes.io/projected/0d38a9cc-3031-42c1-a421-c7446e2ae2a3-kube-api-access-7lzsr\") pod \"glance-default-single-0\" (UID: \"0d38a9cc-3031-42c1-a421-c7446e2ae2a3\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 01 08:58:53 crc kubenswrapper[4813]: I1201 08:58:53.132595 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/4464d6de-546d-4efe-b4bb-1e49b2b0f147-dev\") pod \"glance-default-single-1\" (UID: \"4464d6de-546d-4efe-b4bb-1e49b2b0f147\") " pod="glance-kuttl-tests/glance-default-single-1" Dec 01 08:58:53 crc kubenswrapper[4813]: I1201 08:58:53.132618 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage16-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage16-crc\") pod \"glance-default-single-0\" (UID: \"0d38a9cc-3031-42c1-a421-c7446e2ae2a3\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 01 08:58:53 crc kubenswrapper[4813]: I1201 08:58:53.132642 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wpmkd\" (UniqueName: \"kubernetes.io/projected/4464d6de-546d-4efe-b4bb-1e49b2b0f147-kube-api-access-wpmkd\") pod \"glance-default-single-1\" (UID: \"4464d6de-546d-4efe-b4bb-1e49b2b0f147\") " pod="glance-kuttl-tests/glance-default-single-1" Dec 01 08:58:53 crc kubenswrapper[4813]: I1201 08:58:53.132670 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/4464d6de-546d-4efe-b4bb-1e49b2b0f147-etc-iscsi\") pod \"glance-default-single-1\" (UID: \"4464d6de-546d-4efe-b4bb-1e49b2b0f147\") " pod="glance-kuttl-tests/glance-default-single-1" Dec 01 08:58:53 crc kubenswrapper[4813]: I1201 08:58:53.132699 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/0d38a9cc-3031-42c1-a421-c7446e2ae2a3-etc-nvme\") pod \"glance-default-single-0\" (UID: \"0d38a9cc-3031-42c1-a421-c7446e2ae2a3\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 01 08:58:53 crc kubenswrapper[4813]: I1201 08:58:53.132721 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/4464d6de-546d-4efe-b4bb-1e49b2b0f147-etc-nvme\") pod \"glance-default-single-1\" (UID: \"4464d6de-546d-4efe-b4bb-1e49b2b0f147\") " pod="glance-kuttl-tests/glance-default-single-1" Dec 01 08:58:53 crc kubenswrapper[4813]: I1201 08:58:53.132735 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-single-0\" (UID: \"0d38a9cc-3031-42c1-a421-c7446e2ae2a3\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 01 08:58:53 crc kubenswrapper[4813]: I1201 08:58:53.132752 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0d38a9cc-3031-42c1-a421-c7446e2ae2a3-config-data\") pod \"glance-default-single-0\" (UID: \"0d38a9cc-3031-42c1-a421-c7446e2ae2a3\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 01 08:58:53 crc kubenswrapper[4813]: I1201 08:58:53.132779 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/4464d6de-546d-4efe-b4bb-1e49b2b0f147-lib-modules\") pod \"glance-default-single-1\" (UID: \"4464d6de-546d-4efe-b4bb-1e49b2b0f147\") " pod="glance-kuttl-tests/glance-default-single-1" Dec 01 08:58:53 crc kubenswrapper[4813]: I1201 08:58:53.132796 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/4464d6de-546d-4efe-b4bb-1e49b2b0f147-httpd-run\") pod \"glance-default-single-1\" (UID: \"4464d6de-546d-4efe-b4bb-1e49b2b0f147\") " pod="glance-kuttl-tests/glance-default-single-1" Dec 01 08:58:53 crc kubenswrapper[4813]: I1201 08:58:53.132811 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/0d38a9cc-3031-42c1-a421-c7446e2ae2a3-httpd-run\") pod \"glance-default-single-0\" (UID: \"0d38a9cc-3031-42c1-a421-c7446e2ae2a3\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 01 08:58:53 crc kubenswrapper[4813]: I1201 08:58:53.132831 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/4464d6de-546d-4efe-b4bb-1e49b2b0f147-sys\") pod \"glance-default-single-1\" (UID: \"4464d6de-546d-4efe-b4bb-1e49b2b0f147\") " pod="glance-kuttl-tests/glance-default-single-1" Dec 01 08:58:53 crc kubenswrapper[4813]: I1201 08:58:53.132853 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-single-1\" (UID: \"4464d6de-546d-4efe-b4bb-1e49b2b0f147\") " pod="glance-kuttl-tests/glance-default-single-1" Dec 01 08:58:53 crc kubenswrapper[4813]: I1201 08:58:53.132867 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4464d6de-546d-4efe-b4bb-1e49b2b0f147-config-data\") pod \"glance-default-single-1\" (UID: \"4464d6de-546d-4efe-b4bb-1e49b2b0f147\") " pod="glance-kuttl-tests/glance-default-single-1" Dec 01 08:58:53 crc kubenswrapper[4813]: I1201 08:58:53.132881 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/0d38a9cc-3031-42c1-a421-c7446e2ae2a3-sys\") pod \"glance-default-single-0\" (UID: \"0d38a9cc-3031-42c1-a421-c7446e2ae2a3\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 01 08:58:53 crc kubenswrapper[4813]: I1201 08:58:53.233940 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/4464d6de-546d-4efe-b4bb-1e49b2b0f147-dev\") pod \"glance-default-single-1\" (UID: \"4464d6de-546d-4efe-b4bb-1e49b2b0f147\") " pod="glance-kuttl-tests/glance-default-single-1" Dec 01 08:58:53 crc kubenswrapper[4813]: I1201 08:58:53.234043 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage16-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage16-crc\") pod \"glance-default-single-0\" (UID: \"0d38a9cc-3031-42c1-a421-c7446e2ae2a3\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 01 08:58:53 crc kubenswrapper[4813]: I1201 08:58:53.234069 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wpmkd\" (UniqueName: \"kubernetes.io/projected/4464d6de-546d-4efe-b4bb-1e49b2b0f147-kube-api-access-wpmkd\") pod \"glance-default-single-1\" (UID: \"4464d6de-546d-4efe-b4bb-1e49b2b0f147\") " pod="glance-kuttl-tests/glance-default-single-1" Dec 01 08:58:53 crc kubenswrapper[4813]: I1201 08:58:53.234086 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/4464d6de-546d-4efe-b4bb-1e49b2b0f147-dev\") pod \"glance-default-single-1\" (UID: \"4464d6de-546d-4efe-b4bb-1e49b2b0f147\") " pod="glance-kuttl-tests/glance-default-single-1" Dec 01 08:58:53 crc kubenswrapper[4813]: I1201 08:58:53.234508 4813 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage16-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage16-crc\") pod \"glance-default-single-0\" (UID: \"0d38a9cc-3031-42c1-a421-c7446e2ae2a3\") device mount path \"/mnt/openstack/pv16\"" pod="glance-kuttl-tests/glance-default-single-0" Dec 01 08:58:53 crc kubenswrapper[4813]: I1201 08:58:53.242123 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/4464d6de-546d-4efe-b4bb-1e49b2b0f147-etc-iscsi\") pod \"glance-default-single-1\" (UID: \"4464d6de-546d-4efe-b4bb-1e49b2b0f147\") " pod="glance-kuttl-tests/glance-default-single-1" Dec 01 08:58:53 crc kubenswrapper[4813]: I1201 08:58:53.242190 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/4464d6de-546d-4efe-b4bb-1e49b2b0f147-etc-iscsi\") pod \"glance-default-single-1\" (UID: \"4464d6de-546d-4efe-b4bb-1e49b2b0f147\") " pod="glance-kuttl-tests/glance-default-single-1" Dec 01 08:58:53 crc kubenswrapper[4813]: I1201 08:58:53.242220 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/0d38a9cc-3031-42c1-a421-c7446e2ae2a3-etc-nvme\") pod \"glance-default-single-0\" (UID: \"0d38a9cc-3031-42c1-a421-c7446e2ae2a3\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 01 08:58:53 crc kubenswrapper[4813]: I1201 08:58:53.242268 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/4464d6de-546d-4efe-b4bb-1e49b2b0f147-etc-nvme\") pod \"glance-default-single-1\" (UID: \"4464d6de-546d-4efe-b4bb-1e49b2b0f147\") " pod="glance-kuttl-tests/glance-default-single-1" Dec 01 08:58:53 crc kubenswrapper[4813]: I1201 08:58:53.242301 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-single-0\" (UID: \"0d38a9cc-3031-42c1-a421-c7446e2ae2a3\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 01 08:58:53 crc kubenswrapper[4813]: I1201 08:58:53.242339 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0d38a9cc-3031-42c1-a421-c7446e2ae2a3-config-data\") pod \"glance-default-single-0\" (UID: \"0d38a9cc-3031-42c1-a421-c7446e2ae2a3\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 01 08:58:53 crc kubenswrapper[4813]: I1201 08:58:53.242373 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/4464d6de-546d-4efe-b4bb-1e49b2b0f147-lib-modules\") pod \"glance-default-single-1\" (UID: \"4464d6de-546d-4efe-b4bb-1e49b2b0f147\") " pod="glance-kuttl-tests/glance-default-single-1" Dec 01 08:58:53 crc kubenswrapper[4813]: I1201 08:58:53.242422 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/4464d6de-546d-4efe-b4bb-1e49b2b0f147-httpd-run\") pod \"glance-default-single-1\" (UID: \"4464d6de-546d-4efe-b4bb-1e49b2b0f147\") " pod="glance-kuttl-tests/glance-default-single-1" Dec 01 08:58:53 crc kubenswrapper[4813]: I1201 08:58:53.242455 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/0d38a9cc-3031-42c1-a421-c7446e2ae2a3-httpd-run\") pod \"glance-default-single-0\" (UID: \"0d38a9cc-3031-42c1-a421-c7446e2ae2a3\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 01 08:58:53 crc kubenswrapper[4813]: I1201 08:58:53.242460 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/0d38a9cc-3031-42c1-a421-c7446e2ae2a3-etc-nvme\") pod \"glance-default-single-0\" (UID: \"0d38a9cc-3031-42c1-a421-c7446e2ae2a3\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 01 08:58:53 crc kubenswrapper[4813]: I1201 08:58:53.242481 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-single-1\" (UID: \"4464d6de-546d-4efe-b4bb-1e49b2b0f147\") " pod="glance-kuttl-tests/glance-default-single-1" Dec 01 08:58:53 crc kubenswrapper[4813]: I1201 08:58:53.242570 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/4464d6de-546d-4efe-b4bb-1e49b2b0f147-sys\") pod \"glance-default-single-1\" (UID: \"4464d6de-546d-4efe-b4bb-1e49b2b0f147\") " pod="glance-kuttl-tests/glance-default-single-1" Dec 01 08:58:53 crc kubenswrapper[4813]: I1201 08:58:53.242587 4813 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-single-1\" (UID: \"4464d6de-546d-4efe-b4bb-1e49b2b0f147\") device mount path \"/mnt/openstack/pv01\"" pod="glance-kuttl-tests/glance-default-single-1" Dec 01 08:58:53 crc kubenswrapper[4813]: I1201 08:58:53.242615 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4464d6de-546d-4efe-b4bb-1e49b2b0f147-config-data\") pod \"glance-default-single-1\" (UID: \"4464d6de-546d-4efe-b4bb-1e49b2b0f147\") " pod="glance-kuttl-tests/glance-default-single-1" Dec 01 08:58:53 crc kubenswrapper[4813]: I1201 08:58:53.242651 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/0d38a9cc-3031-42c1-a421-c7446e2ae2a3-sys\") pod \"glance-default-single-0\" (UID: \"0d38a9cc-3031-42c1-a421-c7446e2ae2a3\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 01 08:58:53 crc kubenswrapper[4813]: I1201 08:58:53.242675 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/0d38a9cc-3031-42c1-a421-c7446e2ae2a3-run\") pod \"glance-default-single-0\" (UID: \"0d38a9cc-3031-42c1-a421-c7446e2ae2a3\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 01 08:58:53 crc kubenswrapper[4813]: I1201 08:58:53.242695 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/0d38a9cc-3031-42c1-a421-c7446e2ae2a3-etc-iscsi\") pod \"glance-default-single-0\" (UID: \"0d38a9cc-3031-42c1-a421-c7446e2ae2a3\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 01 08:58:53 crc kubenswrapper[4813]: I1201 08:58:53.242720 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/4464d6de-546d-4efe-b4bb-1e49b2b0f147-run\") pod \"glance-default-single-1\" (UID: \"4464d6de-546d-4efe-b4bb-1e49b2b0f147\") " pod="glance-kuttl-tests/glance-default-single-1" Dec 01 08:58:53 crc kubenswrapper[4813]: I1201 08:58:53.242779 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/0d38a9cc-3031-42c1-a421-c7446e2ae2a3-dev\") pod \"glance-default-single-0\" (UID: \"0d38a9cc-3031-42c1-a421-c7446e2ae2a3\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 01 08:58:53 crc kubenswrapper[4813]: I1201 08:58:53.242811 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/4464d6de-546d-4efe-b4bb-1e49b2b0f147-var-locks-brick\") pod \"glance-default-single-1\" (UID: \"4464d6de-546d-4efe-b4bb-1e49b2b0f147\") " pod="glance-kuttl-tests/glance-default-single-1" Dec 01 08:58:53 crc kubenswrapper[4813]: I1201 08:58:53.242847 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/0d38a9cc-3031-42c1-a421-c7446e2ae2a3-sys\") pod \"glance-default-single-0\" (UID: \"0d38a9cc-3031-42c1-a421-c7446e2ae2a3\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 01 08:58:53 crc kubenswrapper[4813]: I1201 08:58:53.242872 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-single-1\" (UID: \"4464d6de-546d-4efe-b4bb-1e49b2b0f147\") " pod="glance-kuttl-tests/glance-default-single-1" Dec 01 08:58:53 crc kubenswrapper[4813]: I1201 08:58:53.243000 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4464d6de-546d-4efe-b4bb-1e49b2b0f147-logs\") pod \"glance-default-single-1\" (UID: \"4464d6de-546d-4efe-b4bb-1e49b2b0f147\") " pod="glance-kuttl-tests/glance-default-single-1" Dec 01 08:58:53 crc kubenswrapper[4813]: I1201 08:58:53.243031 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/0d38a9cc-3031-42c1-a421-c7446e2ae2a3-run\") pod \"glance-default-single-0\" (UID: \"0d38a9cc-3031-42c1-a421-c7446e2ae2a3\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 01 08:58:53 crc kubenswrapper[4813]: I1201 08:58:53.243054 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/0d38a9cc-3031-42c1-a421-c7446e2ae2a3-var-locks-brick\") pod \"glance-default-single-0\" (UID: \"0d38a9cc-3031-42c1-a421-c7446e2ae2a3\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 01 08:58:53 crc kubenswrapper[4813]: I1201 08:58:53.243062 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/0d38a9cc-3031-42c1-a421-c7446e2ae2a3-etc-iscsi\") pod \"glance-default-single-0\" (UID: \"0d38a9cc-3031-42c1-a421-c7446e2ae2a3\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 01 08:58:53 crc kubenswrapper[4813]: I1201 08:58:53.243088 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/4464d6de-546d-4efe-b4bb-1e49b2b0f147-run\") pod \"glance-default-single-1\" (UID: \"4464d6de-546d-4efe-b4bb-1e49b2b0f147\") " pod="glance-kuttl-tests/glance-default-single-1" Dec 01 08:58:53 crc kubenswrapper[4813]: I1201 08:58:53.243081 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0d38a9cc-3031-42c1-a421-c7446e2ae2a3-scripts\") pod \"glance-default-single-0\" (UID: \"0d38a9cc-3031-42c1-a421-c7446e2ae2a3\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 01 08:58:53 crc kubenswrapper[4813]: I1201 08:58:53.243107 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/0d38a9cc-3031-42c1-a421-c7446e2ae2a3-dev\") pod \"glance-default-single-0\" (UID: \"0d38a9cc-3031-42c1-a421-c7446e2ae2a3\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 01 08:58:53 crc kubenswrapper[4813]: I1201 08:58:53.243009 4813 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-single-1\" (UID: \"4464d6de-546d-4efe-b4bb-1e49b2b0f147\") device mount path \"/mnt/openstack/pv05\"" pod="glance-kuttl-tests/glance-default-single-1" Dec 01 08:58:53 crc kubenswrapper[4813]: I1201 08:58:53.243140 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/4464d6de-546d-4efe-b4bb-1e49b2b0f147-sys\") pod \"glance-default-single-1\" (UID: \"4464d6de-546d-4efe-b4bb-1e49b2b0f147\") " pod="glance-kuttl-tests/glance-default-single-1" Dec 01 08:58:53 crc kubenswrapper[4813]: I1201 08:58:53.243433 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/4464d6de-546d-4efe-b4bb-1e49b2b0f147-var-locks-brick\") pod \"glance-default-single-1\" (UID: \"4464d6de-546d-4efe-b4bb-1e49b2b0f147\") " pod="glance-kuttl-tests/glance-default-single-1" Dec 01 08:58:53 crc kubenswrapper[4813]: I1201 08:58:53.243487 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/0d38a9cc-3031-42c1-a421-c7446e2ae2a3-var-locks-brick\") pod \"glance-default-single-0\" (UID: \"0d38a9cc-3031-42c1-a421-c7446e2ae2a3\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 01 08:58:53 crc kubenswrapper[4813]: I1201 08:58:53.243482 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4464d6de-546d-4efe-b4bb-1e49b2b0f147-scripts\") pod \"glance-default-single-1\" (UID: \"4464d6de-546d-4efe-b4bb-1e49b2b0f147\") " pod="glance-kuttl-tests/glance-default-single-1" Dec 01 08:58:53 crc kubenswrapper[4813]: I1201 08:58:53.243789 4813 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-single-0\" (UID: \"0d38a9cc-3031-42c1-a421-c7446e2ae2a3\") device mount path \"/mnt/openstack/pv10\"" pod="glance-kuttl-tests/glance-default-single-0" Dec 01 08:58:53 crc kubenswrapper[4813]: I1201 08:58:53.243837 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/4464d6de-546d-4efe-b4bb-1e49b2b0f147-etc-nvme\") pod \"glance-default-single-1\" (UID: \"4464d6de-546d-4efe-b4bb-1e49b2b0f147\") " pod="glance-kuttl-tests/glance-default-single-1" Dec 01 08:58:53 crc kubenswrapper[4813]: I1201 08:58:53.244000 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/4464d6de-546d-4efe-b4bb-1e49b2b0f147-lib-modules\") pod \"glance-default-single-1\" (UID: \"4464d6de-546d-4efe-b4bb-1e49b2b0f147\") " pod="glance-kuttl-tests/glance-default-single-1" Dec 01 08:58:53 crc kubenswrapper[4813]: I1201 08:58:53.244569 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4464d6de-546d-4efe-b4bb-1e49b2b0f147-logs\") pod \"glance-default-single-1\" (UID: \"4464d6de-546d-4efe-b4bb-1e49b2b0f147\") " pod="glance-kuttl-tests/glance-default-single-1" Dec 01 08:58:53 crc kubenswrapper[4813]: I1201 08:58:53.244634 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/0d38a9cc-3031-42c1-a421-c7446e2ae2a3-lib-modules\") pod \"glance-default-single-0\" (UID: \"0d38a9cc-3031-42c1-a421-c7446e2ae2a3\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 01 08:58:53 crc kubenswrapper[4813]: I1201 08:58:53.244632 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/4464d6de-546d-4efe-b4bb-1e49b2b0f147-httpd-run\") pod \"glance-default-single-1\" (UID: \"4464d6de-546d-4efe-b4bb-1e49b2b0f147\") " pod="glance-kuttl-tests/glance-default-single-1" Dec 01 08:58:53 crc kubenswrapper[4813]: I1201 08:58:53.244661 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/0d38a9cc-3031-42c1-a421-c7446e2ae2a3-httpd-run\") pod \"glance-default-single-0\" (UID: \"0d38a9cc-3031-42c1-a421-c7446e2ae2a3\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 01 08:58:53 crc kubenswrapper[4813]: I1201 08:58:53.244673 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0d38a9cc-3031-42c1-a421-c7446e2ae2a3-logs\") pod \"glance-default-single-0\" (UID: \"0d38a9cc-3031-42c1-a421-c7446e2ae2a3\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 01 08:58:53 crc kubenswrapper[4813]: I1201 08:58:53.244720 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/0d38a9cc-3031-42c1-a421-c7446e2ae2a3-lib-modules\") pod \"glance-default-single-0\" (UID: \"0d38a9cc-3031-42c1-a421-c7446e2ae2a3\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 01 08:58:53 crc kubenswrapper[4813]: I1201 08:58:53.244787 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7lzsr\" (UniqueName: \"kubernetes.io/projected/0d38a9cc-3031-42c1-a421-c7446e2ae2a3-kube-api-access-7lzsr\") pod \"glance-default-single-0\" (UID: \"0d38a9cc-3031-42c1-a421-c7446e2ae2a3\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 01 08:58:53 crc kubenswrapper[4813]: I1201 08:58:53.244922 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0d38a9cc-3031-42c1-a421-c7446e2ae2a3-logs\") pod \"glance-default-single-0\" (UID: \"0d38a9cc-3031-42c1-a421-c7446e2ae2a3\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 01 08:58:53 crc kubenswrapper[4813]: I1201 08:58:53.249115 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4464d6de-546d-4efe-b4bb-1e49b2b0f147-config-data\") pod \"glance-default-single-1\" (UID: \"4464d6de-546d-4efe-b4bb-1e49b2b0f147\") " pod="glance-kuttl-tests/glance-default-single-1" Dec 01 08:58:53 crc kubenswrapper[4813]: I1201 08:58:53.249931 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0d38a9cc-3031-42c1-a421-c7446e2ae2a3-config-data\") pod \"glance-default-single-0\" (UID: \"0d38a9cc-3031-42c1-a421-c7446e2ae2a3\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 01 08:58:53 crc kubenswrapper[4813]: I1201 08:58:53.254791 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4464d6de-546d-4efe-b4bb-1e49b2b0f147-scripts\") pod \"glance-default-single-1\" (UID: \"4464d6de-546d-4efe-b4bb-1e49b2b0f147\") " pod="glance-kuttl-tests/glance-default-single-1" Dec 01 08:58:53 crc kubenswrapper[4813]: I1201 08:58:53.255732 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0d38a9cc-3031-42c1-a421-c7446e2ae2a3-scripts\") pod \"glance-default-single-0\" (UID: \"0d38a9cc-3031-42c1-a421-c7446e2ae2a3\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 01 08:58:53 crc kubenswrapper[4813]: I1201 08:58:53.257959 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wpmkd\" (UniqueName: \"kubernetes.io/projected/4464d6de-546d-4efe-b4bb-1e49b2b0f147-kube-api-access-wpmkd\") pod \"glance-default-single-1\" (UID: \"4464d6de-546d-4efe-b4bb-1e49b2b0f147\") " pod="glance-kuttl-tests/glance-default-single-1" Dec 01 08:58:53 crc kubenswrapper[4813]: I1201 08:58:53.265780 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7lzsr\" (UniqueName: \"kubernetes.io/projected/0d38a9cc-3031-42c1-a421-c7446e2ae2a3-kube-api-access-7lzsr\") pod \"glance-default-single-0\" (UID: \"0d38a9cc-3031-42c1-a421-c7446e2ae2a3\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 01 08:58:53 crc kubenswrapper[4813]: I1201 08:58:53.267908 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-single-1\" (UID: \"4464d6de-546d-4efe-b4bb-1e49b2b0f147\") " pod="glance-kuttl-tests/glance-default-single-1" Dec 01 08:58:53 crc kubenswrapper[4813]: I1201 08:58:53.269022 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage16-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage16-crc\") pod \"glance-default-single-0\" (UID: \"0d38a9cc-3031-42c1-a421-c7446e2ae2a3\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 01 08:58:53 crc kubenswrapper[4813]: I1201 08:58:53.272531 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-single-0\" (UID: \"0d38a9cc-3031-42c1-a421-c7446e2ae2a3\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 01 08:58:53 crc kubenswrapper[4813]: I1201 08:58:53.286847 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-single-1\" (UID: \"4464d6de-546d-4efe-b4bb-1e49b2b0f147\") " pod="glance-kuttl-tests/glance-default-single-1" Dec 01 08:58:53 crc kubenswrapper[4813]: I1201 08:58:53.336772 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Dec 01 08:58:53 crc kubenswrapper[4813]: I1201 08:58:53.342596 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-1" Dec 01 08:58:53 crc kubenswrapper[4813]: I1201 08:58:53.622929 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-single-1"] Dec 01 08:58:53 crc kubenswrapper[4813]: I1201 08:58:53.883657 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Dec 01 08:58:53 crc kubenswrapper[4813]: W1201 08:58:53.894503 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0d38a9cc_3031_42c1_a421_c7446e2ae2a3.slice/crio-4f33c3a56aec33700d8e3a91dbd58b8eec9c92ec11b9ac2285cf6406d62000fe WatchSource:0}: Error finding container 4f33c3a56aec33700d8e3a91dbd58b8eec9c92ec11b9ac2285cf6406d62000fe: Status 404 returned error can't find the container with id 4f33c3a56aec33700d8e3a91dbd58b8eec9c92ec11b9ac2285cf6406d62000fe Dec 01 08:58:53 crc kubenswrapper[4813]: I1201 08:58:53.942321 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-single-1"] Dec 01 08:58:54 crc kubenswrapper[4813]: I1201 08:58:54.522406 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-1" event={"ID":"4464d6de-546d-4efe-b4bb-1e49b2b0f147","Type":"ContainerStarted","Data":"790b97b55d3327841584a1dd224c5546f228ad1b89f0630f02a6776e248c3cb9"} Dec 01 08:58:54 crc kubenswrapper[4813]: I1201 08:58:54.522883 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-1" event={"ID":"4464d6de-546d-4efe-b4bb-1e49b2b0f147","Type":"ContainerStarted","Data":"237b4ecb3e6eea489ecce967ff55d0f720fb8512180385520ef66d90709cbfe0"} Dec 01 08:58:54 crc kubenswrapper[4813]: I1201 08:58:54.522897 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-1" event={"ID":"4464d6de-546d-4efe-b4bb-1e49b2b0f147","Type":"ContainerStarted","Data":"a5627d024217aaba2ce1a1d28172dd1ea70b3f37dff258e0deec8efcc4f50dde"} Dec 01 08:58:54 crc kubenswrapper[4813]: I1201 08:58:54.522555 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-single-1" podUID="4464d6de-546d-4efe-b4bb-1e49b2b0f147" containerName="glance-log" containerID="cri-o://237b4ecb3e6eea489ecce967ff55d0f720fb8512180385520ef66d90709cbfe0" gracePeriod=30 Dec 01 08:58:54 crc kubenswrapper[4813]: I1201 08:58:54.523184 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-single-1" podUID="4464d6de-546d-4efe-b4bb-1e49b2b0f147" containerName="glance-httpd" containerID="cri-o://790b97b55d3327841584a1dd224c5546f228ad1b89f0630f02a6776e248c3cb9" gracePeriod=30 Dec 01 08:58:54 crc kubenswrapper[4813]: I1201 08:58:54.524945 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"0d38a9cc-3031-42c1-a421-c7446e2ae2a3","Type":"ContainerStarted","Data":"03a85442a2c23d47d6eddbd2d9763103f3d54019d97d99f507e36e857406e531"} Dec 01 08:58:54 crc kubenswrapper[4813]: I1201 08:58:54.525047 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"0d38a9cc-3031-42c1-a421-c7446e2ae2a3","Type":"ContainerStarted","Data":"bdfa25a97aa82668ddf9c2bab8a06b729c0d15ad8b601bba18edeafb8c780149"} Dec 01 08:58:54 crc kubenswrapper[4813]: I1201 08:58:54.525067 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"0d38a9cc-3031-42c1-a421-c7446e2ae2a3","Type":"ContainerStarted","Data":"4f33c3a56aec33700d8e3a91dbd58b8eec9c92ec11b9ac2285cf6406d62000fe"} Dec 01 08:58:54 crc kubenswrapper[4813]: I1201 08:58:54.554535 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-default-single-1" podStartSLOduration=1.554516585 podStartE2EDuration="1.554516585s" podCreationTimestamp="2025-12-01 08:58:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:58:54.551718506 +0000 UTC m=+1119.174640092" watchObservedRunningTime="2025-12-01 08:58:54.554516585 +0000 UTC m=+1119.177438171" Dec 01 08:58:54 crc kubenswrapper[4813]: I1201 08:58:54.590728 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-default-single-0" podStartSLOduration=2.590664372 podStartE2EDuration="2.590664372s" podCreationTimestamp="2025-12-01 08:58:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:58:54.582200344 +0000 UTC m=+1119.205121930" watchObservedRunningTime="2025-12-01 08:58:54.590664372 +0000 UTC m=+1119.213585988" Dec 01 08:58:54 crc kubenswrapper[4813]: I1201 08:58:54.928715 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-1" Dec 01 08:58:54 crc kubenswrapper[4813]: I1201 08:58:54.974890 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/4464d6de-546d-4efe-b4bb-1e49b2b0f147-dev\") pod \"4464d6de-546d-4efe-b4bb-1e49b2b0f147\" (UID: \"4464d6de-546d-4efe-b4bb-1e49b2b0f147\") " Dec 01 08:58:54 crc kubenswrapper[4813]: I1201 08:58:54.974951 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/4464d6de-546d-4efe-b4bb-1e49b2b0f147-sys\") pod \"4464d6de-546d-4efe-b4bb-1e49b2b0f147\" (UID: \"4464d6de-546d-4efe-b4bb-1e49b2b0f147\") " Dec 01 08:58:54 crc kubenswrapper[4813]: I1201 08:58:54.975022 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/4464d6de-546d-4efe-b4bb-1e49b2b0f147-var-locks-brick\") pod \"4464d6de-546d-4efe-b4bb-1e49b2b0f147\" (UID: \"4464d6de-546d-4efe-b4bb-1e49b2b0f147\") " Dec 01 08:58:54 crc kubenswrapper[4813]: I1201 08:58:54.975072 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4464d6de-546d-4efe-b4bb-1e49b2b0f147-dev" (OuterVolumeSpecName: "dev") pod "4464d6de-546d-4efe-b4bb-1e49b2b0f147" (UID: "4464d6de-546d-4efe-b4bb-1e49b2b0f147"). InnerVolumeSpecName "dev". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 08:58:54 crc kubenswrapper[4813]: I1201 08:58:54.975097 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"4464d6de-546d-4efe-b4bb-1e49b2b0f147\" (UID: \"4464d6de-546d-4efe-b4bb-1e49b2b0f147\") " Dec 01 08:58:54 crc kubenswrapper[4813]: I1201 08:58:54.975140 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4464d6de-546d-4efe-b4bb-1e49b2b0f147-scripts\") pod \"4464d6de-546d-4efe-b4bb-1e49b2b0f147\" (UID: \"4464d6de-546d-4efe-b4bb-1e49b2b0f147\") " Dec 01 08:58:54 crc kubenswrapper[4813]: I1201 08:58:54.975149 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4464d6de-546d-4efe-b4bb-1e49b2b0f147-var-locks-brick" (OuterVolumeSpecName: "var-locks-brick") pod "4464d6de-546d-4efe-b4bb-1e49b2b0f147" (UID: "4464d6de-546d-4efe-b4bb-1e49b2b0f147"). InnerVolumeSpecName "var-locks-brick". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 08:58:54 crc kubenswrapper[4813]: I1201 08:58:54.975168 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/4464d6de-546d-4efe-b4bb-1e49b2b0f147-etc-iscsi\") pod \"4464d6de-546d-4efe-b4bb-1e49b2b0f147\" (UID: \"4464d6de-546d-4efe-b4bb-1e49b2b0f147\") " Dec 01 08:58:54 crc kubenswrapper[4813]: I1201 08:58:54.975158 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4464d6de-546d-4efe-b4bb-1e49b2b0f147-sys" (OuterVolumeSpecName: "sys") pod "4464d6de-546d-4efe-b4bb-1e49b2b0f147" (UID: "4464d6de-546d-4efe-b4bb-1e49b2b0f147"). InnerVolumeSpecName "sys". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 08:58:54 crc kubenswrapper[4813]: I1201 08:58:54.975244 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/4464d6de-546d-4efe-b4bb-1e49b2b0f147-lib-modules\") pod \"4464d6de-546d-4efe-b4bb-1e49b2b0f147\" (UID: \"4464d6de-546d-4efe-b4bb-1e49b2b0f147\") " Dec 01 08:58:54 crc kubenswrapper[4813]: I1201 08:58:54.975297 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4464d6de-546d-4efe-b4bb-1e49b2b0f147-etc-iscsi" (OuterVolumeSpecName: "etc-iscsi") pod "4464d6de-546d-4efe-b4bb-1e49b2b0f147" (UID: "4464d6de-546d-4efe-b4bb-1e49b2b0f147"). InnerVolumeSpecName "etc-iscsi". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 08:58:54 crc kubenswrapper[4813]: I1201 08:58:54.975270 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance-cache\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"4464d6de-546d-4efe-b4bb-1e49b2b0f147\" (UID: \"4464d6de-546d-4efe-b4bb-1e49b2b0f147\") " Dec 01 08:58:54 crc kubenswrapper[4813]: I1201 08:58:54.975323 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4464d6de-546d-4efe-b4bb-1e49b2b0f147-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "4464d6de-546d-4efe-b4bb-1e49b2b0f147" (UID: "4464d6de-546d-4efe-b4bb-1e49b2b0f147"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 08:58:54 crc kubenswrapper[4813]: I1201 08:58:54.975362 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4464d6de-546d-4efe-b4bb-1e49b2b0f147-logs\") pod \"4464d6de-546d-4efe-b4bb-1e49b2b0f147\" (UID: \"4464d6de-546d-4efe-b4bb-1e49b2b0f147\") " Dec 01 08:58:54 crc kubenswrapper[4813]: I1201 08:58:54.975396 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/4464d6de-546d-4efe-b4bb-1e49b2b0f147-etc-nvme\") pod \"4464d6de-546d-4efe-b4bb-1e49b2b0f147\" (UID: \"4464d6de-546d-4efe-b4bb-1e49b2b0f147\") " Dec 01 08:58:54 crc kubenswrapper[4813]: I1201 08:58:54.975489 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4464d6de-546d-4efe-b4bb-1e49b2b0f147-config-data\") pod \"4464d6de-546d-4efe-b4bb-1e49b2b0f147\" (UID: \"4464d6de-546d-4efe-b4bb-1e49b2b0f147\") " Dec 01 08:58:54 crc kubenswrapper[4813]: I1201 08:58:54.975546 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wpmkd\" (UniqueName: \"kubernetes.io/projected/4464d6de-546d-4efe-b4bb-1e49b2b0f147-kube-api-access-wpmkd\") pod \"4464d6de-546d-4efe-b4bb-1e49b2b0f147\" (UID: \"4464d6de-546d-4efe-b4bb-1e49b2b0f147\") " Dec 01 08:58:54 crc kubenswrapper[4813]: I1201 08:58:54.975587 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/4464d6de-546d-4efe-b4bb-1e49b2b0f147-httpd-run\") pod \"4464d6de-546d-4efe-b4bb-1e49b2b0f147\" (UID: \"4464d6de-546d-4efe-b4bb-1e49b2b0f147\") " Dec 01 08:58:54 crc kubenswrapper[4813]: I1201 08:58:54.975640 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/4464d6de-546d-4efe-b4bb-1e49b2b0f147-run\") pod \"4464d6de-546d-4efe-b4bb-1e49b2b0f147\" (UID: \"4464d6de-546d-4efe-b4bb-1e49b2b0f147\") " Dec 01 08:58:54 crc kubenswrapper[4813]: I1201 08:58:54.975936 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4464d6de-546d-4efe-b4bb-1e49b2b0f147-etc-nvme" (OuterVolumeSpecName: "etc-nvme") pod "4464d6de-546d-4efe-b4bb-1e49b2b0f147" (UID: "4464d6de-546d-4efe-b4bb-1e49b2b0f147"). InnerVolumeSpecName "etc-nvme". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 08:58:54 crc kubenswrapper[4813]: I1201 08:58:54.976076 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4464d6de-546d-4efe-b4bb-1e49b2b0f147-run" (OuterVolumeSpecName: "run") pod "4464d6de-546d-4efe-b4bb-1e49b2b0f147" (UID: "4464d6de-546d-4efe-b4bb-1e49b2b0f147"). InnerVolumeSpecName "run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 08:58:54 crc kubenswrapper[4813]: I1201 08:58:54.976138 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4464d6de-546d-4efe-b4bb-1e49b2b0f147-logs" (OuterVolumeSpecName: "logs") pod "4464d6de-546d-4efe-b4bb-1e49b2b0f147" (UID: "4464d6de-546d-4efe-b4bb-1e49b2b0f147"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 08:58:54 crc kubenswrapper[4813]: I1201 08:58:54.976325 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4464d6de-546d-4efe-b4bb-1e49b2b0f147-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "4464d6de-546d-4efe-b4bb-1e49b2b0f147" (UID: "4464d6de-546d-4efe-b4bb-1e49b2b0f147"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 08:58:54 crc kubenswrapper[4813]: I1201 08:58:54.978062 4813 reconciler_common.go:293] "Volume detached for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/4464d6de-546d-4efe-b4bb-1e49b2b0f147-dev\") on node \"crc\" DevicePath \"\"" Dec 01 08:58:54 crc kubenswrapper[4813]: I1201 08:58:54.978110 4813 reconciler_common.go:293] "Volume detached for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/4464d6de-546d-4efe-b4bb-1e49b2b0f147-sys\") on node \"crc\" DevicePath \"\"" Dec 01 08:58:54 crc kubenswrapper[4813]: I1201 08:58:54.978137 4813 reconciler_common.go:293] "Volume detached for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/4464d6de-546d-4efe-b4bb-1e49b2b0f147-var-locks-brick\") on node \"crc\" DevicePath \"\"" Dec 01 08:58:54 crc kubenswrapper[4813]: I1201 08:58:54.978163 4813 reconciler_common.go:293] "Volume detached for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/4464d6de-546d-4efe-b4bb-1e49b2b0f147-etc-iscsi\") on node \"crc\" DevicePath \"\"" Dec 01 08:58:54 crc kubenswrapper[4813]: I1201 08:58:54.978185 4813 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/4464d6de-546d-4efe-b4bb-1e49b2b0f147-lib-modules\") on node \"crc\" DevicePath \"\"" Dec 01 08:58:54 crc kubenswrapper[4813]: I1201 08:58:54.978208 4813 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4464d6de-546d-4efe-b4bb-1e49b2b0f147-logs\") on node \"crc\" DevicePath \"\"" Dec 01 08:58:54 crc kubenswrapper[4813]: I1201 08:58:54.978233 4813 reconciler_common.go:293] "Volume detached for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/4464d6de-546d-4efe-b4bb-1e49b2b0f147-etc-nvme\") on node \"crc\" DevicePath \"\"" Dec 01 08:58:54 crc kubenswrapper[4813]: I1201 08:58:54.978255 4813 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/4464d6de-546d-4efe-b4bb-1e49b2b0f147-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 01 08:58:54 crc kubenswrapper[4813]: I1201 08:58:54.978276 4813 reconciler_common.go:293] "Volume detached for volume \"run\" (UniqueName: \"kubernetes.io/host-path/4464d6de-546d-4efe-b4bb-1e49b2b0f147-run\") on node \"crc\" DevicePath \"\"" Dec 01 08:58:54 crc kubenswrapper[4813]: I1201 08:58:54.980047 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage05-crc" (OuterVolumeSpecName: "glance") pod "4464d6de-546d-4efe-b4bb-1e49b2b0f147" (UID: "4464d6de-546d-4efe-b4bb-1e49b2b0f147"). InnerVolumeSpecName "local-storage05-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 01 08:58:54 crc kubenswrapper[4813]: I1201 08:58:54.980162 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4464d6de-546d-4efe-b4bb-1e49b2b0f147-kube-api-access-wpmkd" (OuterVolumeSpecName: "kube-api-access-wpmkd") pod "4464d6de-546d-4efe-b4bb-1e49b2b0f147" (UID: "4464d6de-546d-4efe-b4bb-1e49b2b0f147"). InnerVolumeSpecName "kube-api-access-wpmkd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:58:54 crc kubenswrapper[4813]: I1201 08:58:54.981212 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4464d6de-546d-4efe-b4bb-1e49b2b0f147-scripts" (OuterVolumeSpecName: "scripts") pod "4464d6de-546d-4efe-b4bb-1e49b2b0f147" (UID: "4464d6de-546d-4efe-b4bb-1e49b2b0f147"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:58:54 crc kubenswrapper[4813]: I1201 08:58:54.981556 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage01-crc" (OuterVolumeSpecName: "glance-cache") pod "4464d6de-546d-4efe-b4bb-1e49b2b0f147" (UID: "4464d6de-546d-4efe-b4bb-1e49b2b0f147"). InnerVolumeSpecName "local-storage01-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 01 08:58:55 crc kubenswrapper[4813]: I1201 08:58:55.014589 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4464d6de-546d-4efe-b4bb-1e49b2b0f147-config-data" (OuterVolumeSpecName: "config-data") pod "4464d6de-546d-4efe-b4bb-1e49b2b0f147" (UID: "4464d6de-546d-4efe-b4bb-1e49b2b0f147"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:58:55 crc kubenswrapper[4813]: I1201 08:58:55.080927 4813 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" " Dec 01 08:58:55 crc kubenswrapper[4813]: I1201 08:58:55.081221 4813 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4464d6de-546d-4efe-b4bb-1e49b2b0f147-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 08:58:55 crc kubenswrapper[4813]: I1201 08:58:55.081261 4813 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" " Dec 01 08:58:55 crc kubenswrapper[4813]: I1201 08:58:55.081282 4813 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4464d6de-546d-4efe-b4bb-1e49b2b0f147-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 08:58:55 crc kubenswrapper[4813]: I1201 08:58:55.081304 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wpmkd\" (UniqueName: \"kubernetes.io/projected/4464d6de-546d-4efe-b4bb-1e49b2b0f147-kube-api-access-wpmkd\") on node \"crc\" DevicePath \"\"" Dec 01 08:58:55 crc kubenswrapper[4813]: I1201 08:58:55.095786 4813 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage01-crc" (UniqueName: "kubernetes.io/local-volume/local-storage01-crc") on node "crc" Dec 01 08:58:55 crc kubenswrapper[4813]: I1201 08:58:55.098466 4813 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage05-crc" (UniqueName: "kubernetes.io/local-volume/local-storage05-crc") on node "crc" Dec 01 08:58:55 crc kubenswrapper[4813]: I1201 08:58:55.182637 4813 reconciler_common.go:293] "Volume detached for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" DevicePath \"\"" Dec 01 08:58:55 crc kubenswrapper[4813]: I1201 08:58:55.182668 4813 reconciler_common.go:293] "Volume detached for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" DevicePath \"\"" Dec 01 08:58:55 crc kubenswrapper[4813]: I1201 08:58:55.532737 4813 generic.go:334] "Generic (PLEG): container finished" podID="4464d6de-546d-4efe-b4bb-1e49b2b0f147" containerID="790b97b55d3327841584a1dd224c5546f228ad1b89f0630f02a6776e248c3cb9" exitCode=143 Dec 01 08:58:55 crc kubenswrapper[4813]: I1201 08:58:55.532767 4813 generic.go:334] "Generic (PLEG): container finished" podID="4464d6de-546d-4efe-b4bb-1e49b2b0f147" containerID="237b4ecb3e6eea489ecce967ff55d0f720fb8512180385520ef66d90709cbfe0" exitCode=143 Dec 01 08:58:55 crc kubenswrapper[4813]: I1201 08:58:55.533243 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-1" event={"ID":"4464d6de-546d-4efe-b4bb-1e49b2b0f147","Type":"ContainerDied","Data":"790b97b55d3327841584a1dd224c5546f228ad1b89f0630f02a6776e248c3cb9"} Dec 01 08:58:55 crc kubenswrapper[4813]: I1201 08:58:55.533318 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-1" event={"ID":"4464d6de-546d-4efe-b4bb-1e49b2b0f147","Type":"ContainerDied","Data":"237b4ecb3e6eea489ecce967ff55d0f720fb8512180385520ef66d90709cbfe0"} Dec 01 08:58:55 crc kubenswrapper[4813]: I1201 08:58:55.533336 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-1" event={"ID":"4464d6de-546d-4efe-b4bb-1e49b2b0f147","Type":"ContainerDied","Data":"a5627d024217aaba2ce1a1d28172dd1ea70b3f37dff258e0deec8efcc4f50dde"} Dec 01 08:58:55 crc kubenswrapper[4813]: I1201 08:58:55.533357 4813 scope.go:117] "RemoveContainer" containerID="790b97b55d3327841584a1dd224c5546f228ad1b89f0630f02a6776e248c3cb9" Dec 01 08:58:55 crc kubenswrapper[4813]: I1201 08:58:55.533352 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-1" Dec 01 08:58:55 crc kubenswrapper[4813]: I1201 08:58:55.569236 4813 scope.go:117] "RemoveContainer" containerID="237b4ecb3e6eea489ecce967ff55d0f720fb8512180385520ef66d90709cbfe0" Dec 01 08:58:55 crc kubenswrapper[4813]: I1201 08:58:55.575640 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-single-1"] Dec 01 08:58:55 crc kubenswrapper[4813]: I1201 08:58:55.584305 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-default-single-1"] Dec 01 08:58:55 crc kubenswrapper[4813]: I1201 08:58:55.603327 4813 scope.go:117] "RemoveContainer" containerID="790b97b55d3327841584a1dd224c5546f228ad1b89f0630f02a6776e248c3cb9" Dec 01 08:58:55 crc kubenswrapper[4813]: E1201 08:58:55.604283 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"790b97b55d3327841584a1dd224c5546f228ad1b89f0630f02a6776e248c3cb9\": container with ID starting with 790b97b55d3327841584a1dd224c5546f228ad1b89f0630f02a6776e248c3cb9 not found: ID does not exist" containerID="790b97b55d3327841584a1dd224c5546f228ad1b89f0630f02a6776e248c3cb9" Dec 01 08:58:55 crc kubenswrapper[4813]: I1201 08:58:55.604331 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"790b97b55d3327841584a1dd224c5546f228ad1b89f0630f02a6776e248c3cb9"} err="failed to get container status \"790b97b55d3327841584a1dd224c5546f228ad1b89f0630f02a6776e248c3cb9\": rpc error: code = NotFound desc = could not find container \"790b97b55d3327841584a1dd224c5546f228ad1b89f0630f02a6776e248c3cb9\": container with ID starting with 790b97b55d3327841584a1dd224c5546f228ad1b89f0630f02a6776e248c3cb9 not found: ID does not exist" Dec 01 08:58:55 crc kubenswrapper[4813]: I1201 08:58:55.604364 4813 scope.go:117] "RemoveContainer" containerID="237b4ecb3e6eea489ecce967ff55d0f720fb8512180385520ef66d90709cbfe0" Dec 01 08:58:55 crc kubenswrapper[4813]: E1201 08:58:55.605515 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"237b4ecb3e6eea489ecce967ff55d0f720fb8512180385520ef66d90709cbfe0\": container with ID starting with 237b4ecb3e6eea489ecce967ff55d0f720fb8512180385520ef66d90709cbfe0 not found: ID does not exist" containerID="237b4ecb3e6eea489ecce967ff55d0f720fb8512180385520ef66d90709cbfe0" Dec 01 08:58:55 crc kubenswrapper[4813]: I1201 08:58:55.605590 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"237b4ecb3e6eea489ecce967ff55d0f720fb8512180385520ef66d90709cbfe0"} err="failed to get container status \"237b4ecb3e6eea489ecce967ff55d0f720fb8512180385520ef66d90709cbfe0\": rpc error: code = NotFound desc = could not find container \"237b4ecb3e6eea489ecce967ff55d0f720fb8512180385520ef66d90709cbfe0\": container with ID starting with 237b4ecb3e6eea489ecce967ff55d0f720fb8512180385520ef66d90709cbfe0 not found: ID does not exist" Dec 01 08:58:55 crc kubenswrapper[4813]: I1201 08:58:55.605643 4813 scope.go:117] "RemoveContainer" containerID="790b97b55d3327841584a1dd224c5546f228ad1b89f0630f02a6776e248c3cb9" Dec 01 08:58:55 crc kubenswrapper[4813]: I1201 08:58:55.606262 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"790b97b55d3327841584a1dd224c5546f228ad1b89f0630f02a6776e248c3cb9"} err="failed to get container status \"790b97b55d3327841584a1dd224c5546f228ad1b89f0630f02a6776e248c3cb9\": rpc error: code = NotFound desc = could not find container \"790b97b55d3327841584a1dd224c5546f228ad1b89f0630f02a6776e248c3cb9\": container with ID starting with 790b97b55d3327841584a1dd224c5546f228ad1b89f0630f02a6776e248c3cb9 not found: ID does not exist" Dec 01 08:58:55 crc kubenswrapper[4813]: I1201 08:58:55.606301 4813 scope.go:117] "RemoveContainer" containerID="237b4ecb3e6eea489ecce967ff55d0f720fb8512180385520ef66d90709cbfe0" Dec 01 08:58:55 crc kubenswrapper[4813]: I1201 08:58:55.606584 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"237b4ecb3e6eea489ecce967ff55d0f720fb8512180385520ef66d90709cbfe0"} err="failed to get container status \"237b4ecb3e6eea489ecce967ff55d0f720fb8512180385520ef66d90709cbfe0\": rpc error: code = NotFound desc = could not find container \"237b4ecb3e6eea489ecce967ff55d0f720fb8512180385520ef66d90709cbfe0\": container with ID starting with 237b4ecb3e6eea489ecce967ff55d0f720fb8512180385520ef66d90709cbfe0 not found: ID does not exist" Dec 01 08:58:55 crc kubenswrapper[4813]: I1201 08:58:55.616829 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-single-1"] Dec 01 08:58:55 crc kubenswrapper[4813]: E1201 08:58:55.617253 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4464d6de-546d-4efe-b4bb-1e49b2b0f147" containerName="glance-httpd" Dec 01 08:58:55 crc kubenswrapper[4813]: I1201 08:58:55.617283 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="4464d6de-546d-4efe-b4bb-1e49b2b0f147" containerName="glance-httpd" Dec 01 08:58:55 crc kubenswrapper[4813]: E1201 08:58:55.617309 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4464d6de-546d-4efe-b4bb-1e49b2b0f147" containerName="glance-log" Dec 01 08:58:55 crc kubenswrapper[4813]: I1201 08:58:55.617319 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="4464d6de-546d-4efe-b4bb-1e49b2b0f147" containerName="glance-log" Dec 01 08:58:55 crc kubenswrapper[4813]: I1201 08:58:55.617499 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="4464d6de-546d-4efe-b4bb-1e49b2b0f147" containerName="glance-log" Dec 01 08:58:55 crc kubenswrapper[4813]: I1201 08:58:55.617515 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="4464d6de-546d-4efe-b4bb-1e49b2b0f147" containerName="glance-httpd" Dec 01 08:58:55 crc kubenswrapper[4813]: I1201 08:58:55.618435 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-1" Dec 01 08:58:55 crc kubenswrapper[4813]: I1201 08:58:55.643080 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-single-1"] Dec 01 08:58:55 crc kubenswrapper[4813]: I1201 08:58:55.690866 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/b231270a-aff0-442c-9da0-ece1c0540f29-sys\") pod \"glance-default-single-1\" (UID: \"b231270a-aff0-442c-9da0-ece1c0540f29\") " pod="glance-kuttl-tests/glance-default-single-1" Dec 01 08:58:55 crc kubenswrapper[4813]: I1201 08:58:55.691286 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z88bx\" (UniqueName: \"kubernetes.io/projected/b231270a-aff0-442c-9da0-ece1c0540f29-kube-api-access-z88bx\") pod \"glance-default-single-1\" (UID: \"b231270a-aff0-442c-9da0-ece1c0540f29\") " pod="glance-kuttl-tests/glance-default-single-1" Dec 01 08:58:55 crc kubenswrapper[4813]: I1201 08:58:55.691314 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/b231270a-aff0-442c-9da0-ece1c0540f29-dev\") pod \"glance-default-single-1\" (UID: \"b231270a-aff0-442c-9da0-ece1c0540f29\") " pod="glance-kuttl-tests/glance-default-single-1" Dec 01 08:58:55 crc kubenswrapper[4813]: I1201 08:58:55.691331 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b231270a-aff0-442c-9da0-ece1c0540f29-logs\") pod \"glance-default-single-1\" (UID: \"b231270a-aff0-442c-9da0-ece1c0540f29\") " pod="glance-kuttl-tests/glance-default-single-1" Dec 01 08:58:55 crc kubenswrapper[4813]: I1201 08:58:55.691368 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/b231270a-aff0-442c-9da0-ece1c0540f29-etc-nvme\") pod \"glance-default-single-1\" (UID: \"b231270a-aff0-442c-9da0-ece1c0540f29\") " pod="glance-kuttl-tests/glance-default-single-1" Dec 01 08:58:55 crc kubenswrapper[4813]: I1201 08:58:55.691385 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b231270a-aff0-442c-9da0-ece1c0540f29-scripts\") pod \"glance-default-single-1\" (UID: \"b231270a-aff0-442c-9da0-ece1c0540f29\") " pod="glance-kuttl-tests/glance-default-single-1" Dec 01 08:58:55 crc kubenswrapper[4813]: I1201 08:58:55.691413 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b231270a-aff0-442c-9da0-ece1c0540f29-httpd-run\") pod \"glance-default-single-1\" (UID: \"b231270a-aff0-442c-9da0-ece1c0540f29\") " pod="glance-kuttl-tests/glance-default-single-1" Dec 01 08:58:55 crc kubenswrapper[4813]: I1201 08:58:55.691471 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-single-1\" (UID: \"b231270a-aff0-442c-9da0-ece1c0540f29\") " pod="glance-kuttl-tests/glance-default-single-1" Dec 01 08:58:55 crc kubenswrapper[4813]: I1201 08:58:55.691507 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/b231270a-aff0-442c-9da0-ece1c0540f29-var-locks-brick\") pod \"glance-default-single-1\" (UID: \"b231270a-aff0-442c-9da0-ece1c0540f29\") " pod="glance-kuttl-tests/glance-default-single-1" Dec 01 08:58:55 crc kubenswrapper[4813]: I1201 08:58:55.691545 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-single-1\" (UID: \"b231270a-aff0-442c-9da0-ece1c0540f29\") " pod="glance-kuttl-tests/glance-default-single-1" Dec 01 08:58:55 crc kubenswrapper[4813]: I1201 08:58:55.691564 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/b231270a-aff0-442c-9da0-ece1c0540f29-lib-modules\") pod \"glance-default-single-1\" (UID: \"b231270a-aff0-442c-9da0-ece1c0540f29\") " pod="glance-kuttl-tests/glance-default-single-1" Dec 01 08:58:55 crc kubenswrapper[4813]: I1201 08:58:55.691600 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/b231270a-aff0-442c-9da0-ece1c0540f29-run\") pod \"glance-default-single-1\" (UID: \"b231270a-aff0-442c-9da0-ece1c0540f29\") " pod="glance-kuttl-tests/glance-default-single-1" Dec 01 08:58:55 crc kubenswrapper[4813]: I1201 08:58:55.691620 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b231270a-aff0-442c-9da0-ece1c0540f29-config-data\") pod \"glance-default-single-1\" (UID: \"b231270a-aff0-442c-9da0-ece1c0540f29\") " pod="glance-kuttl-tests/glance-default-single-1" Dec 01 08:58:55 crc kubenswrapper[4813]: I1201 08:58:55.691644 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/b231270a-aff0-442c-9da0-ece1c0540f29-etc-iscsi\") pod \"glance-default-single-1\" (UID: \"b231270a-aff0-442c-9da0-ece1c0540f29\") " pod="glance-kuttl-tests/glance-default-single-1" Dec 01 08:58:55 crc kubenswrapper[4813]: I1201 08:58:55.793265 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/b231270a-aff0-442c-9da0-ece1c0540f29-sys\") pod \"glance-default-single-1\" (UID: \"b231270a-aff0-442c-9da0-ece1c0540f29\") " pod="glance-kuttl-tests/glance-default-single-1" Dec 01 08:58:55 crc kubenswrapper[4813]: I1201 08:58:55.793333 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z88bx\" (UniqueName: \"kubernetes.io/projected/b231270a-aff0-442c-9da0-ece1c0540f29-kube-api-access-z88bx\") pod \"glance-default-single-1\" (UID: \"b231270a-aff0-442c-9da0-ece1c0540f29\") " pod="glance-kuttl-tests/glance-default-single-1" Dec 01 08:58:55 crc kubenswrapper[4813]: I1201 08:58:55.793361 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/b231270a-aff0-442c-9da0-ece1c0540f29-dev\") pod \"glance-default-single-1\" (UID: \"b231270a-aff0-442c-9da0-ece1c0540f29\") " pod="glance-kuttl-tests/glance-default-single-1" Dec 01 08:58:55 crc kubenswrapper[4813]: I1201 08:58:55.793385 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b231270a-aff0-442c-9da0-ece1c0540f29-logs\") pod \"glance-default-single-1\" (UID: \"b231270a-aff0-442c-9da0-ece1c0540f29\") " pod="glance-kuttl-tests/glance-default-single-1" Dec 01 08:58:55 crc kubenswrapper[4813]: I1201 08:58:55.793414 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/b231270a-aff0-442c-9da0-ece1c0540f29-sys\") pod \"glance-default-single-1\" (UID: \"b231270a-aff0-442c-9da0-ece1c0540f29\") " pod="glance-kuttl-tests/glance-default-single-1" Dec 01 08:58:55 crc kubenswrapper[4813]: I1201 08:58:55.793433 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/b231270a-aff0-442c-9da0-ece1c0540f29-etc-nvme\") pod \"glance-default-single-1\" (UID: \"b231270a-aff0-442c-9da0-ece1c0540f29\") " pod="glance-kuttl-tests/glance-default-single-1" Dec 01 08:58:55 crc kubenswrapper[4813]: I1201 08:58:55.793501 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b231270a-aff0-442c-9da0-ece1c0540f29-scripts\") pod \"glance-default-single-1\" (UID: \"b231270a-aff0-442c-9da0-ece1c0540f29\") " pod="glance-kuttl-tests/glance-default-single-1" Dec 01 08:58:55 crc kubenswrapper[4813]: I1201 08:58:55.793508 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/b231270a-aff0-442c-9da0-ece1c0540f29-etc-nvme\") pod \"glance-default-single-1\" (UID: \"b231270a-aff0-442c-9da0-ece1c0540f29\") " pod="glance-kuttl-tests/glance-default-single-1" Dec 01 08:58:55 crc kubenswrapper[4813]: I1201 08:58:55.793532 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b231270a-aff0-442c-9da0-ece1c0540f29-httpd-run\") pod \"glance-default-single-1\" (UID: \"b231270a-aff0-442c-9da0-ece1c0540f29\") " pod="glance-kuttl-tests/glance-default-single-1" Dec 01 08:58:55 crc kubenswrapper[4813]: I1201 08:58:55.793562 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/b231270a-aff0-442c-9da0-ece1c0540f29-dev\") pod \"glance-default-single-1\" (UID: \"b231270a-aff0-442c-9da0-ece1c0540f29\") " pod="glance-kuttl-tests/glance-default-single-1" Dec 01 08:58:55 crc kubenswrapper[4813]: I1201 08:58:55.793602 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-single-1\" (UID: \"b231270a-aff0-442c-9da0-ece1c0540f29\") " pod="glance-kuttl-tests/glance-default-single-1" Dec 01 08:58:55 crc kubenswrapper[4813]: I1201 08:58:55.793640 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/b231270a-aff0-442c-9da0-ece1c0540f29-var-locks-brick\") pod \"glance-default-single-1\" (UID: \"b231270a-aff0-442c-9da0-ece1c0540f29\") " pod="glance-kuttl-tests/glance-default-single-1" Dec 01 08:58:55 crc kubenswrapper[4813]: I1201 08:58:55.793687 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-single-1\" (UID: \"b231270a-aff0-442c-9da0-ece1c0540f29\") " pod="glance-kuttl-tests/glance-default-single-1" Dec 01 08:58:55 crc kubenswrapper[4813]: I1201 08:58:55.793708 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/b231270a-aff0-442c-9da0-ece1c0540f29-lib-modules\") pod \"glance-default-single-1\" (UID: \"b231270a-aff0-442c-9da0-ece1c0540f29\") " pod="glance-kuttl-tests/glance-default-single-1" Dec 01 08:58:55 crc kubenswrapper[4813]: I1201 08:58:55.793752 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/b231270a-aff0-442c-9da0-ece1c0540f29-run\") pod \"glance-default-single-1\" (UID: \"b231270a-aff0-442c-9da0-ece1c0540f29\") " pod="glance-kuttl-tests/glance-default-single-1" Dec 01 08:58:55 crc kubenswrapper[4813]: I1201 08:58:55.793783 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b231270a-aff0-442c-9da0-ece1c0540f29-config-data\") pod \"glance-default-single-1\" (UID: \"b231270a-aff0-442c-9da0-ece1c0540f29\") " pod="glance-kuttl-tests/glance-default-single-1" Dec 01 08:58:55 crc kubenswrapper[4813]: I1201 08:58:55.793812 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/b231270a-aff0-442c-9da0-ece1c0540f29-etc-iscsi\") pod \"glance-default-single-1\" (UID: \"b231270a-aff0-442c-9da0-ece1c0540f29\") " pod="glance-kuttl-tests/glance-default-single-1" Dec 01 08:58:55 crc kubenswrapper[4813]: I1201 08:58:55.793897 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/b231270a-aff0-442c-9da0-ece1c0540f29-etc-iscsi\") pod \"glance-default-single-1\" (UID: \"b231270a-aff0-442c-9da0-ece1c0540f29\") " pod="glance-kuttl-tests/glance-default-single-1" Dec 01 08:58:55 crc kubenswrapper[4813]: I1201 08:58:55.794018 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b231270a-aff0-442c-9da0-ece1c0540f29-httpd-run\") pod \"glance-default-single-1\" (UID: \"b231270a-aff0-442c-9da0-ece1c0540f29\") " pod="glance-kuttl-tests/glance-default-single-1" Dec 01 08:58:55 crc kubenswrapper[4813]: I1201 08:58:55.794160 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/b231270a-aff0-442c-9da0-ece1c0540f29-run\") pod \"glance-default-single-1\" (UID: \"b231270a-aff0-442c-9da0-ece1c0540f29\") " pod="glance-kuttl-tests/glance-default-single-1" Dec 01 08:58:55 crc kubenswrapper[4813]: I1201 08:58:55.794187 4813 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-single-1\" (UID: \"b231270a-aff0-442c-9da0-ece1c0540f29\") device mount path \"/mnt/openstack/pv05\"" pod="glance-kuttl-tests/glance-default-single-1" Dec 01 08:58:55 crc kubenswrapper[4813]: I1201 08:58:55.794216 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/b231270a-aff0-442c-9da0-ece1c0540f29-lib-modules\") pod \"glance-default-single-1\" (UID: \"b231270a-aff0-442c-9da0-ece1c0540f29\") " pod="glance-kuttl-tests/glance-default-single-1" Dec 01 08:58:55 crc kubenswrapper[4813]: I1201 08:58:55.794315 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/b231270a-aff0-442c-9da0-ece1c0540f29-var-locks-brick\") pod \"glance-default-single-1\" (UID: \"b231270a-aff0-442c-9da0-ece1c0540f29\") " pod="glance-kuttl-tests/glance-default-single-1" Dec 01 08:58:55 crc kubenswrapper[4813]: I1201 08:58:55.794320 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b231270a-aff0-442c-9da0-ece1c0540f29-logs\") pod \"glance-default-single-1\" (UID: \"b231270a-aff0-442c-9da0-ece1c0540f29\") " pod="glance-kuttl-tests/glance-default-single-1" Dec 01 08:58:55 crc kubenswrapper[4813]: I1201 08:58:55.794505 4813 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-single-1\" (UID: \"b231270a-aff0-442c-9da0-ece1c0540f29\") device mount path \"/mnt/openstack/pv01\"" pod="glance-kuttl-tests/glance-default-single-1" Dec 01 08:58:55 crc kubenswrapper[4813]: I1201 08:58:55.803676 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b231270a-aff0-442c-9da0-ece1c0540f29-scripts\") pod \"glance-default-single-1\" (UID: \"b231270a-aff0-442c-9da0-ece1c0540f29\") " pod="glance-kuttl-tests/glance-default-single-1" Dec 01 08:58:55 crc kubenswrapper[4813]: I1201 08:58:55.804890 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b231270a-aff0-442c-9da0-ece1c0540f29-config-data\") pod \"glance-default-single-1\" (UID: \"b231270a-aff0-442c-9da0-ece1c0540f29\") " pod="glance-kuttl-tests/glance-default-single-1" Dec 01 08:58:55 crc kubenswrapper[4813]: I1201 08:58:55.817412 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-single-1\" (UID: \"b231270a-aff0-442c-9da0-ece1c0540f29\") " pod="glance-kuttl-tests/glance-default-single-1" Dec 01 08:58:55 crc kubenswrapper[4813]: I1201 08:58:55.820821 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-single-1\" (UID: \"b231270a-aff0-442c-9da0-ece1c0540f29\") " pod="glance-kuttl-tests/glance-default-single-1" Dec 01 08:58:55 crc kubenswrapper[4813]: I1201 08:58:55.822005 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z88bx\" (UniqueName: \"kubernetes.io/projected/b231270a-aff0-442c-9da0-ece1c0540f29-kube-api-access-z88bx\") pod \"glance-default-single-1\" (UID: \"b231270a-aff0-442c-9da0-ece1c0540f29\") " pod="glance-kuttl-tests/glance-default-single-1" Dec 01 08:58:55 crc kubenswrapper[4813]: I1201 08:58:55.935395 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-1" Dec 01 08:58:56 crc kubenswrapper[4813]: I1201 08:58:56.369466 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-single-1"] Dec 01 08:58:56 crc kubenswrapper[4813]: W1201 08:58:56.376305 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb231270a_aff0_442c_9da0_ece1c0540f29.slice/crio-012eb03987f00dc3a6b76d7725b4deb8494670bddbdb9af4b9ad2fb6d610ee96 WatchSource:0}: Error finding container 012eb03987f00dc3a6b76d7725b4deb8494670bddbdb9af4b9ad2fb6d610ee96: Status 404 returned error can't find the container with id 012eb03987f00dc3a6b76d7725b4deb8494670bddbdb9af4b9ad2fb6d610ee96 Dec 01 08:58:56 crc kubenswrapper[4813]: I1201 08:58:56.414291 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4464d6de-546d-4efe-b4bb-1e49b2b0f147" path="/var/lib/kubelet/pods/4464d6de-546d-4efe-b4bb-1e49b2b0f147/volumes" Dec 01 08:58:56 crc kubenswrapper[4813]: I1201 08:58:56.548131 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-1" event={"ID":"b231270a-aff0-442c-9da0-ece1c0540f29","Type":"ContainerStarted","Data":"012eb03987f00dc3a6b76d7725b4deb8494670bddbdb9af4b9ad2fb6d610ee96"} Dec 01 08:58:57 crc kubenswrapper[4813]: I1201 08:58:57.561278 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-1" event={"ID":"b231270a-aff0-442c-9da0-ece1c0540f29","Type":"ContainerStarted","Data":"5d49aad4c62c639f57ff39f38b2afd9ac5d3c22513ab579c17ebcce6f29db9a0"} Dec 01 08:58:57 crc kubenswrapper[4813]: I1201 08:58:57.562757 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-1" event={"ID":"b231270a-aff0-442c-9da0-ece1c0540f29","Type":"ContainerStarted","Data":"8180eaad882643280dbedf20c976cbd80a759bd80dfdfb74e42c9ec5beb7973a"} Dec 01 08:58:57 crc kubenswrapper[4813]: I1201 08:58:57.601677 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-default-single-1" podStartSLOduration=2.601639291 podStartE2EDuration="2.601639291s" podCreationTimestamp="2025-12-01 08:58:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:58:57.599329256 +0000 UTC m=+1122.222250852" watchObservedRunningTime="2025-12-01 08:58:57.601639291 +0000 UTC m=+1122.224560867" Dec 01 08:59:03 crc kubenswrapper[4813]: I1201 08:59:03.337566 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-single-0" Dec 01 08:59:03 crc kubenswrapper[4813]: I1201 08:59:03.338113 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-single-0" Dec 01 08:59:03 crc kubenswrapper[4813]: I1201 08:59:03.366716 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-single-0" Dec 01 08:59:03 crc kubenswrapper[4813]: I1201 08:59:03.386612 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-single-0" Dec 01 08:59:03 crc kubenswrapper[4813]: I1201 08:59:03.613738 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-single-0" Dec 01 08:59:03 crc kubenswrapper[4813]: I1201 08:59:03.613800 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-single-0" Dec 01 08:59:05 crc kubenswrapper[4813]: I1201 08:59:05.936774 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-single-1" Dec 01 08:59:05 crc kubenswrapper[4813]: I1201 08:59:05.937142 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-single-1" Dec 01 08:59:05 crc kubenswrapper[4813]: I1201 08:59:05.997455 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-single-1" Dec 01 08:59:06 crc kubenswrapper[4813]: I1201 08:59:06.018662 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-single-1" Dec 01 08:59:06 crc kubenswrapper[4813]: I1201 08:59:06.639882 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-single-1" Dec 01 08:59:06 crc kubenswrapper[4813]: I1201 08:59:06.639944 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-single-1" Dec 01 08:59:06 crc kubenswrapper[4813]: I1201 08:59:06.986596 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-single-0" Dec 01 08:59:06 crc kubenswrapper[4813]: I1201 08:59:06.986801 4813 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 01 08:59:06 crc kubenswrapper[4813]: I1201 08:59:06.988315 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-single-0" Dec 01 08:59:08 crc kubenswrapper[4813]: I1201 08:59:08.556642 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-single-1" Dec 01 08:59:08 crc kubenswrapper[4813]: I1201 08:59:08.666286 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-single-1" Dec 01 08:59:08 crc kubenswrapper[4813]: I1201 08:59:08.707827 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Dec 01 08:59:08 crc kubenswrapper[4813]: I1201 08:59:08.708116 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-single-0" podUID="0d38a9cc-3031-42c1-a421-c7446e2ae2a3" containerName="glance-log" containerID="cri-o://bdfa25a97aa82668ddf9c2bab8a06b729c0d15ad8b601bba18edeafb8c780149" gracePeriod=30 Dec 01 08:59:08 crc kubenswrapper[4813]: I1201 08:59:08.708649 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-single-0" podUID="0d38a9cc-3031-42c1-a421-c7446e2ae2a3" containerName="glance-httpd" containerID="cri-o://03a85442a2c23d47d6eddbd2d9763103f3d54019d97d99f507e36e857406e531" gracePeriod=30 Dec 01 08:59:09 crc kubenswrapper[4813]: I1201 08:59:09.686534 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"0d38a9cc-3031-42c1-a421-c7446e2ae2a3","Type":"ContainerDied","Data":"bdfa25a97aa82668ddf9c2bab8a06b729c0d15ad8b601bba18edeafb8c780149"} Dec 01 08:59:09 crc kubenswrapper[4813]: I1201 08:59:09.686501 4813 generic.go:334] "Generic (PLEG): container finished" podID="0d38a9cc-3031-42c1-a421-c7446e2ae2a3" containerID="bdfa25a97aa82668ddf9c2bab8a06b729c0d15ad8b601bba18edeafb8c780149" exitCode=143 Dec 01 08:59:12 crc kubenswrapper[4813]: I1201 08:59:12.322665 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Dec 01 08:59:12 crc kubenswrapper[4813]: I1201 08:59:12.421747 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/0d38a9cc-3031-42c1-a421-c7446e2ae2a3-var-locks-brick\") pod \"0d38a9cc-3031-42c1-a421-c7446e2ae2a3\" (UID: \"0d38a9cc-3031-42c1-a421-c7446e2ae2a3\") " Dec 01 08:59:12 crc kubenswrapper[4813]: I1201 08:59:12.421821 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/0d38a9cc-3031-42c1-a421-c7446e2ae2a3-etc-nvme\") pod \"0d38a9cc-3031-42c1-a421-c7446e2ae2a3\" (UID: \"0d38a9cc-3031-42c1-a421-c7446e2ae2a3\") " Dec 01 08:59:12 crc kubenswrapper[4813]: I1201 08:59:12.421870 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage16-crc\") pod \"0d38a9cc-3031-42c1-a421-c7446e2ae2a3\" (UID: \"0d38a9cc-3031-42c1-a421-c7446e2ae2a3\") " Dec 01 08:59:12 crc kubenswrapper[4813]: I1201 08:59:12.421951 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0d38a9cc-3031-42c1-a421-c7446e2ae2a3-scripts\") pod \"0d38a9cc-3031-42c1-a421-c7446e2ae2a3\" (UID: \"0d38a9cc-3031-42c1-a421-c7446e2ae2a3\") " Dec 01 08:59:12 crc kubenswrapper[4813]: I1201 08:59:12.421951 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0d38a9cc-3031-42c1-a421-c7446e2ae2a3-var-locks-brick" (OuterVolumeSpecName: "var-locks-brick") pod "0d38a9cc-3031-42c1-a421-c7446e2ae2a3" (UID: "0d38a9cc-3031-42c1-a421-c7446e2ae2a3"). InnerVolumeSpecName "var-locks-brick". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 08:59:12 crc kubenswrapper[4813]: I1201 08:59:12.422011 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0d38a9cc-3031-42c1-a421-c7446e2ae2a3-etc-nvme" (OuterVolumeSpecName: "etc-nvme") pod "0d38a9cc-3031-42c1-a421-c7446e2ae2a3" (UID: "0d38a9cc-3031-42c1-a421-c7446e2ae2a3"). InnerVolumeSpecName "etc-nvme". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 08:59:12 crc kubenswrapper[4813]: I1201 08:59:12.422057 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0d38a9cc-3031-42c1-a421-c7446e2ae2a3-run" (OuterVolumeSpecName: "run") pod "0d38a9cc-3031-42c1-a421-c7446e2ae2a3" (UID: "0d38a9cc-3031-42c1-a421-c7446e2ae2a3"). InnerVolumeSpecName "run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 08:59:12 crc kubenswrapper[4813]: I1201 08:59:12.422018 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/0d38a9cc-3031-42c1-a421-c7446e2ae2a3-run\") pod \"0d38a9cc-3031-42c1-a421-c7446e2ae2a3\" (UID: \"0d38a9cc-3031-42c1-a421-c7446e2ae2a3\") " Dec 01 08:59:12 crc kubenswrapper[4813]: I1201 08:59:12.422120 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/0d38a9cc-3031-42c1-a421-c7446e2ae2a3-lib-modules\") pod \"0d38a9cc-3031-42c1-a421-c7446e2ae2a3\" (UID: \"0d38a9cc-3031-42c1-a421-c7446e2ae2a3\") " Dec 01 08:59:12 crc kubenswrapper[4813]: I1201 08:59:12.422168 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance-cache\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"0d38a9cc-3031-42c1-a421-c7446e2ae2a3\" (UID: \"0d38a9cc-3031-42c1-a421-c7446e2ae2a3\") " Dec 01 08:59:12 crc kubenswrapper[4813]: I1201 08:59:12.422200 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/0d38a9cc-3031-42c1-a421-c7446e2ae2a3-sys\") pod \"0d38a9cc-3031-42c1-a421-c7446e2ae2a3\" (UID: \"0d38a9cc-3031-42c1-a421-c7446e2ae2a3\") " Dec 01 08:59:12 crc kubenswrapper[4813]: I1201 08:59:12.422236 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/0d38a9cc-3031-42c1-a421-c7446e2ae2a3-httpd-run\") pod \"0d38a9cc-3031-42c1-a421-c7446e2ae2a3\" (UID: \"0d38a9cc-3031-42c1-a421-c7446e2ae2a3\") " Dec 01 08:59:12 crc kubenswrapper[4813]: I1201 08:59:12.422283 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7lzsr\" (UniqueName: \"kubernetes.io/projected/0d38a9cc-3031-42c1-a421-c7446e2ae2a3-kube-api-access-7lzsr\") pod \"0d38a9cc-3031-42c1-a421-c7446e2ae2a3\" (UID: \"0d38a9cc-3031-42c1-a421-c7446e2ae2a3\") " Dec 01 08:59:12 crc kubenswrapper[4813]: I1201 08:59:12.422316 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0d38a9cc-3031-42c1-a421-c7446e2ae2a3-sys" (OuterVolumeSpecName: "sys") pod "0d38a9cc-3031-42c1-a421-c7446e2ae2a3" (UID: "0d38a9cc-3031-42c1-a421-c7446e2ae2a3"). InnerVolumeSpecName "sys". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 08:59:12 crc kubenswrapper[4813]: I1201 08:59:12.422334 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0d38a9cc-3031-42c1-a421-c7446e2ae2a3-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "0d38a9cc-3031-42c1-a421-c7446e2ae2a3" (UID: "0d38a9cc-3031-42c1-a421-c7446e2ae2a3"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 08:59:12 crc kubenswrapper[4813]: I1201 08:59:12.422407 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0d38a9cc-3031-42c1-a421-c7446e2ae2a3-config-data\") pod \"0d38a9cc-3031-42c1-a421-c7446e2ae2a3\" (UID: \"0d38a9cc-3031-42c1-a421-c7446e2ae2a3\") " Dec 01 08:59:12 crc kubenswrapper[4813]: I1201 08:59:12.422452 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/0d38a9cc-3031-42c1-a421-c7446e2ae2a3-etc-iscsi\") pod \"0d38a9cc-3031-42c1-a421-c7446e2ae2a3\" (UID: \"0d38a9cc-3031-42c1-a421-c7446e2ae2a3\") " Dec 01 08:59:12 crc kubenswrapper[4813]: I1201 08:59:12.422485 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/0d38a9cc-3031-42c1-a421-c7446e2ae2a3-dev\") pod \"0d38a9cc-3031-42c1-a421-c7446e2ae2a3\" (UID: \"0d38a9cc-3031-42c1-a421-c7446e2ae2a3\") " Dec 01 08:59:12 crc kubenswrapper[4813]: I1201 08:59:12.422523 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0d38a9cc-3031-42c1-a421-c7446e2ae2a3-logs\") pod \"0d38a9cc-3031-42c1-a421-c7446e2ae2a3\" (UID: \"0d38a9cc-3031-42c1-a421-c7446e2ae2a3\") " Dec 01 08:59:12 crc kubenswrapper[4813]: I1201 08:59:12.422672 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0d38a9cc-3031-42c1-a421-c7446e2ae2a3-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "0d38a9cc-3031-42c1-a421-c7446e2ae2a3" (UID: "0d38a9cc-3031-42c1-a421-c7446e2ae2a3"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 08:59:12 crc kubenswrapper[4813]: I1201 08:59:12.422741 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0d38a9cc-3031-42c1-a421-c7446e2ae2a3-etc-iscsi" (OuterVolumeSpecName: "etc-iscsi") pod "0d38a9cc-3031-42c1-a421-c7446e2ae2a3" (UID: "0d38a9cc-3031-42c1-a421-c7446e2ae2a3"). InnerVolumeSpecName "etc-iscsi". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 08:59:12 crc kubenswrapper[4813]: I1201 08:59:12.422889 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0d38a9cc-3031-42c1-a421-c7446e2ae2a3-dev" (OuterVolumeSpecName: "dev") pod "0d38a9cc-3031-42c1-a421-c7446e2ae2a3" (UID: "0d38a9cc-3031-42c1-a421-c7446e2ae2a3"). InnerVolumeSpecName "dev". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 08:59:12 crc kubenswrapper[4813]: I1201 08:59:12.423139 4813 reconciler_common.go:293] "Volume detached for volume \"run\" (UniqueName: \"kubernetes.io/host-path/0d38a9cc-3031-42c1-a421-c7446e2ae2a3-run\") on node \"crc\" DevicePath \"\"" Dec 01 08:59:12 crc kubenswrapper[4813]: I1201 08:59:12.423165 4813 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/0d38a9cc-3031-42c1-a421-c7446e2ae2a3-lib-modules\") on node \"crc\" DevicePath \"\"" Dec 01 08:59:12 crc kubenswrapper[4813]: I1201 08:59:12.423181 4813 reconciler_common.go:293] "Volume detached for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/0d38a9cc-3031-42c1-a421-c7446e2ae2a3-sys\") on node \"crc\" DevicePath \"\"" Dec 01 08:59:12 crc kubenswrapper[4813]: I1201 08:59:12.423196 4813 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/0d38a9cc-3031-42c1-a421-c7446e2ae2a3-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 01 08:59:12 crc kubenswrapper[4813]: I1201 08:59:12.423211 4813 reconciler_common.go:293] "Volume detached for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/0d38a9cc-3031-42c1-a421-c7446e2ae2a3-etc-iscsi\") on node \"crc\" DevicePath \"\"" Dec 01 08:59:12 crc kubenswrapper[4813]: I1201 08:59:12.423227 4813 reconciler_common.go:293] "Volume detached for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/0d38a9cc-3031-42c1-a421-c7446e2ae2a3-dev\") on node \"crc\" DevicePath \"\"" Dec 01 08:59:12 crc kubenswrapper[4813]: I1201 08:59:12.423242 4813 reconciler_common.go:293] "Volume detached for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/0d38a9cc-3031-42c1-a421-c7446e2ae2a3-var-locks-brick\") on node \"crc\" DevicePath \"\"" Dec 01 08:59:12 crc kubenswrapper[4813]: I1201 08:59:12.423258 4813 reconciler_common.go:293] "Volume detached for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/0d38a9cc-3031-42c1-a421-c7446e2ae2a3-etc-nvme\") on node \"crc\" DevicePath \"\"" Dec 01 08:59:12 crc kubenswrapper[4813]: I1201 08:59:12.423343 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0d38a9cc-3031-42c1-a421-c7446e2ae2a3-logs" (OuterVolumeSpecName: "logs") pod "0d38a9cc-3031-42c1-a421-c7446e2ae2a3" (UID: "0d38a9cc-3031-42c1-a421-c7446e2ae2a3"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 08:59:12 crc kubenswrapper[4813]: I1201 08:59:12.427280 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0d38a9cc-3031-42c1-a421-c7446e2ae2a3-kube-api-access-7lzsr" (OuterVolumeSpecName: "kube-api-access-7lzsr") pod "0d38a9cc-3031-42c1-a421-c7446e2ae2a3" (UID: "0d38a9cc-3031-42c1-a421-c7446e2ae2a3"). InnerVolumeSpecName "kube-api-access-7lzsr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:59:12 crc kubenswrapper[4813]: I1201 08:59:12.427315 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage16-crc" (OuterVolumeSpecName: "glance") pod "0d38a9cc-3031-42c1-a421-c7446e2ae2a3" (UID: "0d38a9cc-3031-42c1-a421-c7446e2ae2a3"). InnerVolumeSpecName "local-storage16-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 01 08:59:12 crc kubenswrapper[4813]: I1201 08:59:12.428388 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0d38a9cc-3031-42c1-a421-c7446e2ae2a3-scripts" (OuterVolumeSpecName: "scripts") pod "0d38a9cc-3031-42c1-a421-c7446e2ae2a3" (UID: "0d38a9cc-3031-42c1-a421-c7446e2ae2a3"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:59:12 crc kubenswrapper[4813]: I1201 08:59:12.429249 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage10-crc" (OuterVolumeSpecName: "glance-cache") pod "0d38a9cc-3031-42c1-a421-c7446e2ae2a3" (UID: "0d38a9cc-3031-42c1-a421-c7446e2ae2a3"). InnerVolumeSpecName "local-storage10-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 01 08:59:12 crc kubenswrapper[4813]: I1201 08:59:12.479675 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0d38a9cc-3031-42c1-a421-c7446e2ae2a3-config-data" (OuterVolumeSpecName: "config-data") pod "0d38a9cc-3031-42c1-a421-c7446e2ae2a3" (UID: "0d38a9cc-3031-42c1-a421-c7446e2ae2a3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:59:12 crc kubenswrapper[4813]: I1201 08:59:12.525198 4813 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage16-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage16-crc\") on node \"crc\" " Dec 01 08:59:12 crc kubenswrapper[4813]: I1201 08:59:12.525447 4813 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0d38a9cc-3031-42c1-a421-c7446e2ae2a3-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 08:59:12 crc kubenswrapper[4813]: I1201 08:59:12.525530 4813 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" " Dec 01 08:59:12 crc kubenswrapper[4813]: I1201 08:59:12.525599 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7lzsr\" (UniqueName: \"kubernetes.io/projected/0d38a9cc-3031-42c1-a421-c7446e2ae2a3-kube-api-access-7lzsr\") on node \"crc\" DevicePath \"\"" Dec 01 08:59:12 crc kubenswrapper[4813]: I1201 08:59:12.525694 4813 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0d38a9cc-3031-42c1-a421-c7446e2ae2a3-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 08:59:12 crc kubenswrapper[4813]: I1201 08:59:12.525760 4813 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0d38a9cc-3031-42c1-a421-c7446e2ae2a3-logs\") on node \"crc\" DevicePath \"\"" Dec 01 08:59:12 crc kubenswrapper[4813]: I1201 08:59:12.539097 4813 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage10-crc" (UniqueName: "kubernetes.io/local-volume/local-storage10-crc") on node "crc" Dec 01 08:59:12 crc kubenswrapper[4813]: I1201 08:59:12.543141 4813 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage16-crc" (UniqueName: "kubernetes.io/local-volume/local-storage16-crc") on node "crc" Dec 01 08:59:12 crc kubenswrapper[4813]: I1201 08:59:12.626765 4813 reconciler_common.go:293] "Volume detached for volume \"local-storage16-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage16-crc\") on node \"crc\" DevicePath \"\"" Dec 01 08:59:12 crc kubenswrapper[4813]: I1201 08:59:12.626814 4813 reconciler_common.go:293] "Volume detached for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" DevicePath \"\"" Dec 01 08:59:12 crc kubenswrapper[4813]: I1201 08:59:12.717136 4813 generic.go:334] "Generic (PLEG): container finished" podID="0d38a9cc-3031-42c1-a421-c7446e2ae2a3" containerID="03a85442a2c23d47d6eddbd2d9763103f3d54019d97d99f507e36e857406e531" exitCode=0 Dec 01 08:59:12 crc kubenswrapper[4813]: I1201 08:59:12.717216 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Dec 01 08:59:12 crc kubenswrapper[4813]: I1201 08:59:12.717262 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"0d38a9cc-3031-42c1-a421-c7446e2ae2a3","Type":"ContainerDied","Data":"03a85442a2c23d47d6eddbd2d9763103f3d54019d97d99f507e36e857406e531"} Dec 01 08:59:12 crc kubenswrapper[4813]: I1201 08:59:12.721169 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"0d38a9cc-3031-42c1-a421-c7446e2ae2a3","Type":"ContainerDied","Data":"4f33c3a56aec33700d8e3a91dbd58b8eec9c92ec11b9ac2285cf6406d62000fe"} Dec 01 08:59:12 crc kubenswrapper[4813]: I1201 08:59:12.721311 4813 scope.go:117] "RemoveContainer" containerID="03a85442a2c23d47d6eddbd2d9763103f3d54019d97d99f507e36e857406e531" Dec 01 08:59:12 crc kubenswrapper[4813]: I1201 08:59:12.771414 4813 scope.go:117] "RemoveContainer" containerID="bdfa25a97aa82668ddf9c2bab8a06b729c0d15ad8b601bba18edeafb8c780149" Dec 01 08:59:12 crc kubenswrapper[4813]: I1201 08:59:12.777428 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Dec 01 08:59:12 crc kubenswrapper[4813]: I1201 08:59:12.793293 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Dec 01 08:59:12 crc kubenswrapper[4813]: I1201 08:59:12.797820 4813 scope.go:117] "RemoveContainer" containerID="03a85442a2c23d47d6eddbd2d9763103f3d54019d97d99f507e36e857406e531" Dec 01 08:59:12 crc kubenswrapper[4813]: E1201 08:59:12.798320 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"03a85442a2c23d47d6eddbd2d9763103f3d54019d97d99f507e36e857406e531\": container with ID starting with 03a85442a2c23d47d6eddbd2d9763103f3d54019d97d99f507e36e857406e531 not found: ID does not exist" containerID="03a85442a2c23d47d6eddbd2d9763103f3d54019d97d99f507e36e857406e531" Dec 01 08:59:12 crc kubenswrapper[4813]: I1201 08:59:12.798383 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"03a85442a2c23d47d6eddbd2d9763103f3d54019d97d99f507e36e857406e531"} err="failed to get container status \"03a85442a2c23d47d6eddbd2d9763103f3d54019d97d99f507e36e857406e531\": rpc error: code = NotFound desc = could not find container \"03a85442a2c23d47d6eddbd2d9763103f3d54019d97d99f507e36e857406e531\": container with ID starting with 03a85442a2c23d47d6eddbd2d9763103f3d54019d97d99f507e36e857406e531 not found: ID does not exist" Dec 01 08:59:12 crc kubenswrapper[4813]: I1201 08:59:12.798435 4813 scope.go:117] "RemoveContainer" containerID="bdfa25a97aa82668ddf9c2bab8a06b729c0d15ad8b601bba18edeafb8c780149" Dec 01 08:59:12 crc kubenswrapper[4813]: E1201 08:59:12.798674 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bdfa25a97aa82668ddf9c2bab8a06b729c0d15ad8b601bba18edeafb8c780149\": container with ID starting with bdfa25a97aa82668ddf9c2bab8a06b729c0d15ad8b601bba18edeafb8c780149 not found: ID does not exist" containerID="bdfa25a97aa82668ddf9c2bab8a06b729c0d15ad8b601bba18edeafb8c780149" Dec 01 08:59:12 crc kubenswrapper[4813]: I1201 08:59:12.798701 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bdfa25a97aa82668ddf9c2bab8a06b729c0d15ad8b601bba18edeafb8c780149"} err="failed to get container status \"bdfa25a97aa82668ddf9c2bab8a06b729c0d15ad8b601bba18edeafb8c780149\": rpc error: code = NotFound desc = could not find container \"bdfa25a97aa82668ddf9c2bab8a06b729c0d15ad8b601bba18edeafb8c780149\": container with ID starting with bdfa25a97aa82668ddf9c2bab8a06b729c0d15ad8b601bba18edeafb8c780149 not found: ID does not exist" Dec 01 08:59:12 crc kubenswrapper[4813]: I1201 08:59:12.801680 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Dec 01 08:59:12 crc kubenswrapper[4813]: E1201 08:59:12.802277 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0d38a9cc-3031-42c1-a421-c7446e2ae2a3" containerName="glance-httpd" Dec 01 08:59:12 crc kubenswrapper[4813]: I1201 08:59:12.802389 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="0d38a9cc-3031-42c1-a421-c7446e2ae2a3" containerName="glance-httpd" Dec 01 08:59:12 crc kubenswrapper[4813]: E1201 08:59:12.802493 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0d38a9cc-3031-42c1-a421-c7446e2ae2a3" containerName="glance-log" Dec 01 08:59:12 crc kubenswrapper[4813]: I1201 08:59:12.802562 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="0d38a9cc-3031-42c1-a421-c7446e2ae2a3" containerName="glance-log" Dec 01 08:59:12 crc kubenswrapper[4813]: I1201 08:59:12.802820 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="0d38a9cc-3031-42c1-a421-c7446e2ae2a3" containerName="glance-httpd" Dec 01 08:59:12 crc kubenswrapper[4813]: I1201 08:59:12.802907 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="0d38a9cc-3031-42c1-a421-c7446e2ae2a3" containerName="glance-log" Dec 01 08:59:12 crc kubenswrapper[4813]: I1201 08:59:12.804004 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Dec 01 08:59:12 crc kubenswrapper[4813]: I1201 08:59:12.815978 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Dec 01 08:59:12 crc kubenswrapper[4813]: I1201 08:59:12.829111 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/92eeaf32-a7b1-4109-82d3-3f4deb6bb8bf-config-data\") pod \"glance-default-single-0\" (UID: \"92eeaf32-a7b1-4109-82d3-3f4deb6bb8bf\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 01 08:59:12 crc kubenswrapper[4813]: I1201 08:59:12.829180 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/92eeaf32-a7b1-4109-82d3-3f4deb6bb8bf-dev\") pod \"glance-default-single-0\" (UID: \"92eeaf32-a7b1-4109-82d3-3f4deb6bb8bf\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 01 08:59:12 crc kubenswrapper[4813]: I1201 08:59:12.829230 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/92eeaf32-a7b1-4109-82d3-3f4deb6bb8bf-etc-nvme\") pod \"glance-default-single-0\" (UID: \"92eeaf32-a7b1-4109-82d3-3f4deb6bb8bf\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 01 08:59:12 crc kubenswrapper[4813]: I1201 08:59:12.829288 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/92eeaf32-a7b1-4109-82d3-3f4deb6bb8bf-scripts\") pod \"glance-default-single-0\" (UID: \"92eeaf32-a7b1-4109-82d3-3f4deb6bb8bf\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 01 08:59:12 crc kubenswrapper[4813]: I1201 08:59:12.829331 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/92eeaf32-a7b1-4109-82d3-3f4deb6bb8bf-logs\") pod \"glance-default-single-0\" (UID: \"92eeaf32-a7b1-4109-82d3-3f4deb6bb8bf\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 01 08:59:12 crc kubenswrapper[4813]: I1201 08:59:12.829360 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/92eeaf32-a7b1-4109-82d3-3f4deb6bb8bf-etc-iscsi\") pod \"glance-default-single-0\" (UID: \"92eeaf32-a7b1-4109-82d3-3f4deb6bb8bf\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 01 08:59:12 crc kubenswrapper[4813]: I1201 08:59:12.829457 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/92eeaf32-a7b1-4109-82d3-3f4deb6bb8bf-httpd-run\") pod \"glance-default-single-0\" (UID: \"92eeaf32-a7b1-4109-82d3-3f4deb6bb8bf\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 01 08:59:12 crc kubenswrapper[4813]: I1201 08:59:12.829512 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5mj99\" (UniqueName: \"kubernetes.io/projected/92eeaf32-a7b1-4109-82d3-3f4deb6bb8bf-kube-api-access-5mj99\") pod \"glance-default-single-0\" (UID: \"92eeaf32-a7b1-4109-82d3-3f4deb6bb8bf\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 01 08:59:12 crc kubenswrapper[4813]: I1201 08:59:12.829619 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/92eeaf32-a7b1-4109-82d3-3f4deb6bb8bf-run\") pod \"glance-default-single-0\" (UID: \"92eeaf32-a7b1-4109-82d3-3f4deb6bb8bf\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 01 08:59:12 crc kubenswrapper[4813]: I1201 08:59:12.829742 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/92eeaf32-a7b1-4109-82d3-3f4deb6bb8bf-lib-modules\") pod \"glance-default-single-0\" (UID: \"92eeaf32-a7b1-4109-82d3-3f4deb6bb8bf\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 01 08:59:12 crc kubenswrapper[4813]: I1201 08:59:12.829808 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage16-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage16-crc\") pod \"glance-default-single-0\" (UID: \"92eeaf32-a7b1-4109-82d3-3f4deb6bb8bf\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 01 08:59:12 crc kubenswrapper[4813]: I1201 08:59:12.829862 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/92eeaf32-a7b1-4109-82d3-3f4deb6bb8bf-var-locks-brick\") pod \"glance-default-single-0\" (UID: \"92eeaf32-a7b1-4109-82d3-3f4deb6bb8bf\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 01 08:59:12 crc kubenswrapper[4813]: I1201 08:59:12.829956 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-single-0\" (UID: \"92eeaf32-a7b1-4109-82d3-3f4deb6bb8bf\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 01 08:59:12 crc kubenswrapper[4813]: I1201 08:59:12.830026 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/92eeaf32-a7b1-4109-82d3-3f4deb6bb8bf-sys\") pod \"glance-default-single-0\" (UID: \"92eeaf32-a7b1-4109-82d3-3f4deb6bb8bf\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 01 08:59:12 crc kubenswrapper[4813]: I1201 08:59:12.930944 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/92eeaf32-a7b1-4109-82d3-3f4deb6bb8bf-logs\") pod \"glance-default-single-0\" (UID: \"92eeaf32-a7b1-4109-82d3-3f4deb6bb8bf\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 01 08:59:12 crc kubenswrapper[4813]: I1201 08:59:12.931050 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/92eeaf32-a7b1-4109-82d3-3f4deb6bb8bf-etc-iscsi\") pod \"glance-default-single-0\" (UID: \"92eeaf32-a7b1-4109-82d3-3f4deb6bb8bf\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 01 08:59:12 crc kubenswrapper[4813]: I1201 08:59:12.931086 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/92eeaf32-a7b1-4109-82d3-3f4deb6bb8bf-httpd-run\") pod \"glance-default-single-0\" (UID: \"92eeaf32-a7b1-4109-82d3-3f4deb6bb8bf\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 01 08:59:12 crc kubenswrapper[4813]: I1201 08:59:12.931112 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5mj99\" (UniqueName: \"kubernetes.io/projected/92eeaf32-a7b1-4109-82d3-3f4deb6bb8bf-kube-api-access-5mj99\") pod \"glance-default-single-0\" (UID: \"92eeaf32-a7b1-4109-82d3-3f4deb6bb8bf\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 01 08:59:12 crc kubenswrapper[4813]: I1201 08:59:12.931155 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/92eeaf32-a7b1-4109-82d3-3f4deb6bb8bf-run\") pod \"glance-default-single-0\" (UID: \"92eeaf32-a7b1-4109-82d3-3f4deb6bb8bf\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 01 08:59:12 crc kubenswrapper[4813]: I1201 08:59:12.931194 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/92eeaf32-a7b1-4109-82d3-3f4deb6bb8bf-lib-modules\") pod \"glance-default-single-0\" (UID: \"92eeaf32-a7b1-4109-82d3-3f4deb6bb8bf\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 01 08:59:12 crc kubenswrapper[4813]: I1201 08:59:12.931228 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage16-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage16-crc\") pod \"glance-default-single-0\" (UID: \"92eeaf32-a7b1-4109-82d3-3f4deb6bb8bf\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 01 08:59:12 crc kubenswrapper[4813]: I1201 08:59:12.931254 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/92eeaf32-a7b1-4109-82d3-3f4deb6bb8bf-var-locks-brick\") pod \"glance-default-single-0\" (UID: \"92eeaf32-a7b1-4109-82d3-3f4deb6bb8bf\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 01 08:59:12 crc kubenswrapper[4813]: I1201 08:59:12.931300 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-single-0\" (UID: \"92eeaf32-a7b1-4109-82d3-3f4deb6bb8bf\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 01 08:59:12 crc kubenswrapper[4813]: I1201 08:59:12.931328 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/92eeaf32-a7b1-4109-82d3-3f4deb6bb8bf-sys\") pod \"glance-default-single-0\" (UID: \"92eeaf32-a7b1-4109-82d3-3f4deb6bb8bf\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 01 08:59:12 crc kubenswrapper[4813]: I1201 08:59:12.931352 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/92eeaf32-a7b1-4109-82d3-3f4deb6bb8bf-config-data\") pod \"glance-default-single-0\" (UID: \"92eeaf32-a7b1-4109-82d3-3f4deb6bb8bf\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 01 08:59:12 crc kubenswrapper[4813]: I1201 08:59:12.931382 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/92eeaf32-a7b1-4109-82d3-3f4deb6bb8bf-dev\") pod \"glance-default-single-0\" (UID: \"92eeaf32-a7b1-4109-82d3-3f4deb6bb8bf\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 01 08:59:12 crc kubenswrapper[4813]: I1201 08:59:12.931417 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/92eeaf32-a7b1-4109-82d3-3f4deb6bb8bf-etc-nvme\") pod \"glance-default-single-0\" (UID: \"92eeaf32-a7b1-4109-82d3-3f4deb6bb8bf\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 01 08:59:12 crc kubenswrapper[4813]: I1201 08:59:12.931441 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/92eeaf32-a7b1-4109-82d3-3f4deb6bb8bf-scripts\") pod \"glance-default-single-0\" (UID: \"92eeaf32-a7b1-4109-82d3-3f4deb6bb8bf\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 01 08:59:12 crc kubenswrapper[4813]: I1201 08:59:12.932417 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/92eeaf32-a7b1-4109-82d3-3f4deb6bb8bf-sys\") pod \"glance-default-single-0\" (UID: \"92eeaf32-a7b1-4109-82d3-3f4deb6bb8bf\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 01 08:59:12 crc kubenswrapper[4813]: I1201 08:59:12.932580 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/92eeaf32-a7b1-4109-82d3-3f4deb6bb8bf-run\") pod \"glance-default-single-0\" (UID: \"92eeaf32-a7b1-4109-82d3-3f4deb6bb8bf\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 01 08:59:12 crc kubenswrapper[4813]: I1201 08:59:12.932618 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/92eeaf32-a7b1-4109-82d3-3f4deb6bb8bf-var-locks-brick\") pod \"glance-default-single-0\" (UID: \"92eeaf32-a7b1-4109-82d3-3f4deb6bb8bf\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 01 08:59:12 crc kubenswrapper[4813]: I1201 08:59:12.932670 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/92eeaf32-a7b1-4109-82d3-3f4deb6bb8bf-etc-nvme\") pod \"glance-default-single-0\" (UID: \"92eeaf32-a7b1-4109-82d3-3f4deb6bb8bf\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 01 08:59:12 crc kubenswrapper[4813]: I1201 08:59:12.932580 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/92eeaf32-a7b1-4109-82d3-3f4deb6bb8bf-etc-iscsi\") pod \"glance-default-single-0\" (UID: \"92eeaf32-a7b1-4109-82d3-3f4deb6bb8bf\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 01 08:59:12 crc kubenswrapper[4813]: I1201 08:59:12.932496 4813 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage16-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage16-crc\") pod \"glance-default-single-0\" (UID: \"92eeaf32-a7b1-4109-82d3-3f4deb6bb8bf\") device mount path \"/mnt/openstack/pv16\"" pod="glance-kuttl-tests/glance-default-single-0" Dec 01 08:59:12 crc kubenswrapper[4813]: I1201 08:59:12.932721 4813 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-single-0\" (UID: \"92eeaf32-a7b1-4109-82d3-3f4deb6bb8bf\") device mount path \"/mnt/openstack/pv10\"" pod="glance-kuttl-tests/glance-default-single-0" Dec 01 08:59:12 crc kubenswrapper[4813]: I1201 08:59:12.932681 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/92eeaf32-a7b1-4109-82d3-3f4deb6bb8bf-lib-modules\") pod \"glance-default-single-0\" (UID: \"92eeaf32-a7b1-4109-82d3-3f4deb6bb8bf\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 01 08:59:12 crc kubenswrapper[4813]: I1201 08:59:12.932883 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/92eeaf32-a7b1-4109-82d3-3f4deb6bb8bf-logs\") pod \"glance-default-single-0\" (UID: \"92eeaf32-a7b1-4109-82d3-3f4deb6bb8bf\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 01 08:59:12 crc kubenswrapper[4813]: I1201 08:59:12.933002 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/92eeaf32-a7b1-4109-82d3-3f4deb6bb8bf-httpd-run\") pod \"glance-default-single-0\" (UID: \"92eeaf32-a7b1-4109-82d3-3f4deb6bb8bf\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 01 08:59:12 crc kubenswrapper[4813]: I1201 08:59:12.933154 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/92eeaf32-a7b1-4109-82d3-3f4deb6bb8bf-dev\") pod \"glance-default-single-0\" (UID: \"92eeaf32-a7b1-4109-82d3-3f4deb6bb8bf\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 01 08:59:12 crc kubenswrapper[4813]: I1201 08:59:12.935454 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/92eeaf32-a7b1-4109-82d3-3f4deb6bb8bf-scripts\") pod \"glance-default-single-0\" (UID: \"92eeaf32-a7b1-4109-82d3-3f4deb6bb8bf\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 01 08:59:12 crc kubenswrapper[4813]: I1201 08:59:12.936901 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/92eeaf32-a7b1-4109-82d3-3f4deb6bb8bf-config-data\") pod \"glance-default-single-0\" (UID: \"92eeaf32-a7b1-4109-82d3-3f4deb6bb8bf\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 01 08:59:12 crc kubenswrapper[4813]: I1201 08:59:12.952280 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage16-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage16-crc\") pod \"glance-default-single-0\" (UID: \"92eeaf32-a7b1-4109-82d3-3f4deb6bb8bf\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 01 08:59:12 crc kubenswrapper[4813]: I1201 08:59:12.954204 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5mj99\" (UniqueName: \"kubernetes.io/projected/92eeaf32-a7b1-4109-82d3-3f4deb6bb8bf-kube-api-access-5mj99\") pod \"glance-default-single-0\" (UID: \"92eeaf32-a7b1-4109-82d3-3f4deb6bb8bf\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 01 08:59:12 crc kubenswrapper[4813]: I1201 08:59:12.960935 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-single-0\" (UID: \"92eeaf32-a7b1-4109-82d3-3f4deb6bb8bf\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 01 08:59:13 crc kubenswrapper[4813]: I1201 08:59:13.129296 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Dec 01 08:59:13 crc kubenswrapper[4813]: I1201 08:59:13.661748 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Dec 01 08:59:13 crc kubenswrapper[4813]: W1201 08:59:13.665956 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod92eeaf32_a7b1_4109_82d3_3f4deb6bb8bf.slice/crio-df218fe74d63795a3c93d09f2acecc950b610411ad8caae4d59f91896597293b WatchSource:0}: Error finding container df218fe74d63795a3c93d09f2acecc950b610411ad8caae4d59f91896597293b: Status 404 returned error can't find the container with id df218fe74d63795a3c93d09f2acecc950b610411ad8caae4d59f91896597293b Dec 01 08:59:13 crc kubenswrapper[4813]: I1201 08:59:13.730367 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"92eeaf32-a7b1-4109-82d3-3f4deb6bb8bf","Type":"ContainerStarted","Data":"df218fe74d63795a3c93d09f2acecc950b610411ad8caae4d59f91896597293b"} Dec 01 08:59:14 crc kubenswrapper[4813]: I1201 08:59:14.442923 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0d38a9cc-3031-42c1-a421-c7446e2ae2a3" path="/var/lib/kubelet/pods/0d38a9cc-3031-42c1-a421-c7446e2ae2a3/volumes" Dec 01 08:59:14 crc kubenswrapper[4813]: I1201 08:59:14.744244 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"92eeaf32-a7b1-4109-82d3-3f4deb6bb8bf","Type":"ContainerStarted","Data":"dd9f2262f5b6023b129ed5ef64f88a24261537a4a8be9e435d043c151fdc6059"} Dec 01 08:59:14 crc kubenswrapper[4813]: I1201 08:59:14.744294 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"92eeaf32-a7b1-4109-82d3-3f4deb6bb8bf","Type":"ContainerStarted","Data":"67d76dcb531e0c7369cc1c8a9943f18d7dacfde17ed6aef5f657851ffab723dc"} Dec 01 08:59:23 crc kubenswrapper[4813]: I1201 08:59:23.130445 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-single-0" Dec 01 08:59:23 crc kubenswrapper[4813]: I1201 08:59:23.131042 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-single-0" Dec 01 08:59:23 crc kubenswrapper[4813]: I1201 08:59:23.190949 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-single-0" Dec 01 08:59:23 crc kubenswrapper[4813]: I1201 08:59:23.208715 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-single-0" Dec 01 08:59:23 crc kubenswrapper[4813]: I1201 08:59:23.227944 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-default-single-0" podStartSLOduration=11.227742984 podStartE2EDuration="11.227742984s" podCreationTimestamp="2025-12-01 08:59:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:59:14.775731444 +0000 UTC m=+1139.398653030" watchObservedRunningTime="2025-12-01 08:59:23.227742984 +0000 UTC m=+1147.850664570" Dec 01 08:59:23 crc kubenswrapper[4813]: I1201 08:59:23.817183 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-single-0" Dec 01 08:59:23 crc kubenswrapper[4813]: I1201 08:59:23.817461 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-single-0" Dec 01 08:59:25 crc kubenswrapper[4813]: I1201 08:59:25.898805 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-single-0" Dec 01 08:59:25 crc kubenswrapper[4813]: I1201 08:59:25.902614 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-single-0" Dec 01 08:59:41 crc kubenswrapper[4813]: I1201 08:59:41.446052 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-db-sync-lpj7j"] Dec 01 08:59:41 crc kubenswrapper[4813]: I1201 08:59:41.458849 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-db-sync-lpj7j"] Dec 01 08:59:41 crc kubenswrapper[4813]: I1201 08:59:41.547035 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Dec 01 08:59:41 crc kubenswrapper[4813]: I1201 08:59:41.547769 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-single-0" podUID="92eeaf32-a7b1-4109-82d3-3f4deb6bb8bf" containerName="glance-log" containerID="cri-o://67d76dcb531e0c7369cc1c8a9943f18d7dacfde17ed6aef5f657851ffab723dc" gracePeriod=30 Dec 01 08:59:41 crc kubenswrapper[4813]: I1201 08:59:41.548329 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-single-0" podUID="92eeaf32-a7b1-4109-82d3-3f4deb6bb8bf" containerName="glance-httpd" containerID="cri-o://dd9f2262f5b6023b129ed5ef64f88a24261537a4a8be9e435d043c151fdc6059" gracePeriod=30 Dec 01 08:59:41 crc kubenswrapper[4813]: I1201 08:59:41.553914 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-single-1"] Dec 01 08:59:41 crc kubenswrapper[4813]: I1201 08:59:41.554264 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-single-1" podUID="b231270a-aff0-442c-9da0-ece1c0540f29" containerName="glance-log" containerID="cri-o://8180eaad882643280dbedf20c976cbd80a759bd80dfdfb74e42c9ec5beb7973a" gracePeriod=30 Dec 01 08:59:41 crc kubenswrapper[4813]: I1201 08:59:41.554318 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-single-1" podUID="b231270a-aff0-442c-9da0-ece1c0540f29" containerName="glance-httpd" containerID="cri-o://5d49aad4c62c639f57ff39f38b2afd9ac5d3c22513ab579c17ebcce6f29db9a0" gracePeriod=30 Dec 01 08:59:41 crc kubenswrapper[4813]: I1201 08:59:41.576835 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/openstackclient"] Dec 01 08:59:41 crc kubenswrapper[4813]: I1201 08:59:41.577091 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/openstackclient" podUID="79ccaf0c-6d3c-4539-b716-72d794c0eb64" containerName="openstackclient" containerID="cri-o://30a5931573955f0ccc85fb3be8db76491c5ee3689e487770b1c50a739162e8c3" gracePeriod=30 Dec 01 08:59:41 crc kubenswrapper[4813]: I1201 08:59:41.582638 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glanced205-account-delete-s9x5m"] Dec 01 08:59:41 crc kubenswrapper[4813]: I1201 08:59:41.630952 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glanced205-account-delete-s9x5m" Dec 01 08:59:41 crc kubenswrapper[4813]: I1201 08:59:41.641179 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glanced205-account-delete-s9x5m"] Dec 01 08:59:41 crc kubenswrapper[4813]: I1201 08:59:41.655691 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mrd7x\" (UniqueName: \"kubernetes.io/projected/ef8750cf-a451-47f9-a168-380246f7fbfd-kube-api-access-mrd7x\") pod \"glanced205-account-delete-s9x5m\" (UID: \"ef8750cf-a451-47f9-a168-380246f7fbfd\") " pod="glance-kuttl-tests/glanced205-account-delete-s9x5m" Dec 01 08:59:41 crc kubenswrapper[4813]: I1201 08:59:41.655757 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ef8750cf-a451-47f9-a168-380246f7fbfd-operator-scripts\") pod \"glanced205-account-delete-s9x5m\" (UID: \"ef8750cf-a451-47f9-a168-380246f7fbfd\") " pod="glance-kuttl-tests/glanced205-account-delete-s9x5m" Dec 01 08:59:41 crc kubenswrapper[4813]: I1201 08:59:41.757337 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mrd7x\" (UniqueName: \"kubernetes.io/projected/ef8750cf-a451-47f9-a168-380246f7fbfd-kube-api-access-mrd7x\") pod \"glanced205-account-delete-s9x5m\" (UID: \"ef8750cf-a451-47f9-a168-380246f7fbfd\") " pod="glance-kuttl-tests/glanced205-account-delete-s9x5m" Dec 01 08:59:41 crc kubenswrapper[4813]: I1201 08:59:41.757409 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ef8750cf-a451-47f9-a168-380246f7fbfd-operator-scripts\") pod \"glanced205-account-delete-s9x5m\" (UID: \"ef8750cf-a451-47f9-a168-380246f7fbfd\") " pod="glance-kuttl-tests/glanced205-account-delete-s9x5m" Dec 01 08:59:41 crc kubenswrapper[4813]: I1201 08:59:41.758424 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ef8750cf-a451-47f9-a168-380246f7fbfd-operator-scripts\") pod \"glanced205-account-delete-s9x5m\" (UID: \"ef8750cf-a451-47f9-a168-380246f7fbfd\") " pod="glance-kuttl-tests/glanced205-account-delete-s9x5m" Dec 01 08:59:41 crc kubenswrapper[4813]: I1201 08:59:41.795498 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mrd7x\" (UniqueName: \"kubernetes.io/projected/ef8750cf-a451-47f9-a168-380246f7fbfd-kube-api-access-mrd7x\") pod \"glanced205-account-delete-s9x5m\" (UID: \"ef8750cf-a451-47f9-a168-380246f7fbfd\") " pod="glance-kuttl-tests/glanced205-account-delete-s9x5m" Dec 01 08:59:41 crc kubenswrapper[4813]: I1201 08:59:41.991153 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/openstackclient" Dec 01 08:59:42 crc kubenswrapper[4813]: I1201 08:59:42.006599 4813 generic.go:334] "Generic (PLEG): container finished" podID="b231270a-aff0-442c-9da0-ece1c0540f29" containerID="8180eaad882643280dbedf20c976cbd80a759bd80dfdfb74e42c9ec5beb7973a" exitCode=143 Dec 01 08:59:42 crc kubenswrapper[4813]: I1201 08:59:42.006663 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-1" event={"ID":"b231270a-aff0-442c-9da0-ece1c0540f29","Type":"ContainerDied","Data":"8180eaad882643280dbedf20c976cbd80a759bd80dfdfb74e42c9ec5beb7973a"} Dec 01 08:59:42 crc kubenswrapper[4813]: I1201 08:59:42.008481 4813 generic.go:334] "Generic (PLEG): container finished" podID="79ccaf0c-6d3c-4539-b716-72d794c0eb64" containerID="30a5931573955f0ccc85fb3be8db76491c5ee3689e487770b1c50a739162e8c3" exitCode=143 Dec 01 08:59:42 crc kubenswrapper[4813]: I1201 08:59:42.008561 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstackclient" event={"ID":"79ccaf0c-6d3c-4539-b716-72d794c0eb64","Type":"ContainerDied","Data":"30a5931573955f0ccc85fb3be8db76491c5ee3689e487770b1c50a739162e8c3"} Dec 01 08:59:42 crc kubenswrapper[4813]: I1201 08:59:42.008576 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/openstackclient" Dec 01 08:59:42 crc kubenswrapper[4813]: I1201 08:59:42.008717 4813 scope.go:117] "RemoveContainer" containerID="30a5931573955f0ccc85fb3be8db76491c5ee3689e487770b1c50a739162e8c3" Dec 01 08:59:42 crc kubenswrapper[4813]: I1201 08:59:42.008837 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstackclient" event={"ID":"79ccaf0c-6d3c-4539-b716-72d794c0eb64","Type":"ContainerDied","Data":"60fcb86c74ec781407237cddbb211e6fdd1d12376ad6ed875cc559ca6b87f1ad"} Dec 01 08:59:42 crc kubenswrapper[4813]: I1201 08:59:42.011498 4813 generic.go:334] "Generic (PLEG): container finished" podID="92eeaf32-a7b1-4109-82d3-3f4deb6bb8bf" containerID="67d76dcb531e0c7369cc1c8a9943f18d7dacfde17ed6aef5f657851ffab723dc" exitCode=143 Dec 01 08:59:42 crc kubenswrapper[4813]: I1201 08:59:42.011549 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"92eeaf32-a7b1-4109-82d3-3f4deb6bb8bf","Type":"ContainerDied","Data":"67d76dcb531e0c7369cc1c8a9943f18d7dacfde17ed6aef5f657851ffab723dc"} Dec 01 08:59:42 crc kubenswrapper[4813]: I1201 08:59:42.033352 4813 scope.go:117] "RemoveContainer" containerID="30a5931573955f0ccc85fb3be8db76491c5ee3689e487770b1c50a739162e8c3" Dec 01 08:59:42 crc kubenswrapper[4813]: I1201 08:59:42.033541 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glanced205-account-delete-s9x5m" Dec 01 08:59:42 crc kubenswrapper[4813]: E1201 08:59:42.034128 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"30a5931573955f0ccc85fb3be8db76491c5ee3689e487770b1c50a739162e8c3\": container with ID starting with 30a5931573955f0ccc85fb3be8db76491c5ee3689e487770b1c50a739162e8c3 not found: ID does not exist" containerID="30a5931573955f0ccc85fb3be8db76491c5ee3689e487770b1c50a739162e8c3" Dec 01 08:59:42 crc kubenswrapper[4813]: I1201 08:59:42.034181 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"30a5931573955f0ccc85fb3be8db76491c5ee3689e487770b1c50a739162e8c3"} err="failed to get container status \"30a5931573955f0ccc85fb3be8db76491c5ee3689e487770b1c50a739162e8c3\": rpc error: code = NotFound desc = could not find container \"30a5931573955f0ccc85fb3be8db76491c5ee3689e487770b1c50a739162e8c3\": container with ID starting with 30a5931573955f0ccc85fb3be8db76491c5ee3689e487770b1c50a739162e8c3 not found: ID does not exist" Dec 01 08:59:42 crc kubenswrapper[4813]: I1201 08:59:42.167899 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qh2qn\" (UniqueName: \"kubernetes.io/projected/79ccaf0c-6d3c-4539-b716-72d794c0eb64-kube-api-access-qh2qn\") pod \"79ccaf0c-6d3c-4539-b716-72d794c0eb64\" (UID: \"79ccaf0c-6d3c-4539-b716-72d794c0eb64\") " Dec 01 08:59:42 crc kubenswrapper[4813]: I1201 08:59:42.168021 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-scripts\" (UniqueName: \"kubernetes.io/configmap/79ccaf0c-6d3c-4539-b716-72d794c0eb64-openstack-scripts\") pod \"79ccaf0c-6d3c-4539-b716-72d794c0eb64\" (UID: \"79ccaf0c-6d3c-4539-b716-72d794c0eb64\") " Dec 01 08:59:42 crc kubenswrapper[4813]: I1201 08:59:42.168149 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/79ccaf0c-6d3c-4539-b716-72d794c0eb64-openstack-config-secret\") pod \"79ccaf0c-6d3c-4539-b716-72d794c0eb64\" (UID: \"79ccaf0c-6d3c-4539-b716-72d794c0eb64\") " Dec 01 08:59:42 crc kubenswrapper[4813]: I1201 08:59:42.168263 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/79ccaf0c-6d3c-4539-b716-72d794c0eb64-openstack-config\") pod \"79ccaf0c-6d3c-4539-b716-72d794c0eb64\" (UID: \"79ccaf0c-6d3c-4539-b716-72d794c0eb64\") " Dec 01 08:59:42 crc kubenswrapper[4813]: I1201 08:59:42.170084 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/79ccaf0c-6d3c-4539-b716-72d794c0eb64-openstack-scripts" (OuterVolumeSpecName: "openstack-scripts") pod "79ccaf0c-6d3c-4539-b716-72d794c0eb64" (UID: "79ccaf0c-6d3c-4539-b716-72d794c0eb64"). InnerVolumeSpecName "openstack-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:59:42 crc kubenswrapper[4813]: I1201 08:59:42.176981 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/79ccaf0c-6d3c-4539-b716-72d794c0eb64-kube-api-access-qh2qn" (OuterVolumeSpecName: "kube-api-access-qh2qn") pod "79ccaf0c-6d3c-4539-b716-72d794c0eb64" (UID: "79ccaf0c-6d3c-4539-b716-72d794c0eb64"). InnerVolumeSpecName "kube-api-access-qh2qn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:59:42 crc kubenswrapper[4813]: I1201 08:59:42.188932 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/79ccaf0c-6d3c-4539-b716-72d794c0eb64-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "79ccaf0c-6d3c-4539-b716-72d794c0eb64" (UID: "79ccaf0c-6d3c-4539-b716-72d794c0eb64"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:59:42 crc kubenswrapper[4813]: I1201 08:59:42.207172 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/79ccaf0c-6d3c-4539-b716-72d794c0eb64-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "79ccaf0c-6d3c-4539-b716-72d794c0eb64" (UID: "79ccaf0c-6d3c-4539-b716-72d794c0eb64"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:59:42 crc kubenswrapper[4813]: I1201 08:59:42.269537 4813 reconciler_common.go:293] "Volume detached for volume \"openstack-scripts\" (UniqueName: \"kubernetes.io/configmap/79ccaf0c-6d3c-4539-b716-72d794c0eb64-openstack-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 08:59:42 crc kubenswrapper[4813]: I1201 08:59:42.269572 4813 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/79ccaf0c-6d3c-4539-b716-72d794c0eb64-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Dec 01 08:59:42 crc kubenswrapper[4813]: I1201 08:59:42.269583 4813 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/79ccaf0c-6d3c-4539-b716-72d794c0eb64-openstack-config\") on node \"crc\" DevicePath \"\"" Dec 01 08:59:42 crc kubenswrapper[4813]: I1201 08:59:42.269591 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qh2qn\" (UniqueName: \"kubernetes.io/projected/79ccaf0c-6d3c-4539-b716-72d794c0eb64-kube-api-access-qh2qn\") on node \"crc\" DevicePath \"\"" Dec 01 08:59:42 crc kubenswrapper[4813]: I1201 08:59:42.340734 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/openstackclient"] Dec 01 08:59:42 crc kubenswrapper[4813]: I1201 08:59:42.348082 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/openstackclient"] Dec 01 08:59:42 crc kubenswrapper[4813]: I1201 08:59:42.406264 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="51769471-fc41-44cc-8b88-cb2e913188b9" path="/var/lib/kubelet/pods/51769471-fc41-44cc-8b88-cb2e913188b9/volumes" Dec 01 08:59:42 crc kubenswrapper[4813]: I1201 08:59:42.407522 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="79ccaf0c-6d3c-4539-b716-72d794c0eb64" path="/var/lib/kubelet/pods/79ccaf0c-6d3c-4539-b716-72d794c0eb64/volumes" Dec 01 08:59:42 crc kubenswrapper[4813]: I1201 08:59:42.457845 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glanced205-account-delete-s9x5m"] Dec 01 08:59:43 crc kubenswrapper[4813]: I1201 08:59:43.026875 4813 generic.go:334] "Generic (PLEG): container finished" podID="ef8750cf-a451-47f9-a168-380246f7fbfd" containerID="9c0ec35f5d5d3797790bee4bbb0abe2d40e46be0e232315ab4892b4a6bbf2d84" exitCode=0 Dec 01 08:59:43 crc kubenswrapper[4813]: I1201 08:59:43.027024 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glanced205-account-delete-s9x5m" event={"ID":"ef8750cf-a451-47f9-a168-380246f7fbfd","Type":"ContainerDied","Data":"9c0ec35f5d5d3797790bee4bbb0abe2d40e46be0e232315ab4892b4a6bbf2d84"} Dec 01 08:59:43 crc kubenswrapper[4813]: I1201 08:59:43.027410 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glanced205-account-delete-s9x5m" event={"ID":"ef8750cf-a451-47f9-a168-380246f7fbfd","Type":"ContainerStarted","Data":"3607c69e339bf5b392f28deb538245a9c2da94f0379142d6fc57ca7935ca4e97"} Dec 01 08:59:44 crc kubenswrapper[4813]: I1201 08:59:44.436996 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glanced205-account-delete-s9x5m" Dec 01 08:59:44 crc kubenswrapper[4813]: I1201 08:59:44.519654 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ef8750cf-a451-47f9-a168-380246f7fbfd-operator-scripts\") pod \"ef8750cf-a451-47f9-a168-380246f7fbfd\" (UID: \"ef8750cf-a451-47f9-a168-380246f7fbfd\") " Dec 01 08:59:44 crc kubenswrapper[4813]: I1201 08:59:44.520115 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mrd7x\" (UniqueName: \"kubernetes.io/projected/ef8750cf-a451-47f9-a168-380246f7fbfd-kube-api-access-mrd7x\") pod \"ef8750cf-a451-47f9-a168-380246f7fbfd\" (UID: \"ef8750cf-a451-47f9-a168-380246f7fbfd\") " Dec 01 08:59:44 crc kubenswrapper[4813]: I1201 08:59:44.520492 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ef8750cf-a451-47f9-a168-380246f7fbfd-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "ef8750cf-a451-47f9-a168-380246f7fbfd" (UID: "ef8750cf-a451-47f9-a168-380246f7fbfd"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:59:44 crc kubenswrapper[4813]: I1201 08:59:44.525774 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ef8750cf-a451-47f9-a168-380246f7fbfd-kube-api-access-mrd7x" (OuterVolumeSpecName: "kube-api-access-mrd7x") pod "ef8750cf-a451-47f9-a168-380246f7fbfd" (UID: "ef8750cf-a451-47f9-a168-380246f7fbfd"). InnerVolumeSpecName "kube-api-access-mrd7x". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:59:44 crc kubenswrapper[4813]: I1201 08:59:44.629891 4813 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ef8750cf-a451-47f9-a168-380246f7fbfd-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 08:59:44 crc kubenswrapper[4813]: I1201 08:59:44.629956 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mrd7x\" (UniqueName: \"kubernetes.io/projected/ef8750cf-a451-47f9-a168-380246f7fbfd-kube-api-access-mrd7x\") on node \"crc\" DevicePath \"\"" Dec 01 08:59:44 crc kubenswrapper[4813]: I1201 08:59:44.988042 4813 prober.go:107] "Probe failed" probeType="Readiness" pod="glance-kuttl-tests/glance-default-single-0" podUID="92eeaf32-a7b1-4109-82d3-3f4deb6bb8bf" containerName="glance-httpd" probeResult="failure" output="Get \"http://10.217.0.101:9292/healthcheck\": read tcp 10.217.0.2:55694->10.217.0.101:9292: read: connection reset by peer" Dec 01 08:59:44 crc kubenswrapper[4813]: I1201 08:59:44.988086 4813 prober.go:107] "Probe failed" probeType="Readiness" pod="glance-kuttl-tests/glance-default-single-0" podUID="92eeaf32-a7b1-4109-82d3-3f4deb6bb8bf" containerName="glance-log" probeResult="failure" output="Get \"http://10.217.0.101:9292/healthcheck\": read tcp 10.217.0.2:55698->10.217.0.101:9292: read: connection reset by peer" Dec 01 08:59:45 crc kubenswrapper[4813]: I1201 08:59:45.045164 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glanced205-account-delete-s9x5m" event={"ID":"ef8750cf-a451-47f9-a168-380246f7fbfd","Type":"ContainerDied","Data":"3607c69e339bf5b392f28deb538245a9c2da94f0379142d6fc57ca7935ca4e97"} Dec 01 08:59:45 crc kubenswrapper[4813]: I1201 08:59:45.045203 4813 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3607c69e339bf5b392f28deb538245a9c2da94f0379142d6fc57ca7935ca4e97" Dec 01 08:59:45 crc kubenswrapper[4813]: I1201 08:59:45.045228 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glanced205-account-delete-s9x5m" Dec 01 08:59:45 crc kubenswrapper[4813]: I1201 08:59:45.413668 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Dec 01 08:59:45 crc kubenswrapper[4813]: I1201 08:59:45.479434 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-1" Dec 01 08:59:45 crc kubenswrapper[4813]: I1201 08:59:45.550302 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/92eeaf32-a7b1-4109-82d3-3f4deb6bb8bf-dev\") pod \"92eeaf32-a7b1-4109-82d3-3f4deb6bb8bf\" (UID: \"92eeaf32-a7b1-4109-82d3-3f4deb6bb8bf\") " Dec 01 08:59:45 crc kubenswrapper[4813]: I1201 08:59:45.550350 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/92eeaf32-a7b1-4109-82d3-3f4deb6bb8bf-sys\") pod \"92eeaf32-a7b1-4109-82d3-3f4deb6bb8bf\" (UID: \"92eeaf32-a7b1-4109-82d3-3f4deb6bb8bf\") " Dec 01 08:59:45 crc kubenswrapper[4813]: I1201 08:59:45.550400 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/92eeaf32-a7b1-4109-82d3-3f4deb6bb8bf-config-data\") pod \"92eeaf32-a7b1-4109-82d3-3f4deb6bb8bf\" (UID: \"92eeaf32-a7b1-4109-82d3-3f4deb6bb8bf\") " Dec 01 08:59:45 crc kubenswrapper[4813]: I1201 08:59:45.550423 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage16-crc\") pod \"92eeaf32-a7b1-4109-82d3-3f4deb6bb8bf\" (UID: \"92eeaf32-a7b1-4109-82d3-3f4deb6bb8bf\") " Dec 01 08:59:45 crc kubenswrapper[4813]: I1201 08:59:45.550408 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/92eeaf32-a7b1-4109-82d3-3f4deb6bb8bf-dev" (OuterVolumeSpecName: "dev") pod "92eeaf32-a7b1-4109-82d3-3f4deb6bb8bf" (UID: "92eeaf32-a7b1-4109-82d3-3f4deb6bb8bf"). InnerVolumeSpecName "dev". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 08:59:45 crc kubenswrapper[4813]: I1201 08:59:45.550444 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/92eeaf32-a7b1-4109-82d3-3f4deb6bb8bf-etc-iscsi\") pod \"92eeaf32-a7b1-4109-82d3-3f4deb6bb8bf\" (UID: \"92eeaf32-a7b1-4109-82d3-3f4deb6bb8bf\") " Dec 01 08:59:45 crc kubenswrapper[4813]: I1201 08:59:45.550463 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/92eeaf32-a7b1-4109-82d3-3f4deb6bb8bf-run\") pod \"92eeaf32-a7b1-4109-82d3-3f4deb6bb8bf\" (UID: \"92eeaf32-a7b1-4109-82d3-3f4deb6bb8bf\") " Dec 01 08:59:45 crc kubenswrapper[4813]: I1201 08:59:45.550478 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/92eeaf32-a7b1-4109-82d3-3f4deb6bb8bf-etc-nvme\") pod \"92eeaf32-a7b1-4109-82d3-3f4deb6bb8bf\" (UID: \"92eeaf32-a7b1-4109-82d3-3f4deb6bb8bf\") " Dec 01 08:59:45 crc kubenswrapper[4813]: I1201 08:59:45.550530 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/92eeaf32-a7b1-4109-82d3-3f4deb6bb8bf-scripts\") pod \"92eeaf32-a7b1-4109-82d3-3f4deb6bb8bf\" (UID: \"92eeaf32-a7b1-4109-82d3-3f4deb6bb8bf\") " Dec 01 08:59:45 crc kubenswrapper[4813]: I1201 08:59:45.550552 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance-cache\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"92eeaf32-a7b1-4109-82d3-3f4deb6bb8bf\" (UID: \"92eeaf32-a7b1-4109-82d3-3f4deb6bb8bf\") " Dec 01 08:59:45 crc kubenswrapper[4813]: I1201 08:59:45.550613 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/92eeaf32-a7b1-4109-82d3-3f4deb6bb8bf-httpd-run\") pod \"92eeaf32-a7b1-4109-82d3-3f4deb6bb8bf\" (UID: \"92eeaf32-a7b1-4109-82d3-3f4deb6bb8bf\") " Dec 01 08:59:45 crc kubenswrapper[4813]: I1201 08:59:45.550875 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5mj99\" (UniqueName: \"kubernetes.io/projected/92eeaf32-a7b1-4109-82d3-3f4deb6bb8bf-kube-api-access-5mj99\") pod \"92eeaf32-a7b1-4109-82d3-3f4deb6bb8bf\" (UID: \"92eeaf32-a7b1-4109-82d3-3f4deb6bb8bf\") " Dec 01 08:59:45 crc kubenswrapper[4813]: I1201 08:59:45.550906 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/b231270a-aff0-442c-9da0-ece1c0540f29-sys\") pod \"b231270a-aff0-442c-9da0-ece1c0540f29\" (UID: \"b231270a-aff0-442c-9da0-ece1c0540f29\") " Dec 01 08:59:45 crc kubenswrapper[4813]: I1201 08:59:45.550935 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/92eeaf32-a7b1-4109-82d3-3f4deb6bb8bf-lib-modules\") pod \"92eeaf32-a7b1-4109-82d3-3f4deb6bb8bf\" (UID: \"92eeaf32-a7b1-4109-82d3-3f4deb6bb8bf\") " Dec 01 08:59:45 crc kubenswrapper[4813]: I1201 08:59:45.550949 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/92eeaf32-a7b1-4109-82d3-3f4deb6bb8bf-var-locks-brick\") pod \"92eeaf32-a7b1-4109-82d3-3f4deb6bb8bf\" (UID: \"92eeaf32-a7b1-4109-82d3-3f4deb6bb8bf\") " Dec 01 08:59:45 crc kubenswrapper[4813]: I1201 08:59:45.550994 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b231270a-aff0-442c-9da0-ece1c0540f29-config-data\") pod \"b231270a-aff0-442c-9da0-ece1c0540f29\" (UID: \"b231270a-aff0-442c-9da0-ece1c0540f29\") " Dec 01 08:59:45 crc kubenswrapper[4813]: I1201 08:59:45.551018 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/92eeaf32-a7b1-4109-82d3-3f4deb6bb8bf-logs\") pod \"92eeaf32-a7b1-4109-82d3-3f4deb6bb8bf\" (UID: \"92eeaf32-a7b1-4109-82d3-3f4deb6bb8bf\") " Dec 01 08:59:45 crc kubenswrapper[4813]: I1201 08:59:45.551032 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/b231270a-aff0-442c-9da0-ece1c0540f29-etc-iscsi\") pod \"b231270a-aff0-442c-9da0-ece1c0540f29\" (UID: \"b231270a-aff0-442c-9da0-ece1c0540f29\") " Dec 01 08:59:45 crc kubenswrapper[4813]: I1201 08:59:45.550514 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/92eeaf32-a7b1-4109-82d3-3f4deb6bb8bf-sys" (OuterVolumeSpecName: "sys") pod "92eeaf32-a7b1-4109-82d3-3f4deb6bb8bf" (UID: "92eeaf32-a7b1-4109-82d3-3f4deb6bb8bf"). InnerVolumeSpecName "sys". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 08:59:45 crc kubenswrapper[4813]: I1201 08:59:45.551146 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b231270a-aff0-442c-9da0-ece1c0540f29-sys" (OuterVolumeSpecName: "sys") pod "b231270a-aff0-442c-9da0-ece1c0540f29" (UID: "b231270a-aff0-442c-9da0-ece1c0540f29"). InnerVolumeSpecName "sys". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 08:59:45 crc kubenswrapper[4813]: I1201 08:59:45.551431 4813 reconciler_common.go:293] "Volume detached for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/92eeaf32-a7b1-4109-82d3-3f4deb6bb8bf-dev\") on node \"crc\" DevicePath \"\"" Dec 01 08:59:45 crc kubenswrapper[4813]: I1201 08:59:45.551458 4813 reconciler_common.go:293] "Volume detached for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/92eeaf32-a7b1-4109-82d3-3f4deb6bb8bf-sys\") on node \"crc\" DevicePath \"\"" Dec 01 08:59:45 crc kubenswrapper[4813]: I1201 08:59:45.551466 4813 reconciler_common.go:293] "Volume detached for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/b231270a-aff0-442c-9da0-ece1c0540f29-sys\") on node \"crc\" DevicePath \"\"" Dec 01 08:59:45 crc kubenswrapper[4813]: I1201 08:59:45.552366 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/92eeaf32-a7b1-4109-82d3-3f4deb6bb8bf-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "92eeaf32-a7b1-4109-82d3-3f4deb6bb8bf" (UID: "92eeaf32-a7b1-4109-82d3-3f4deb6bb8bf"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 08:59:45 crc kubenswrapper[4813]: I1201 08:59:45.552409 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/92eeaf32-a7b1-4109-82d3-3f4deb6bb8bf-var-locks-brick" (OuterVolumeSpecName: "var-locks-brick") pod "92eeaf32-a7b1-4109-82d3-3f4deb6bb8bf" (UID: "92eeaf32-a7b1-4109-82d3-3f4deb6bb8bf"). InnerVolumeSpecName "var-locks-brick". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 08:59:45 crc kubenswrapper[4813]: I1201 08:59:45.553279 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b231270a-aff0-442c-9da0-ece1c0540f29-etc-iscsi" (OuterVolumeSpecName: "etc-iscsi") pod "b231270a-aff0-442c-9da0-ece1c0540f29" (UID: "b231270a-aff0-442c-9da0-ece1c0540f29"). InnerVolumeSpecName "etc-iscsi". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 08:59:45 crc kubenswrapper[4813]: I1201 08:59:45.553350 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/92eeaf32-a7b1-4109-82d3-3f4deb6bb8bf-run" (OuterVolumeSpecName: "run") pod "92eeaf32-a7b1-4109-82d3-3f4deb6bb8bf" (UID: "92eeaf32-a7b1-4109-82d3-3f4deb6bb8bf"). InnerVolumeSpecName "run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 08:59:45 crc kubenswrapper[4813]: I1201 08:59:45.553376 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/92eeaf32-a7b1-4109-82d3-3f4deb6bb8bf-etc-iscsi" (OuterVolumeSpecName: "etc-iscsi") pod "92eeaf32-a7b1-4109-82d3-3f4deb6bb8bf" (UID: "92eeaf32-a7b1-4109-82d3-3f4deb6bb8bf"). InnerVolumeSpecName "etc-iscsi". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 08:59:45 crc kubenswrapper[4813]: I1201 08:59:45.553503 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/92eeaf32-a7b1-4109-82d3-3f4deb6bb8bf-etc-nvme" (OuterVolumeSpecName: "etc-nvme") pod "92eeaf32-a7b1-4109-82d3-3f4deb6bb8bf" (UID: "92eeaf32-a7b1-4109-82d3-3f4deb6bb8bf"). InnerVolumeSpecName "etc-nvme". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 08:59:45 crc kubenswrapper[4813]: I1201 08:59:45.553430 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/92eeaf32-a7b1-4109-82d3-3f4deb6bb8bf-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "92eeaf32-a7b1-4109-82d3-3f4deb6bb8bf" (UID: "92eeaf32-a7b1-4109-82d3-3f4deb6bb8bf"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 08:59:45 crc kubenswrapper[4813]: I1201 08:59:45.553849 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/92eeaf32-a7b1-4109-82d3-3f4deb6bb8bf-logs" (OuterVolumeSpecName: "logs") pod "92eeaf32-a7b1-4109-82d3-3f4deb6bb8bf" (UID: "92eeaf32-a7b1-4109-82d3-3f4deb6bb8bf"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 08:59:45 crc kubenswrapper[4813]: I1201 08:59:45.555816 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/92eeaf32-a7b1-4109-82d3-3f4deb6bb8bf-scripts" (OuterVolumeSpecName: "scripts") pod "92eeaf32-a7b1-4109-82d3-3f4deb6bb8bf" (UID: "92eeaf32-a7b1-4109-82d3-3f4deb6bb8bf"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:59:45 crc kubenswrapper[4813]: I1201 08:59:45.559138 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage10-crc" (OuterVolumeSpecName: "glance-cache") pod "92eeaf32-a7b1-4109-82d3-3f4deb6bb8bf" (UID: "92eeaf32-a7b1-4109-82d3-3f4deb6bb8bf"). InnerVolumeSpecName "local-storage10-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 01 08:59:45 crc kubenswrapper[4813]: I1201 08:59:45.559146 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/92eeaf32-a7b1-4109-82d3-3f4deb6bb8bf-kube-api-access-5mj99" (OuterVolumeSpecName: "kube-api-access-5mj99") pod "92eeaf32-a7b1-4109-82d3-3f4deb6bb8bf" (UID: "92eeaf32-a7b1-4109-82d3-3f4deb6bb8bf"). InnerVolumeSpecName "kube-api-access-5mj99". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:59:45 crc kubenswrapper[4813]: I1201 08:59:45.563327 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage16-crc" (OuterVolumeSpecName: "glance") pod "92eeaf32-a7b1-4109-82d3-3f4deb6bb8bf" (UID: "92eeaf32-a7b1-4109-82d3-3f4deb6bb8bf"). InnerVolumeSpecName "local-storage16-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 01 08:59:45 crc kubenswrapper[4813]: I1201 08:59:45.598477 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b231270a-aff0-442c-9da0-ece1c0540f29-config-data" (OuterVolumeSpecName: "config-data") pod "b231270a-aff0-442c-9da0-ece1c0540f29" (UID: "b231270a-aff0-442c-9da0-ece1c0540f29"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:59:45 crc kubenswrapper[4813]: I1201 08:59:45.608283 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/92eeaf32-a7b1-4109-82d3-3f4deb6bb8bf-config-data" (OuterVolumeSpecName: "config-data") pod "92eeaf32-a7b1-4109-82d3-3f4deb6bb8bf" (UID: "92eeaf32-a7b1-4109-82d3-3f4deb6bb8bf"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:59:45 crc kubenswrapper[4813]: I1201 08:59:45.652131 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b231270a-aff0-442c-9da0-ece1c0540f29-httpd-run\") pod \"b231270a-aff0-442c-9da0-ece1c0540f29\" (UID: \"b231270a-aff0-442c-9da0-ece1c0540f29\") " Dec 01 08:59:45 crc kubenswrapper[4813]: I1201 08:59:45.652185 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z88bx\" (UniqueName: \"kubernetes.io/projected/b231270a-aff0-442c-9da0-ece1c0540f29-kube-api-access-z88bx\") pod \"b231270a-aff0-442c-9da0-ece1c0540f29\" (UID: \"b231270a-aff0-442c-9da0-ece1c0540f29\") " Dec 01 08:59:45 crc kubenswrapper[4813]: I1201 08:59:45.652215 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/b231270a-aff0-442c-9da0-ece1c0540f29-etc-nvme\") pod \"b231270a-aff0-442c-9da0-ece1c0540f29\" (UID: \"b231270a-aff0-442c-9da0-ece1c0540f29\") " Dec 01 08:59:45 crc kubenswrapper[4813]: I1201 08:59:45.652240 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/b231270a-aff0-442c-9da0-ece1c0540f29-lib-modules\") pod \"b231270a-aff0-442c-9da0-ece1c0540f29\" (UID: \"b231270a-aff0-442c-9da0-ece1c0540f29\") " Dec 01 08:59:45 crc kubenswrapper[4813]: I1201 08:59:45.652274 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/b231270a-aff0-442c-9da0-ece1c0540f29-dev\") pod \"b231270a-aff0-442c-9da0-ece1c0540f29\" (UID: \"b231270a-aff0-442c-9da0-ece1c0540f29\") " Dec 01 08:59:45 crc kubenswrapper[4813]: I1201 08:59:45.652291 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/b231270a-aff0-442c-9da0-ece1c0540f29-var-locks-brick\") pod \"b231270a-aff0-442c-9da0-ece1c0540f29\" (UID: \"b231270a-aff0-442c-9da0-ece1c0540f29\") " Dec 01 08:59:45 crc kubenswrapper[4813]: I1201 08:59:45.652346 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/b231270a-aff0-442c-9da0-ece1c0540f29-run\") pod \"b231270a-aff0-442c-9da0-ece1c0540f29\" (UID: \"b231270a-aff0-442c-9da0-ece1c0540f29\") " Dec 01 08:59:45 crc kubenswrapper[4813]: I1201 08:59:45.652362 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b231270a-aff0-442c-9da0-ece1c0540f29-scripts\") pod \"b231270a-aff0-442c-9da0-ece1c0540f29\" (UID: \"b231270a-aff0-442c-9da0-ece1c0540f29\") " Dec 01 08:59:45 crc kubenswrapper[4813]: I1201 08:59:45.652387 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"b231270a-aff0-442c-9da0-ece1c0540f29\" (UID: \"b231270a-aff0-442c-9da0-ece1c0540f29\") " Dec 01 08:59:45 crc kubenswrapper[4813]: I1201 08:59:45.652407 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b231270a-aff0-442c-9da0-ece1c0540f29-logs\") pod \"b231270a-aff0-442c-9da0-ece1c0540f29\" (UID: \"b231270a-aff0-442c-9da0-ece1c0540f29\") " Dec 01 08:59:45 crc kubenswrapper[4813]: I1201 08:59:45.652432 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance-cache\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"b231270a-aff0-442c-9da0-ece1c0540f29\" (UID: \"b231270a-aff0-442c-9da0-ece1c0540f29\") " Dec 01 08:59:45 crc kubenswrapper[4813]: I1201 08:59:45.652460 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b231270a-aff0-442c-9da0-ece1c0540f29-dev" (OuterVolumeSpecName: "dev") pod "b231270a-aff0-442c-9da0-ece1c0540f29" (UID: "b231270a-aff0-442c-9da0-ece1c0540f29"). InnerVolumeSpecName "dev". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 08:59:45 crc kubenswrapper[4813]: I1201 08:59:45.652603 4813 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/92eeaf32-a7b1-4109-82d3-3f4deb6bb8bf-lib-modules\") on node \"crc\" DevicePath \"\"" Dec 01 08:59:45 crc kubenswrapper[4813]: I1201 08:59:45.652615 4813 reconciler_common.go:293] "Volume detached for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/92eeaf32-a7b1-4109-82d3-3f4deb6bb8bf-var-locks-brick\") on node \"crc\" DevicePath \"\"" Dec 01 08:59:45 crc kubenswrapper[4813]: I1201 08:59:45.652625 4813 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b231270a-aff0-442c-9da0-ece1c0540f29-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 08:59:45 crc kubenswrapper[4813]: I1201 08:59:45.652633 4813 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/92eeaf32-a7b1-4109-82d3-3f4deb6bb8bf-logs\") on node \"crc\" DevicePath \"\"" Dec 01 08:59:45 crc kubenswrapper[4813]: I1201 08:59:45.652641 4813 reconciler_common.go:293] "Volume detached for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/b231270a-aff0-442c-9da0-ece1c0540f29-etc-iscsi\") on node \"crc\" DevicePath \"\"" Dec 01 08:59:45 crc kubenswrapper[4813]: I1201 08:59:45.652651 4813 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/92eeaf32-a7b1-4109-82d3-3f4deb6bb8bf-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 08:59:45 crc kubenswrapper[4813]: I1201 08:59:45.652681 4813 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage16-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage16-crc\") on node \"crc\" " Dec 01 08:59:45 crc kubenswrapper[4813]: I1201 08:59:45.652690 4813 reconciler_common.go:293] "Volume detached for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/92eeaf32-a7b1-4109-82d3-3f4deb6bb8bf-etc-iscsi\") on node \"crc\" DevicePath \"\"" Dec 01 08:59:45 crc kubenswrapper[4813]: I1201 08:59:45.652699 4813 reconciler_common.go:293] "Volume detached for volume \"run\" (UniqueName: \"kubernetes.io/host-path/92eeaf32-a7b1-4109-82d3-3f4deb6bb8bf-run\") on node \"crc\" DevicePath \"\"" Dec 01 08:59:45 crc kubenswrapper[4813]: I1201 08:59:45.652707 4813 reconciler_common.go:293] "Volume detached for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/92eeaf32-a7b1-4109-82d3-3f4deb6bb8bf-etc-nvme\") on node \"crc\" DevicePath \"\"" Dec 01 08:59:45 crc kubenswrapper[4813]: I1201 08:59:45.652708 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b231270a-aff0-442c-9da0-ece1c0540f29-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "b231270a-aff0-442c-9da0-ece1c0540f29" (UID: "b231270a-aff0-442c-9da0-ece1c0540f29"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 08:59:45 crc kubenswrapper[4813]: I1201 08:59:45.652714 4813 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/92eeaf32-a7b1-4109-82d3-3f4deb6bb8bf-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 08:59:45 crc kubenswrapper[4813]: I1201 08:59:45.652739 4813 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" " Dec 01 08:59:45 crc kubenswrapper[4813]: I1201 08:59:45.652750 4813 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/92eeaf32-a7b1-4109-82d3-3f4deb6bb8bf-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 01 08:59:45 crc kubenswrapper[4813]: I1201 08:59:45.652760 4813 reconciler_common.go:293] "Volume detached for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/b231270a-aff0-442c-9da0-ece1c0540f29-dev\") on node \"crc\" DevicePath \"\"" Dec 01 08:59:45 crc kubenswrapper[4813]: I1201 08:59:45.652770 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5mj99\" (UniqueName: \"kubernetes.io/projected/92eeaf32-a7b1-4109-82d3-3f4deb6bb8bf-kube-api-access-5mj99\") on node \"crc\" DevicePath \"\"" Dec 01 08:59:45 crc kubenswrapper[4813]: I1201 08:59:45.664239 4813 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage10-crc" (UniqueName: "kubernetes.io/local-volume/local-storage10-crc") on node "crc" Dec 01 08:59:45 crc kubenswrapper[4813]: I1201 08:59:45.664382 4813 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage16-crc" (UniqueName: "kubernetes.io/local-volume/local-storage16-crc") on node "crc" Dec 01 08:59:45 crc kubenswrapper[4813]: I1201 08:59:45.665047 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b231270a-aff0-442c-9da0-ece1c0540f29-etc-nvme" (OuterVolumeSpecName: "etc-nvme") pod "b231270a-aff0-442c-9da0-ece1c0540f29" (UID: "b231270a-aff0-442c-9da0-ece1c0540f29"). InnerVolumeSpecName "etc-nvme". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 08:59:45 crc kubenswrapper[4813]: I1201 08:59:45.665149 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b231270a-aff0-442c-9da0-ece1c0540f29-logs" (OuterVolumeSpecName: "logs") pod "b231270a-aff0-442c-9da0-ece1c0540f29" (UID: "b231270a-aff0-442c-9da0-ece1c0540f29"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 08:59:45 crc kubenswrapper[4813]: I1201 08:59:45.665216 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b231270a-aff0-442c-9da0-ece1c0540f29-run" (OuterVolumeSpecName: "run") pod "b231270a-aff0-442c-9da0-ece1c0540f29" (UID: "b231270a-aff0-442c-9da0-ece1c0540f29"). InnerVolumeSpecName "run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 08:59:45 crc kubenswrapper[4813]: I1201 08:59:45.665225 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b231270a-aff0-442c-9da0-ece1c0540f29-var-locks-brick" (OuterVolumeSpecName: "var-locks-brick") pod "b231270a-aff0-442c-9da0-ece1c0540f29" (UID: "b231270a-aff0-442c-9da0-ece1c0540f29"). InnerVolumeSpecName "var-locks-brick". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 08:59:45 crc kubenswrapper[4813]: I1201 08:59:45.665238 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b231270a-aff0-442c-9da0-ece1c0540f29-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "b231270a-aff0-442c-9da0-ece1c0540f29" (UID: "b231270a-aff0-442c-9da0-ece1c0540f29"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 08:59:45 crc kubenswrapper[4813]: I1201 08:59:45.667124 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage01-crc" (OuterVolumeSpecName: "glance-cache") pod "b231270a-aff0-442c-9da0-ece1c0540f29" (UID: "b231270a-aff0-442c-9da0-ece1c0540f29"). InnerVolumeSpecName "local-storage01-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 01 08:59:45 crc kubenswrapper[4813]: I1201 08:59:45.667946 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b231270a-aff0-442c-9da0-ece1c0540f29-scripts" (OuterVolumeSpecName: "scripts") pod "b231270a-aff0-442c-9da0-ece1c0540f29" (UID: "b231270a-aff0-442c-9da0-ece1c0540f29"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:59:45 crc kubenswrapper[4813]: I1201 08:59:45.668264 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b231270a-aff0-442c-9da0-ece1c0540f29-kube-api-access-z88bx" (OuterVolumeSpecName: "kube-api-access-z88bx") pod "b231270a-aff0-442c-9da0-ece1c0540f29" (UID: "b231270a-aff0-442c-9da0-ece1c0540f29"). InnerVolumeSpecName "kube-api-access-z88bx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:59:45 crc kubenswrapper[4813]: I1201 08:59:45.670038 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage05-crc" (OuterVolumeSpecName: "glance") pod "b231270a-aff0-442c-9da0-ece1c0540f29" (UID: "b231270a-aff0-442c-9da0-ece1c0540f29"). InnerVolumeSpecName "local-storage05-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 01 08:59:45 crc kubenswrapper[4813]: I1201 08:59:45.754262 4813 reconciler_common.go:293] "Volume detached for volume \"run\" (UniqueName: \"kubernetes.io/host-path/b231270a-aff0-442c-9da0-ece1c0540f29-run\") on node \"crc\" DevicePath \"\"" Dec 01 08:59:45 crc kubenswrapper[4813]: I1201 08:59:45.754328 4813 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b231270a-aff0-442c-9da0-ece1c0540f29-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 08:59:45 crc kubenswrapper[4813]: I1201 08:59:45.754367 4813 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" " Dec 01 08:59:45 crc kubenswrapper[4813]: I1201 08:59:45.754411 4813 reconciler_common.go:293] "Volume detached for volume \"local-storage16-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage16-crc\") on node \"crc\" DevicePath \"\"" Dec 01 08:59:45 crc kubenswrapper[4813]: I1201 08:59:45.754424 4813 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b231270a-aff0-442c-9da0-ece1c0540f29-logs\") on node \"crc\" DevicePath \"\"" Dec 01 08:59:45 crc kubenswrapper[4813]: I1201 08:59:45.754442 4813 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" " Dec 01 08:59:45 crc kubenswrapper[4813]: I1201 08:59:45.754455 4813 reconciler_common.go:293] "Volume detached for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" DevicePath \"\"" Dec 01 08:59:45 crc kubenswrapper[4813]: I1201 08:59:45.754500 4813 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b231270a-aff0-442c-9da0-ece1c0540f29-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 01 08:59:45 crc kubenswrapper[4813]: I1201 08:59:45.754515 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z88bx\" (UniqueName: \"kubernetes.io/projected/b231270a-aff0-442c-9da0-ece1c0540f29-kube-api-access-z88bx\") on node \"crc\" DevicePath \"\"" Dec 01 08:59:45 crc kubenswrapper[4813]: I1201 08:59:45.754525 4813 reconciler_common.go:293] "Volume detached for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/b231270a-aff0-442c-9da0-ece1c0540f29-etc-nvme\") on node \"crc\" DevicePath \"\"" Dec 01 08:59:45 crc kubenswrapper[4813]: I1201 08:59:45.754536 4813 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/b231270a-aff0-442c-9da0-ece1c0540f29-lib-modules\") on node \"crc\" DevicePath \"\"" Dec 01 08:59:45 crc kubenswrapper[4813]: I1201 08:59:45.754576 4813 reconciler_common.go:293] "Volume detached for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/b231270a-aff0-442c-9da0-ece1c0540f29-var-locks-brick\") on node \"crc\" DevicePath \"\"" Dec 01 08:59:45 crc kubenswrapper[4813]: I1201 08:59:45.769096 4813 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage01-crc" (UniqueName: "kubernetes.io/local-volume/local-storage01-crc") on node "crc" Dec 01 08:59:45 crc kubenswrapper[4813]: I1201 08:59:45.784997 4813 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage05-crc" (UniqueName: "kubernetes.io/local-volume/local-storage05-crc") on node "crc" Dec 01 08:59:45 crc kubenswrapper[4813]: I1201 08:59:45.855920 4813 reconciler_common.go:293] "Volume detached for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" DevicePath \"\"" Dec 01 08:59:45 crc kubenswrapper[4813]: I1201 08:59:45.856032 4813 reconciler_common.go:293] "Volume detached for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" DevicePath \"\"" Dec 01 08:59:46 crc kubenswrapper[4813]: I1201 08:59:46.059739 4813 generic.go:334] "Generic (PLEG): container finished" podID="92eeaf32-a7b1-4109-82d3-3f4deb6bb8bf" containerID="dd9f2262f5b6023b129ed5ef64f88a24261537a4a8be9e435d043c151fdc6059" exitCode=0 Dec 01 08:59:46 crc kubenswrapper[4813]: I1201 08:59:46.059812 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"92eeaf32-a7b1-4109-82d3-3f4deb6bb8bf","Type":"ContainerDied","Data":"dd9f2262f5b6023b129ed5ef64f88a24261537a4a8be9e435d043c151fdc6059"} Dec 01 08:59:46 crc kubenswrapper[4813]: I1201 08:59:46.059842 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"92eeaf32-a7b1-4109-82d3-3f4deb6bb8bf","Type":"ContainerDied","Data":"df218fe74d63795a3c93d09f2acecc950b610411ad8caae4d59f91896597293b"} Dec 01 08:59:46 crc kubenswrapper[4813]: I1201 08:59:46.059837 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Dec 01 08:59:46 crc kubenswrapper[4813]: I1201 08:59:46.059891 4813 scope.go:117] "RemoveContainer" containerID="dd9f2262f5b6023b129ed5ef64f88a24261537a4a8be9e435d043c151fdc6059" Dec 01 08:59:46 crc kubenswrapper[4813]: I1201 08:59:46.064107 4813 generic.go:334] "Generic (PLEG): container finished" podID="b231270a-aff0-442c-9da0-ece1c0540f29" containerID="5d49aad4c62c639f57ff39f38b2afd9ac5d3c22513ab579c17ebcce6f29db9a0" exitCode=0 Dec 01 08:59:46 crc kubenswrapper[4813]: I1201 08:59:46.064146 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-1" event={"ID":"b231270a-aff0-442c-9da0-ece1c0540f29","Type":"ContainerDied","Data":"5d49aad4c62c639f57ff39f38b2afd9ac5d3c22513ab579c17ebcce6f29db9a0"} Dec 01 08:59:46 crc kubenswrapper[4813]: I1201 08:59:46.064225 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-1" event={"ID":"b231270a-aff0-442c-9da0-ece1c0540f29","Type":"ContainerDied","Data":"012eb03987f00dc3a6b76d7725b4deb8494670bddbdb9af4b9ad2fb6d610ee96"} Dec 01 08:59:46 crc kubenswrapper[4813]: I1201 08:59:46.065192 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-1" Dec 01 08:59:46 crc kubenswrapper[4813]: I1201 08:59:46.103187 4813 scope.go:117] "RemoveContainer" containerID="67d76dcb531e0c7369cc1c8a9943f18d7dacfde17ed6aef5f657851ffab723dc" Dec 01 08:59:46 crc kubenswrapper[4813]: I1201 08:59:46.121249 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Dec 01 08:59:46 crc kubenswrapper[4813]: I1201 08:59:46.128026 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Dec 01 08:59:46 crc kubenswrapper[4813]: I1201 08:59:46.139430 4813 scope.go:117] "RemoveContainer" containerID="dd9f2262f5b6023b129ed5ef64f88a24261537a4a8be9e435d043c151fdc6059" Dec 01 08:59:46 crc kubenswrapper[4813]: I1201 08:59:46.139760 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-single-1"] Dec 01 08:59:46 crc kubenswrapper[4813]: E1201 08:59:46.140368 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dd9f2262f5b6023b129ed5ef64f88a24261537a4a8be9e435d043c151fdc6059\": container with ID starting with dd9f2262f5b6023b129ed5ef64f88a24261537a4a8be9e435d043c151fdc6059 not found: ID does not exist" containerID="dd9f2262f5b6023b129ed5ef64f88a24261537a4a8be9e435d043c151fdc6059" Dec 01 08:59:46 crc kubenswrapper[4813]: I1201 08:59:46.141041 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dd9f2262f5b6023b129ed5ef64f88a24261537a4a8be9e435d043c151fdc6059"} err="failed to get container status \"dd9f2262f5b6023b129ed5ef64f88a24261537a4a8be9e435d043c151fdc6059\": rpc error: code = NotFound desc = could not find container \"dd9f2262f5b6023b129ed5ef64f88a24261537a4a8be9e435d043c151fdc6059\": container with ID starting with dd9f2262f5b6023b129ed5ef64f88a24261537a4a8be9e435d043c151fdc6059 not found: ID does not exist" Dec 01 08:59:46 crc kubenswrapper[4813]: I1201 08:59:46.141353 4813 scope.go:117] "RemoveContainer" containerID="67d76dcb531e0c7369cc1c8a9943f18d7dacfde17ed6aef5f657851ffab723dc" Dec 01 08:59:46 crc kubenswrapper[4813]: E1201 08:59:46.145564 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"67d76dcb531e0c7369cc1c8a9943f18d7dacfde17ed6aef5f657851ffab723dc\": container with ID starting with 67d76dcb531e0c7369cc1c8a9943f18d7dacfde17ed6aef5f657851ffab723dc not found: ID does not exist" containerID="67d76dcb531e0c7369cc1c8a9943f18d7dacfde17ed6aef5f657851ffab723dc" Dec 01 08:59:46 crc kubenswrapper[4813]: I1201 08:59:46.145719 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"67d76dcb531e0c7369cc1c8a9943f18d7dacfde17ed6aef5f657851ffab723dc"} err="failed to get container status \"67d76dcb531e0c7369cc1c8a9943f18d7dacfde17ed6aef5f657851ffab723dc\": rpc error: code = NotFound desc = could not find container \"67d76dcb531e0c7369cc1c8a9943f18d7dacfde17ed6aef5f657851ffab723dc\": container with ID starting with 67d76dcb531e0c7369cc1c8a9943f18d7dacfde17ed6aef5f657851ffab723dc not found: ID does not exist" Dec 01 08:59:46 crc kubenswrapper[4813]: I1201 08:59:46.145844 4813 scope.go:117] "RemoveContainer" containerID="5d49aad4c62c639f57ff39f38b2afd9ac5d3c22513ab579c17ebcce6f29db9a0" Dec 01 08:59:46 crc kubenswrapper[4813]: I1201 08:59:46.146857 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-default-single-1"] Dec 01 08:59:46 crc kubenswrapper[4813]: I1201 08:59:46.173841 4813 scope.go:117] "RemoveContainer" containerID="8180eaad882643280dbedf20c976cbd80a759bd80dfdfb74e42c9ec5beb7973a" Dec 01 08:59:46 crc kubenswrapper[4813]: I1201 08:59:46.187373 4813 scope.go:117] "RemoveContainer" containerID="5d49aad4c62c639f57ff39f38b2afd9ac5d3c22513ab579c17ebcce6f29db9a0" Dec 01 08:59:46 crc kubenswrapper[4813]: E1201 08:59:46.188058 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5d49aad4c62c639f57ff39f38b2afd9ac5d3c22513ab579c17ebcce6f29db9a0\": container with ID starting with 5d49aad4c62c639f57ff39f38b2afd9ac5d3c22513ab579c17ebcce6f29db9a0 not found: ID does not exist" containerID="5d49aad4c62c639f57ff39f38b2afd9ac5d3c22513ab579c17ebcce6f29db9a0" Dec 01 08:59:46 crc kubenswrapper[4813]: I1201 08:59:46.188215 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5d49aad4c62c639f57ff39f38b2afd9ac5d3c22513ab579c17ebcce6f29db9a0"} err="failed to get container status \"5d49aad4c62c639f57ff39f38b2afd9ac5d3c22513ab579c17ebcce6f29db9a0\": rpc error: code = NotFound desc = could not find container \"5d49aad4c62c639f57ff39f38b2afd9ac5d3c22513ab579c17ebcce6f29db9a0\": container with ID starting with 5d49aad4c62c639f57ff39f38b2afd9ac5d3c22513ab579c17ebcce6f29db9a0 not found: ID does not exist" Dec 01 08:59:46 crc kubenswrapper[4813]: I1201 08:59:46.188312 4813 scope.go:117] "RemoveContainer" containerID="8180eaad882643280dbedf20c976cbd80a759bd80dfdfb74e42c9ec5beb7973a" Dec 01 08:59:46 crc kubenswrapper[4813]: E1201 08:59:46.188700 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8180eaad882643280dbedf20c976cbd80a759bd80dfdfb74e42c9ec5beb7973a\": container with ID starting with 8180eaad882643280dbedf20c976cbd80a759bd80dfdfb74e42c9ec5beb7973a not found: ID does not exist" containerID="8180eaad882643280dbedf20c976cbd80a759bd80dfdfb74e42c9ec5beb7973a" Dec 01 08:59:46 crc kubenswrapper[4813]: I1201 08:59:46.188735 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8180eaad882643280dbedf20c976cbd80a759bd80dfdfb74e42c9ec5beb7973a"} err="failed to get container status \"8180eaad882643280dbedf20c976cbd80a759bd80dfdfb74e42c9ec5beb7973a\": rpc error: code = NotFound desc = could not find container \"8180eaad882643280dbedf20c976cbd80a759bd80dfdfb74e42c9ec5beb7973a\": container with ID starting with 8180eaad882643280dbedf20c976cbd80a759bd80dfdfb74e42c9ec5beb7973a not found: ID does not exist" Dec 01 08:59:46 crc kubenswrapper[4813]: I1201 08:59:46.407504 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="92eeaf32-a7b1-4109-82d3-3f4deb6bb8bf" path="/var/lib/kubelet/pods/92eeaf32-a7b1-4109-82d3-3f4deb6bb8bf/volumes" Dec 01 08:59:46 crc kubenswrapper[4813]: I1201 08:59:46.409050 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b231270a-aff0-442c-9da0-ece1c0540f29" path="/var/lib/kubelet/pods/b231270a-aff0-442c-9da0-ece1c0540f29/volumes" Dec 01 08:59:46 crc kubenswrapper[4813]: I1201 08:59:46.655225 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-db-create-fb2hf"] Dec 01 08:59:46 crc kubenswrapper[4813]: I1201 08:59:46.669450 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-db-create-fb2hf"] Dec 01 08:59:46 crc kubenswrapper[4813]: I1201 08:59:46.678436 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-d205-account-create-update-mgb49"] Dec 01 08:59:46 crc kubenswrapper[4813]: I1201 08:59:46.684365 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glanced205-account-delete-s9x5m"] Dec 01 08:59:46 crc kubenswrapper[4813]: I1201 08:59:46.689276 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-d205-account-create-update-mgb49"] Dec 01 08:59:46 crc kubenswrapper[4813]: I1201 08:59:46.694670 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glanced205-account-delete-s9x5m"] Dec 01 08:59:48 crc kubenswrapper[4813]: I1201 08:59:48.333467 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-db-create-cg9jd"] Dec 01 08:59:48 crc kubenswrapper[4813]: E1201 08:59:48.333856 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b231270a-aff0-442c-9da0-ece1c0540f29" containerName="glance-log" Dec 01 08:59:48 crc kubenswrapper[4813]: I1201 08:59:48.333889 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="b231270a-aff0-442c-9da0-ece1c0540f29" containerName="glance-log" Dec 01 08:59:48 crc kubenswrapper[4813]: E1201 08:59:48.333899 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b231270a-aff0-442c-9da0-ece1c0540f29" containerName="glance-httpd" Dec 01 08:59:48 crc kubenswrapper[4813]: I1201 08:59:48.333908 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="b231270a-aff0-442c-9da0-ece1c0540f29" containerName="glance-httpd" Dec 01 08:59:48 crc kubenswrapper[4813]: E1201 08:59:48.333923 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="79ccaf0c-6d3c-4539-b716-72d794c0eb64" containerName="openstackclient" Dec 01 08:59:48 crc kubenswrapper[4813]: I1201 08:59:48.333930 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="79ccaf0c-6d3c-4539-b716-72d794c0eb64" containerName="openstackclient" Dec 01 08:59:48 crc kubenswrapper[4813]: E1201 08:59:48.333952 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="92eeaf32-a7b1-4109-82d3-3f4deb6bb8bf" containerName="glance-httpd" Dec 01 08:59:48 crc kubenswrapper[4813]: I1201 08:59:48.333958 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="92eeaf32-a7b1-4109-82d3-3f4deb6bb8bf" containerName="glance-httpd" Dec 01 08:59:48 crc kubenswrapper[4813]: E1201 08:59:48.333990 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ef8750cf-a451-47f9-a168-380246f7fbfd" containerName="mariadb-account-delete" Dec 01 08:59:48 crc kubenswrapper[4813]: I1201 08:59:48.333996 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="ef8750cf-a451-47f9-a168-380246f7fbfd" containerName="mariadb-account-delete" Dec 01 08:59:48 crc kubenswrapper[4813]: E1201 08:59:48.334007 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="92eeaf32-a7b1-4109-82d3-3f4deb6bb8bf" containerName="glance-log" Dec 01 08:59:48 crc kubenswrapper[4813]: I1201 08:59:48.334013 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="92eeaf32-a7b1-4109-82d3-3f4deb6bb8bf" containerName="glance-log" Dec 01 08:59:48 crc kubenswrapper[4813]: I1201 08:59:48.334159 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="b231270a-aff0-442c-9da0-ece1c0540f29" containerName="glance-httpd" Dec 01 08:59:48 crc kubenswrapper[4813]: I1201 08:59:48.334180 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="92eeaf32-a7b1-4109-82d3-3f4deb6bb8bf" containerName="glance-httpd" Dec 01 08:59:48 crc kubenswrapper[4813]: I1201 08:59:48.334191 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="92eeaf32-a7b1-4109-82d3-3f4deb6bb8bf" containerName="glance-log" Dec 01 08:59:48 crc kubenswrapper[4813]: I1201 08:59:48.334199 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="ef8750cf-a451-47f9-a168-380246f7fbfd" containerName="mariadb-account-delete" Dec 01 08:59:48 crc kubenswrapper[4813]: I1201 08:59:48.334210 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="b231270a-aff0-442c-9da0-ece1c0540f29" containerName="glance-log" Dec 01 08:59:48 crc kubenswrapper[4813]: I1201 08:59:48.334217 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="79ccaf0c-6d3c-4539-b716-72d794c0eb64" containerName="openstackclient" Dec 01 08:59:48 crc kubenswrapper[4813]: I1201 08:59:48.334735 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-cg9jd" Dec 01 08:59:48 crc kubenswrapper[4813]: I1201 08:59:48.342836 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-b0ca-account-create-update-7wbnc"] Dec 01 08:59:48 crc kubenswrapper[4813]: I1201 08:59:48.344807 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-b0ca-account-create-update-7wbnc" Dec 01 08:59:48 crc kubenswrapper[4813]: I1201 08:59:48.348567 4813 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-db-secret" Dec 01 08:59:48 crc kubenswrapper[4813]: I1201 08:59:48.349551 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-create-cg9jd"] Dec 01 08:59:48 crc kubenswrapper[4813]: I1201 08:59:48.355762 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-b0ca-account-create-update-7wbnc"] Dec 01 08:59:48 crc kubenswrapper[4813]: I1201 08:59:48.407406 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7dc297d9-6c3f-49e7-bb11-25697b36da14" path="/var/lib/kubelet/pods/7dc297d9-6c3f-49e7-bb11-25697b36da14/volumes" Dec 01 08:59:48 crc kubenswrapper[4813]: I1201 08:59:48.408371 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="905a377a-0608-4830-abc3-10b9404846b7" path="/var/lib/kubelet/pods/905a377a-0608-4830-abc3-10b9404846b7/volumes" Dec 01 08:59:48 crc kubenswrapper[4813]: I1201 08:59:48.409173 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ef8750cf-a451-47f9-a168-380246f7fbfd" path="/var/lib/kubelet/pods/ef8750cf-a451-47f9-a168-380246f7fbfd/volumes" Dec 01 08:59:48 crc kubenswrapper[4813]: I1201 08:59:48.452815 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2e29ec70-392c-45e5-b8bb-28c604bfbcb9-operator-scripts\") pod \"glance-b0ca-account-create-update-7wbnc\" (UID: \"2e29ec70-392c-45e5-b8bb-28c604bfbcb9\") " pod="glance-kuttl-tests/glance-b0ca-account-create-update-7wbnc" Dec 01 08:59:48 crc kubenswrapper[4813]: I1201 08:59:48.453239 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8fd46b3b-b57c-45ff-ad5e-45a10050cfa7-operator-scripts\") pod \"glance-db-create-cg9jd\" (UID: \"8fd46b3b-b57c-45ff-ad5e-45a10050cfa7\") " pod="glance-kuttl-tests/glance-db-create-cg9jd" Dec 01 08:59:48 crc kubenswrapper[4813]: I1201 08:59:48.453557 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6sgrg\" (UniqueName: \"kubernetes.io/projected/2e29ec70-392c-45e5-b8bb-28c604bfbcb9-kube-api-access-6sgrg\") pod \"glance-b0ca-account-create-update-7wbnc\" (UID: \"2e29ec70-392c-45e5-b8bb-28c604bfbcb9\") " pod="glance-kuttl-tests/glance-b0ca-account-create-update-7wbnc" Dec 01 08:59:48 crc kubenswrapper[4813]: I1201 08:59:48.453884 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dnfhx\" (UniqueName: \"kubernetes.io/projected/8fd46b3b-b57c-45ff-ad5e-45a10050cfa7-kube-api-access-dnfhx\") pod \"glance-db-create-cg9jd\" (UID: \"8fd46b3b-b57c-45ff-ad5e-45a10050cfa7\") " pod="glance-kuttl-tests/glance-db-create-cg9jd" Dec 01 08:59:48 crc kubenswrapper[4813]: I1201 08:59:48.555347 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dnfhx\" (UniqueName: \"kubernetes.io/projected/8fd46b3b-b57c-45ff-ad5e-45a10050cfa7-kube-api-access-dnfhx\") pod \"glance-db-create-cg9jd\" (UID: \"8fd46b3b-b57c-45ff-ad5e-45a10050cfa7\") " pod="glance-kuttl-tests/glance-db-create-cg9jd" Dec 01 08:59:48 crc kubenswrapper[4813]: I1201 08:59:48.555407 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2e29ec70-392c-45e5-b8bb-28c604bfbcb9-operator-scripts\") pod \"glance-b0ca-account-create-update-7wbnc\" (UID: \"2e29ec70-392c-45e5-b8bb-28c604bfbcb9\") " pod="glance-kuttl-tests/glance-b0ca-account-create-update-7wbnc" Dec 01 08:59:48 crc kubenswrapper[4813]: I1201 08:59:48.555455 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8fd46b3b-b57c-45ff-ad5e-45a10050cfa7-operator-scripts\") pod \"glance-db-create-cg9jd\" (UID: \"8fd46b3b-b57c-45ff-ad5e-45a10050cfa7\") " pod="glance-kuttl-tests/glance-db-create-cg9jd" Dec 01 08:59:48 crc kubenswrapper[4813]: I1201 08:59:48.555517 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6sgrg\" (UniqueName: \"kubernetes.io/projected/2e29ec70-392c-45e5-b8bb-28c604bfbcb9-kube-api-access-6sgrg\") pod \"glance-b0ca-account-create-update-7wbnc\" (UID: \"2e29ec70-392c-45e5-b8bb-28c604bfbcb9\") " pod="glance-kuttl-tests/glance-b0ca-account-create-update-7wbnc" Dec 01 08:59:48 crc kubenswrapper[4813]: I1201 08:59:48.556550 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8fd46b3b-b57c-45ff-ad5e-45a10050cfa7-operator-scripts\") pod \"glance-db-create-cg9jd\" (UID: \"8fd46b3b-b57c-45ff-ad5e-45a10050cfa7\") " pod="glance-kuttl-tests/glance-db-create-cg9jd" Dec 01 08:59:48 crc kubenswrapper[4813]: I1201 08:59:48.556803 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2e29ec70-392c-45e5-b8bb-28c604bfbcb9-operator-scripts\") pod \"glance-b0ca-account-create-update-7wbnc\" (UID: \"2e29ec70-392c-45e5-b8bb-28c604bfbcb9\") " pod="glance-kuttl-tests/glance-b0ca-account-create-update-7wbnc" Dec 01 08:59:48 crc kubenswrapper[4813]: I1201 08:59:48.579560 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6sgrg\" (UniqueName: \"kubernetes.io/projected/2e29ec70-392c-45e5-b8bb-28c604bfbcb9-kube-api-access-6sgrg\") pod \"glance-b0ca-account-create-update-7wbnc\" (UID: \"2e29ec70-392c-45e5-b8bb-28c604bfbcb9\") " pod="glance-kuttl-tests/glance-b0ca-account-create-update-7wbnc" Dec 01 08:59:48 crc kubenswrapper[4813]: I1201 08:59:48.580048 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dnfhx\" (UniqueName: \"kubernetes.io/projected/8fd46b3b-b57c-45ff-ad5e-45a10050cfa7-kube-api-access-dnfhx\") pod \"glance-db-create-cg9jd\" (UID: \"8fd46b3b-b57c-45ff-ad5e-45a10050cfa7\") " pod="glance-kuttl-tests/glance-db-create-cg9jd" Dec 01 08:59:48 crc kubenswrapper[4813]: I1201 08:59:48.658389 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-cg9jd" Dec 01 08:59:48 crc kubenswrapper[4813]: I1201 08:59:48.679685 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-b0ca-account-create-update-7wbnc" Dec 01 08:59:49 crc kubenswrapper[4813]: I1201 08:59:49.202415 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-create-cg9jd"] Dec 01 08:59:49 crc kubenswrapper[4813]: I1201 08:59:49.209641 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-b0ca-account-create-update-7wbnc"] Dec 01 08:59:50 crc kubenswrapper[4813]: I1201 08:59:50.108783 4813 generic.go:334] "Generic (PLEG): container finished" podID="8fd46b3b-b57c-45ff-ad5e-45a10050cfa7" containerID="69c747470314cd28ca10e4f24bc635fc4aec16e2a4fc9c19015931af4655e644" exitCode=0 Dec 01 08:59:50 crc kubenswrapper[4813]: I1201 08:59:50.108924 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-create-cg9jd" event={"ID":"8fd46b3b-b57c-45ff-ad5e-45a10050cfa7","Type":"ContainerDied","Data":"69c747470314cd28ca10e4f24bc635fc4aec16e2a4fc9c19015931af4655e644"} Dec 01 08:59:50 crc kubenswrapper[4813]: I1201 08:59:50.109105 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-create-cg9jd" event={"ID":"8fd46b3b-b57c-45ff-ad5e-45a10050cfa7","Type":"ContainerStarted","Data":"0c9b27956f2f0e36668056ed10ee1ea5f5d797148c0da197203bd830e4db81d2"} Dec 01 08:59:50 crc kubenswrapper[4813]: I1201 08:59:50.111383 4813 generic.go:334] "Generic (PLEG): container finished" podID="2e29ec70-392c-45e5-b8bb-28c604bfbcb9" containerID="25b3a5735cd7eaa24c2967139820a3df74e032619e2a68233879342cdf48fa2c" exitCode=0 Dec 01 08:59:50 crc kubenswrapper[4813]: I1201 08:59:50.111411 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-b0ca-account-create-update-7wbnc" event={"ID":"2e29ec70-392c-45e5-b8bb-28c604bfbcb9","Type":"ContainerDied","Data":"25b3a5735cd7eaa24c2967139820a3df74e032619e2a68233879342cdf48fa2c"} Dec 01 08:59:50 crc kubenswrapper[4813]: I1201 08:59:50.111428 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-b0ca-account-create-update-7wbnc" event={"ID":"2e29ec70-392c-45e5-b8bb-28c604bfbcb9","Type":"ContainerStarted","Data":"a372d0c2a50a1c0f30d0c4df9624342e0190b377bd037327fa85c0d9a969c9e4"} Dec 01 08:59:51 crc kubenswrapper[4813]: I1201 08:59:51.530474 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-cg9jd" Dec 01 08:59:51 crc kubenswrapper[4813]: I1201 08:59:51.536842 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-b0ca-account-create-update-7wbnc" Dec 01 08:59:51 crc kubenswrapper[4813]: I1201 08:59:51.611645 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6sgrg\" (UniqueName: \"kubernetes.io/projected/2e29ec70-392c-45e5-b8bb-28c604bfbcb9-kube-api-access-6sgrg\") pod \"2e29ec70-392c-45e5-b8bb-28c604bfbcb9\" (UID: \"2e29ec70-392c-45e5-b8bb-28c604bfbcb9\") " Dec 01 08:59:51 crc kubenswrapper[4813]: I1201 08:59:51.611746 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2e29ec70-392c-45e5-b8bb-28c604bfbcb9-operator-scripts\") pod \"2e29ec70-392c-45e5-b8bb-28c604bfbcb9\" (UID: \"2e29ec70-392c-45e5-b8bb-28c604bfbcb9\") " Dec 01 08:59:51 crc kubenswrapper[4813]: I1201 08:59:51.611795 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8fd46b3b-b57c-45ff-ad5e-45a10050cfa7-operator-scripts\") pod \"8fd46b3b-b57c-45ff-ad5e-45a10050cfa7\" (UID: \"8fd46b3b-b57c-45ff-ad5e-45a10050cfa7\") " Dec 01 08:59:51 crc kubenswrapper[4813]: I1201 08:59:51.611891 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dnfhx\" (UniqueName: \"kubernetes.io/projected/8fd46b3b-b57c-45ff-ad5e-45a10050cfa7-kube-api-access-dnfhx\") pod \"8fd46b3b-b57c-45ff-ad5e-45a10050cfa7\" (UID: \"8fd46b3b-b57c-45ff-ad5e-45a10050cfa7\") " Dec 01 08:59:51 crc kubenswrapper[4813]: I1201 08:59:51.612609 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2e29ec70-392c-45e5-b8bb-28c604bfbcb9-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "2e29ec70-392c-45e5-b8bb-28c604bfbcb9" (UID: "2e29ec70-392c-45e5-b8bb-28c604bfbcb9"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:59:51 crc kubenswrapper[4813]: I1201 08:59:51.613018 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8fd46b3b-b57c-45ff-ad5e-45a10050cfa7-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "8fd46b3b-b57c-45ff-ad5e-45a10050cfa7" (UID: "8fd46b3b-b57c-45ff-ad5e-45a10050cfa7"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 08:59:51 crc kubenswrapper[4813]: I1201 08:59:51.619431 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2e29ec70-392c-45e5-b8bb-28c604bfbcb9-kube-api-access-6sgrg" (OuterVolumeSpecName: "kube-api-access-6sgrg") pod "2e29ec70-392c-45e5-b8bb-28c604bfbcb9" (UID: "2e29ec70-392c-45e5-b8bb-28c604bfbcb9"). InnerVolumeSpecName "kube-api-access-6sgrg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:59:51 crc kubenswrapper[4813]: I1201 08:59:51.621572 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8fd46b3b-b57c-45ff-ad5e-45a10050cfa7-kube-api-access-dnfhx" (OuterVolumeSpecName: "kube-api-access-dnfhx") pod "8fd46b3b-b57c-45ff-ad5e-45a10050cfa7" (UID: "8fd46b3b-b57c-45ff-ad5e-45a10050cfa7"). InnerVolumeSpecName "kube-api-access-dnfhx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:59:51 crc kubenswrapper[4813]: I1201 08:59:51.713341 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6sgrg\" (UniqueName: \"kubernetes.io/projected/2e29ec70-392c-45e5-b8bb-28c604bfbcb9-kube-api-access-6sgrg\") on node \"crc\" DevicePath \"\"" Dec 01 08:59:51 crc kubenswrapper[4813]: I1201 08:59:51.713552 4813 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2e29ec70-392c-45e5-b8bb-28c604bfbcb9-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 08:59:51 crc kubenswrapper[4813]: I1201 08:59:51.713611 4813 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8fd46b3b-b57c-45ff-ad5e-45a10050cfa7-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 08:59:51 crc kubenswrapper[4813]: I1201 08:59:51.713665 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dnfhx\" (UniqueName: \"kubernetes.io/projected/8fd46b3b-b57c-45ff-ad5e-45a10050cfa7-kube-api-access-dnfhx\") on node \"crc\" DevicePath \"\"" Dec 01 08:59:52 crc kubenswrapper[4813]: I1201 08:59:52.130542 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-cg9jd" Dec 01 08:59:52 crc kubenswrapper[4813]: I1201 08:59:52.130536 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-create-cg9jd" event={"ID":"8fd46b3b-b57c-45ff-ad5e-45a10050cfa7","Type":"ContainerDied","Data":"0c9b27956f2f0e36668056ed10ee1ea5f5d797148c0da197203bd830e4db81d2"} Dec 01 08:59:52 crc kubenswrapper[4813]: I1201 08:59:52.130722 4813 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0c9b27956f2f0e36668056ed10ee1ea5f5d797148c0da197203bd830e4db81d2" Dec 01 08:59:52 crc kubenswrapper[4813]: I1201 08:59:52.132506 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-b0ca-account-create-update-7wbnc" event={"ID":"2e29ec70-392c-45e5-b8bb-28c604bfbcb9","Type":"ContainerDied","Data":"a372d0c2a50a1c0f30d0c4df9624342e0190b377bd037327fa85c0d9a969c9e4"} Dec 01 08:59:52 crc kubenswrapper[4813]: I1201 08:59:52.132548 4813 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a372d0c2a50a1c0f30d0c4df9624342e0190b377bd037327fa85c0d9a969c9e4" Dec 01 08:59:52 crc kubenswrapper[4813]: I1201 08:59:52.132559 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-b0ca-account-create-update-7wbnc" Dec 01 08:59:53 crc kubenswrapper[4813]: I1201 08:59:53.455638 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-db-sync-pkc4b"] Dec 01 08:59:53 crc kubenswrapper[4813]: E1201 08:59:53.456502 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8fd46b3b-b57c-45ff-ad5e-45a10050cfa7" containerName="mariadb-database-create" Dec 01 08:59:53 crc kubenswrapper[4813]: I1201 08:59:53.456516 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="8fd46b3b-b57c-45ff-ad5e-45a10050cfa7" containerName="mariadb-database-create" Dec 01 08:59:53 crc kubenswrapper[4813]: E1201 08:59:53.456535 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2e29ec70-392c-45e5-b8bb-28c604bfbcb9" containerName="mariadb-account-create-update" Dec 01 08:59:53 crc kubenswrapper[4813]: I1201 08:59:53.456541 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="2e29ec70-392c-45e5-b8bb-28c604bfbcb9" containerName="mariadb-account-create-update" Dec 01 08:59:53 crc kubenswrapper[4813]: I1201 08:59:53.456668 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="8fd46b3b-b57c-45ff-ad5e-45a10050cfa7" containerName="mariadb-database-create" Dec 01 08:59:53 crc kubenswrapper[4813]: I1201 08:59:53.456683 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="2e29ec70-392c-45e5-b8bb-28c604bfbcb9" containerName="mariadb-account-create-update" Dec 01 08:59:53 crc kubenswrapper[4813]: I1201 08:59:53.457274 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-pkc4b" Dec 01 08:59:53 crc kubenswrapper[4813]: I1201 08:59:53.460785 4813 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-glance-dockercfg-srxng" Dec 01 08:59:53 crc kubenswrapper[4813]: I1201 08:59:53.461041 4813 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-config-data" Dec 01 08:59:53 crc kubenswrapper[4813]: I1201 08:59:53.461184 4813 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"combined-ca-bundle" Dec 01 08:59:53 crc kubenswrapper[4813]: I1201 08:59:53.467695 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8k8nv\" (UniqueName: \"kubernetes.io/projected/e7487c91-a8ac-4c0a-ac6e-6263890da1f2-kube-api-access-8k8nv\") pod \"glance-db-sync-pkc4b\" (UID: \"e7487c91-a8ac-4c0a-ac6e-6263890da1f2\") " pod="glance-kuttl-tests/glance-db-sync-pkc4b" Dec 01 08:59:53 crc kubenswrapper[4813]: I1201 08:59:53.467813 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e7487c91-a8ac-4c0a-ac6e-6263890da1f2-config-data\") pod \"glance-db-sync-pkc4b\" (UID: \"e7487c91-a8ac-4c0a-ac6e-6263890da1f2\") " pod="glance-kuttl-tests/glance-db-sync-pkc4b" Dec 01 08:59:53 crc kubenswrapper[4813]: I1201 08:59:53.467841 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e7487c91-a8ac-4c0a-ac6e-6263890da1f2-combined-ca-bundle\") pod \"glance-db-sync-pkc4b\" (UID: \"e7487c91-a8ac-4c0a-ac6e-6263890da1f2\") " pod="glance-kuttl-tests/glance-db-sync-pkc4b" Dec 01 08:59:53 crc kubenswrapper[4813]: I1201 08:59:53.467858 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/e7487c91-a8ac-4c0a-ac6e-6263890da1f2-db-sync-config-data\") pod \"glance-db-sync-pkc4b\" (UID: \"e7487c91-a8ac-4c0a-ac6e-6263890da1f2\") " pod="glance-kuttl-tests/glance-db-sync-pkc4b" Dec 01 08:59:53 crc kubenswrapper[4813]: I1201 08:59:53.467936 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-sync-pkc4b"] Dec 01 08:59:53 crc kubenswrapper[4813]: I1201 08:59:53.568711 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e7487c91-a8ac-4c0a-ac6e-6263890da1f2-config-data\") pod \"glance-db-sync-pkc4b\" (UID: \"e7487c91-a8ac-4c0a-ac6e-6263890da1f2\") " pod="glance-kuttl-tests/glance-db-sync-pkc4b" Dec 01 08:59:53 crc kubenswrapper[4813]: I1201 08:59:53.568765 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e7487c91-a8ac-4c0a-ac6e-6263890da1f2-combined-ca-bundle\") pod \"glance-db-sync-pkc4b\" (UID: \"e7487c91-a8ac-4c0a-ac6e-6263890da1f2\") " pod="glance-kuttl-tests/glance-db-sync-pkc4b" Dec 01 08:59:53 crc kubenswrapper[4813]: I1201 08:59:53.568789 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/e7487c91-a8ac-4c0a-ac6e-6263890da1f2-db-sync-config-data\") pod \"glance-db-sync-pkc4b\" (UID: \"e7487c91-a8ac-4c0a-ac6e-6263890da1f2\") " pod="glance-kuttl-tests/glance-db-sync-pkc4b" Dec 01 08:59:53 crc kubenswrapper[4813]: I1201 08:59:53.568843 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8k8nv\" (UniqueName: \"kubernetes.io/projected/e7487c91-a8ac-4c0a-ac6e-6263890da1f2-kube-api-access-8k8nv\") pod \"glance-db-sync-pkc4b\" (UID: \"e7487c91-a8ac-4c0a-ac6e-6263890da1f2\") " pod="glance-kuttl-tests/glance-db-sync-pkc4b" Dec 01 08:59:53 crc kubenswrapper[4813]: I1201 08:59:53.572926 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e7487c91-a8ac-4c0a-ac6e-6263890da1f2-config-data\") pod \"glance-db-sync-pkc4b\" (UID: \"e7487c91-a8ac-4c0a-ac6e-6263890da1f2\") " pod="glance-kuttl-tests/glance-db-sync-pkc4b" Dec 01 08:59:53 crc kubenswrapper[4813]: I1201 08:59:53.572916 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/e7487c91-a8ac-4c0a-ac6e-6263890da1f2-db-sync-config-data\") pod \"glance-db-sync-pkc4b\" (UID: \"e7487c91-a8ac-4c0a-ac6e-6263890da1f2\") " pod="glance-kuttl-tests/glance-db-sync-pkc4b" Dec 01 08:59:53 crc kubenswrapper[4813]: I1201 08:59:53.578448 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e7487c91-a8ac-4c0a-ac6e-6263890da1f2-combined-ca-bundle\") pod \"glance-db-sync-pkc4b\" (UID: \"e7487c91-a8ac-4c0a-ac6e-6263890da1f2\") " pod="glance-kuttl-tests/glance-db-sync-pkc4b" Dec 01 08:59:53 crc kubenswrapper[4813]: I1201 08:59:53.591362 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8k8nv\" (UniqueName: \"kubernetes.io/projected/e7487c91-a8ac-4c0a-ac6e-6263890da1f2-kube-api-access-8k8nv\") pod \"glance-db-sync-pkc4b\" (UID: \"e7487c91-a8ac-4c0a-ac6e-6263890da1f2\") " pod="glance-kuttl-tests/glance-db-sync-pkc4b" Dec 01 08:59:53 crc kubenswrapper[4813]: I1201 08:59:53.811006 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-pkc4b" Dec 01 08:59:54 crc kubenswrapper[4813]: I1201 08:59:54.277652 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-sync-pkc4b"] Dec 01 08:59:55 crc kubenswrapper[4813]: I1201 08:59:55.157757 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-pkc4b" event={"ID":"e7487c91-a8ac-4c0a-ac6e-6263890da1f2","Type":"ContainerStarted","Data":"f85162b0555d9137364b7ea0770528a5d3d405ae6a72a4ed39d2524bc0b97042"} Dec 01 08:59:55 crc kubenswrapper[4813]: I1201 08:59:55.157824 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-pkc4b" event={"ID":"e7487c91-a8ac-4c0a-ac6e-6263890da1f2","Type":"ContainerStarted","Data":"9730bab7170acf2308d4537eb82747314e3cc284a92abc50e1454881d12ad4d6"} Dec 01 08:59:55 crc kubenswrapper[4813]: I1201 08:59:55.176053 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-db-sync-pkc4b" podStartSLOduration=2.176014291 podStartE2EDuration="2.176014291s" podCreationTimestamp="2025-12-01 08:59:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 08:59:55.17027519 +0000 UTC m=+1179.793196776" watchObservedRunningTime="2025-12-01 08:59:55.176014291 +0000 UTC m=+1179.798935877" Dec 01 08:59:58 crc kubenswrapper[4813]: I1201 08:59:58.189699 4813 generic.go:334] "Generic (PLEG): container finished" podID="e7487c91-a8ac-4c0a-ac6e-6263890da1f2" containerID="f85162b0555d9137364b7ea0770528a5d3d405ae6a72a4ed39d2524bc0b97042" exitCode=0 Dec 01 08:59:58 crc kubenswrapper[4813]: I1201 08:59:58.189762 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-pkc4b" event={"ID":"e7487c91-a8ac-4c0a-ac6e-6263890da1f2","Type":"ContainerDied","Data":"f85162b0555d9137364b7ea0770528a5d3d405ae6a72a4ed39d2524bc0b97042"} Dec 01 08:59:59 crc kubenswrapper[4813]: I1201 08:59:59.573943 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-pkc4b" Dec 01 08:59:59 crc kubenswrapper[4813]: I1201 08:59:59.664162 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/e7487c91-a8ac-4c0a-ac6e-6263890da1f2-db-sync-config-data\") pod \"e7487c91-a8ac-4c0a-ac6e-6263890da1f2\" (UID: \"e7487c91-a8ac-4c0a-ac6e-6263890da1f2\") " Dec 01 08:59:59 crc kubenswrapper[4813]: I1201 08:59:59.664220 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e7487c91-a8ac-4c0a-ac6e-6263890da1f2-combined-ca-bundle\") pod \"e7487c91-a8ac-4c0a-ac6e-6263890da1f2\" (UID: \"e7487c91-a8ac-4c0a-ac6e-6263890da1f2\") " Dec 01 08:59:59 crc kubenswrapper[4813]: I1201 08:59:59.664339 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8k8nv\" (UniqueName: \"kubernetes.io/projected/e7487c91-a8ac-4c0a-ac6e-6263890da1f2-kube-api-access-8k8nv\") pod \"e7487c91-a8ac-4c0a-ac6e-6263890da1f2\" (UID: \"e7487c91-a8ac-4c0a-ac6e-6263890da1f2\") " Dec 01 08:59:59 crc kubenswrapper[4813]: I1201 08:59:59.664371 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e7487c91-a8ac-4c0a-ac6e-6263890da1f2-config-data\") pod \"e7487c91-a8ac-4c0a-ac6e-6263890da1f2\" (UID: \"e7487c91-a8ac-4c0a-ac6e-6263890da1f2\") " Dec 01 08:59:59 crc kubenswrapper[4813]: I1201 08:59:59.672292 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7487c91-a8ac-4c0a-ac6e-6263890da1f2-kube-api-access-8k8nv" (OuterVolumeSpecName: "kube-api-access-8k8nv") pod "e7487c91-a8ac-4c0a-ac6e-6263890da1f2" (UID: "e7487c91-a8ac-4c0a-ac6e-6263890da1f2"). InnerVolumeSpecName "kube-api-access-8k8nv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 08:59:59 crc kubenswrapper[4813]: I1201 08:59:59.672326 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7487c91-a8ac-4c0a-ac6e-6263890da1f2-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "e7487c91-a8ac-4c0a-ac6e-6263890da1f2" (UID: "e7487c91-a8ac-4c0a-ac6e-6263890da1f2"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:59:59 crc kubenswrapper[4813]: I1201 08:59:59.694375 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7487c91-a8ac-4c0a-ac6e-6263890da1f2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e7487c91-a8ac-4c0a-ac6e-6263890da1f2" (UID: "e7487c91-a8ac-4c0a-ac6e-6263890da1f2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:59:59 crc kubenswrapper[4813]: I1201 08:59:59.717698 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7487c91-a8ac-4c0a-ac6e-6263890da1f2-config-data" (OuterVolumeSpecName: "config-data") pod "e7487c91-a8ac-4c0a-ac6e-6263890da1f2" (UID: "e7487c91-a8ac-4c0a-ac6e-6263890da1f2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 08:59:59 crc kubenswrapper[4813]: I1201 08:59:59.766431 4813 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/e7487c91-a8ac-4c0a-ac6e-6263890da1f2-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 08:59:59 crc kubenswrapper[4813]: I1201 08:59:59.766473 4813 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e7487c91-a8ac-4c0a-ac6e-6263890da1f2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 08:59:59 crc kubenswrapper[4813]: I1201 08:59:59.766486 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8k8nv\" (UniqueName: \"kubernetes.io/projected/e7487c91-a8ac-4c0a-ac6e-6263890da1f2-kube-api-access-8k8nv\") on node \"crc\" DevicePath \"\"" Dec 01 08:59:59 crc kubenswrapper[4813]: I1201 08:59:59.766500 4813 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e7487c91-a8ac-4c0a-ac6e-6263890da1f2-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 09:00:00 crc kubenswrapper[4813]: I1201 09:00:00.136453 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29409660-wk2cc"] Dec 01 09:00:00 crc kubenswrapper[4813]: E1201 09:00:00.136837 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e7487c91-a8ac-4c0a-ac6e-6263890da1f2" containerName="glance-db-sync" Dec 01 09:00:00 crc kubenswrapper[4813]: I1201 09:00:00.136865 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="e7487c91-a8ac-4c0a-ac6e-6263890da1f2" containerName="glance-db-sync" Dec 01 09:00:00 crc kubenswrapper[4813]: I1201 09:00:00.137074 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="e7487c91-a8ac-4c0a-ac6e-6263890da1f2" containerName="glance-db-sync" Dec 01 09:00:00 crc kubenswrapper[4813]: I1201 09:00:00.137728 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29409660-wk2cc" Dec 01 09:00:00 crc kubenswrapper[4813]: I1201 09:00:00.139851 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 01 09:00:00 crc kubenswrapper[4813]: I1201 09:00:00.140332 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 01 09:00:00 crc kubenswrapper[4813]: I1201 09:00:00.149512 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29409660-wk2cc"] Dec 01 09:00:00 crc kubenswrapper[4813]: I1201 09:00:00.173580 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b6af51c1-d4e6-4990-8aa1-b75c6bb4147e-secret-volume\") pod \"collect-profiles-29409660-wk2cc\" (UID: \"b6af51c1-d4e6-4990-8aa1-b75c6bb4147e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29409660-wk2cc" Dec 01 09:00:00 crc kubenswrapper[4813]: I1201 09:00:00.173965 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b6af51c1-d4e6-4990-8aa1-b75c6bb4147e-config-volume\") pod \"collect-profiles-29409660-wk2cc\" (UID: \"b6af51c1-d4e6-4990-8aa1-b75c6bb4147e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29409660-wk2cc" Dec 01 09:00:00 crc kubenswrapper[4813]: I1201 09:00:00.174147 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-79qd5\" (UniqueName: \"kubernetes.io/projected/b6af51c1-d4e6-4990-8aa1-b75c6bb4147e-kube-api-access-79qd5\") pod \"collect-profiles-29409660-wk2cc\" (UID: \"b6af51c1-d4e6-4990-8aa1-b75c6bb4147e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29409660-wk2cc" Dec 01 09:00:00 crc kubenswrapper[4813]: I1201 09:00:00.206728 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-pkc4b" event={"ID":"e7487c91-a8ac-4c0a-ac6e-6263890da1f2","Type":"ContainerDied","Data":"9730bab7170acf2308d4537eb82747314e3cc284a92abc50e1454881d12ad4d6"} Dec 01 09:00:00 crc kubenswrapper[4813]: I1201 09:00:00.206764 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-pkc4b" Dec 01 09:00:00 crc kubenswrapper[4813]: I1201 09:00:00.206776 4813 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9730bab7170acf2308d4537eb82747314e3cc284a92abc50e1454881d12ad4d6" Dec 01 09:00:00 crc kubenswrapper[4813]: I1201 09:00:00.275831 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-79qd5\" (UniqueName: \"kubernetes.io/projected/b6af51c1-d4e6-4990-8aa1-b75c6bb4147e-kube-api-access-79qd5\") pod \"collect-profiles-29409660-wk2cc\" (UID: \"b6af51c1-d4e6-4990-8aa1-b75c6bb4147e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29409660-wk2cc" Dec 01 09:00:00 crc kubenswrapper[4813]: I1201 09:00:00.275939 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b6af51c1-d4e6-4990-8aa1-b75c6bb4147e-secret-volume\") pod \"collect-profiles-29409660-wk2cc\" (UID: \"b6af51c1-d4e6-4990-8aa1-b75c6bb4147e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29409660-wk2cc" Dec 01 09:00:00 crc kubenswrapper[4813]: I1201 09:00:00.276044 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b6af51c1-d4e6-4990-8aa1-b75c6bb4147e-config-volume\") pod \"collect-profiles-29409660-wk2cc\" (UID: \"b6af51c1-d4e6-4990-8aa1-b75c6bb4147e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29409660-wk2cc" Dec 01 09:00:00 crc kubenswrapper[4813]: I1201 09:00:00.297943 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b6af51c1-d4e6-4990-8aa1-b75c6bb4147e-secret-volume\") pod \"collect-profiles-29409660-wk2cc\" (UID: \"b6af51c1-d4e6-4990-8aa1-b75c6bb4147e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29409660-wk2cc" Dec 01 09:00:00 crc kubenswrapper[4813]: I1201 09:00:00.301667 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b6af51c1-d4e6-4990-8aa1-b75c6bb4147e-config-volume\") pod \"collect-profiles-29409660-wk2cc\" (UID: \"b6af51c1-d4e6-4990-8aa1-b75c6bb4147e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29409660-wk2cc" Dec 01 09:00:00 crc kubenswrapper[4813]: I1201 09:00:00.307877 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-79qd5\" (UniqueName: \"kubernetes.io/projected/b6af51c1-d4e6-4990-8aa1-b75c6bb4147e-kube-api-access-79qd5\") pod \"collect-profiles-29409660-wk2cc\" (UID: \"b6af51c1-d4e6-4990-8aa1-b75c6bb4147e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29409660-wk2cc" Dec 01 09:00:00 crc kubenswrapper[4813]: I1201 09:00:00.457550 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29409660-wk2cc" Dec 01 09:00:00 crc kubenswrapper[4813]: I1201 09:00:00.988807 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29409660-wk2cc"] Dec 01 09:00:01 crc kubenswrapper[4813]: I1201 09:00:01.216469 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29409660-wk2cc" event={"ID":"b6af51c1-d4e6-4990-8aa1-b75c6bb4147e","Type":"ContainerStarted","Data":"00728d18cf3382077a7569f4e444fb56e355532c3e3ed604a6634a3c0cbc81a9"} Dec 01 09:00:01 crc kubenswrapper[4813]: I1201 09:00:01.216529 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29409660-wk2cc" event={"ID":"b6af51c1-d4e6-4990-8aa1-b75c6bb4147e","Type":"ContainerStarted","Data":"ced223d33f92d694c8922e442239f50db23503cd23fbedc60040883bffc59ac9"} Dec 01 09:00:01 crc kubenswrapper[4813]: I1201 09:00:01.259092 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29409660-wk2cc" podStartSLOduration=1.259073523 podStartE2EDuration="1.259073523s" podCreationTimestamp="2025-12-01 09:00:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 09:00:01.25573473 +0000 UTC m=+1185.878656316" watchObservedRunningTime="2025-12-01 09:00:01.259073523 +0000 UTC m=+1185.881995109" Dec 01 09:00:01 crc kubenswrapper[4813]: I1201 09:00:01.523911 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Dec 01 09:00:01 crc kubenswrapper[4813]: I1201 09:00:01.525033 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Dec 01 09:00:01 crc kubenswrapper[4813]: I1201 09:00:01.527043 4813 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"combined-ca-bundle" Dec 01 09:00:01 crc kubenswrapper[4813]: I1201 09:00:01.527448 4813 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-default-single-config-data" Dec 01 09:00:01 crc kubenswrapper[4813]: I1201 09:00:01.528304 4813 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"cert-glance-default-public-svc" Dec 01 09:00:01 crc kubenswrapper[4813]: I1201 09:00:01.528911 4813 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-glance-dockercfg-srxng" Dec 01 09:00:01 crc kubenswrapper[4813]: I1201 09:00:01.529272 4813 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-scripts" Dec 01 09:00:01 crc kubenswrapper[4813]: I1201 09:00:01.529291 4813 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"cert-glance-default-internal-svc" Dec 01 09:00:01 crc kubenswrapper[4813]: I1201 09:00:01.539960 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Dec 01 09:00:01 crc kubenswrapper[4813]: I1201 09:00:01.607908 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-single-0\" (UID: \"45132dc6-65ef-45bb-98a9-994e61a58f2e\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 01 09:00:01 crc kubenswrapper[4813]: I1201 09:00:01.608213 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/45132dc6-65ef-45bb-98a9-994e61a58f2e-public-tls-certs\") pod \"glance-default-single-0\" (UID: \"45132dc6-65ef-45bb-98a9-994e61a58f2e\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 01 09:00:01 crc kubenswrapper[4813]: I1201 09:00:01.608377 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/45132dc6-65ef-45bb-98a9-994e61a58f2e-config-data\") pod \"glance-default-single-0\" (UID: \"45132dc6-65ef-45bb-98a9-994e61a58f2e\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 01 09:00:01 crc kubenswrapper[4813]: I1201 09:00:01.608501 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/45132dc6-65ef-45bb-98a9-994e61a58f2e-httpd-run\") pod \"glance-default-single-0\" (UID: \"45132dc6-65ef-45bb-98a9-994e61a58f2e\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 01 09:00:01 crc kubenswrapper[4813]: I1201 09:00:01.608590 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/45132dc6-65ef-45bb-98a9-994e61a58f2e-scripts\") pod \"glance-default-single-0\" (UID: \"45132dc6-65ef-45bb-98a9-994e61a58f2e\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 01 09:00:01 crc kubenswrapper[4813]: I1201 09:00:01.608688 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/45132dc6-65ef-45bb-98a9-994e61a58f2e-logs\") pod \"glance-default-single-0\" (UID: \"45132dc6-65ef-45bb-98a9-994e61a58f2e\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 01 09:00:01 crc kubenswrapper[4813]: I1201 09:00:01.608779 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/45132dc6-65ef-45bb-98a9-994e61a58f2e-internal-tls-certs\") pod \"glance-default-single-0\" (UID: \"45132dc6-65ef-45bb-98a9-994e61a58f2e\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 01 09:00:01 crc kubenswrapper[4813]: I1201 09:00:01.608950 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/45132dc6-65ef-45bb-98a9-994e61a58f2e-combined-ca-bundle\") pod \"glance-default-single-0\" (UID: \"45132dc6-65ef-45bb-98a9-994e61a58f2e\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 01 09:00:01 crc kubenswrapper[4813]: I1201 09:00:01.609010 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pwvm9\" (UniqueName: \"kubernetes.io/projected/45132dc6-65ef-45bb-98a9-994e61a58f2e-kube-api-access-pwvm9\") pod \"glance-default-single-0\" (UID: \"45132dc6-65ef-45bb-98a9-994e61a58f2e\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 01 09:00:01 crc kubenswrapper[4813]: I1201 09:00:01.650781 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Dec 01 09:00:01 crc kubenswrapper[4813]: E1201 09:00:01.651368 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[combined-ca-bundle config-data glance httpd-run internal-tls-certs kube-api-access-pwvm9 logs public-tls-certs scripts], unattached volumes=[], failed to process volumes=[]: context canceled" pod="glance-kuttl-tests/glance-default-single-0" podUID="45132dc6-65ef-45bb-98a9-994e61a58f2e" Dec 01 09:00:01 crc kubenswrapper[4813]: I1201 09:00:01.709995 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/45132dc6-65ef-45bb-98a9-994e61a58f2e-config-data\") pod \"glance-default-single-0\" (UID: \"45132dc6-65ef-45bb-98a9-994e61a58f2e\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 01 09:00:01 crc kubenswrapper[4813]: I1201 09:00:01.710089 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/45132dc6-65ef-45bb-98a9-994e61a58f2e-httpd-run\") pod \"glance-default-single-0\" (UID: \"45132dc6-65ef-45bb-98a9-994e61a58f2e\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 01 09:00:01 crc kubenswrapper[4813]: I1201 09:00:01.710108 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/45132dc6-65ef-45bb-98a9-994e61a58f2e-scripts\") pod \"glance-default-single-0\" (UID: \"45132dc6-65ef-45bb-98a9-994e61a58f2e\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 01 09:00:01 crc kubenswrapper[4813]: I1201 09:00:01.710133 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/45132dc6-65ef-45bb-98a9-994e61a58f2e-logs\") pod \"glance-default-single-0\" (UID: \"45132dc6-65ef-45bb-98a9-994e61a58f2e\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 01 09:00:01 crc kubenswrapper[4813]: I1201 09:00:01.710156 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/45132dc6-65ef-45bb-98a9-994e61a58f2e-internal-tls-certs\") pod \"glance-default-single-0\" (UID: \"45132dc6-65ef-45bb-98a9-994e61a58f2e\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 01 09:00:01 crc kubenswrapper[4813]: I1201 09:00:01.710198 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/45132dc6-65ef-45bb-98a9-994e61a58f2e-combined-ca-bundle\") pod \"glance-default-single-0\" (UID: \"45132dc6-65ef-45bb-98a9-994e61a58f2e\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 01 09:00:01 crc kubenswrapper[4813]: I1201 09:00:01.710213 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pwvm9\" (UniqueName: \"kubernetes.io/projected/45132dc6-65ef-45bb-98a9-994e61a58f2e-kube-api-access-pwvm9\") pod \"glance-default-single-0\" (UID: \"45132dc6-65ef-45bb-98a9-994e61a58f2e\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 01 09:00:01 crc kubenswrapper[4813]: I1201 09:00:01.710235 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-single-0\" (UID: \"45132dc6-65ef-45bb-98a9-994e61a58f2e\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 01 09:00:01 crc kubenswrapper[4813]: I1201 09:00:01.710253 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/45132dc6-65ef-45bb-98a9-994e61a58f2e-public-tls-certs\") pod \"glance-default-single-0\" (UID: \"45132dc6-65ef-45bb-98a9-994e61a58f2e\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 01 09:00:01 crc kubenswrapper[4813]: I1201 09:00:01.711031 4813 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-single-0\" (UID: \"45132dc6-65ef-45bb-98a9-994e61a58f2e\") device mount path \"/mnt/openstack/pv10\"" pod="glance-kuttl-tests/glance-default-single-0" Dec 01 09:00:01 crc kubenswrapper[4813]: I1201 09:00:01.711410 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/45132dc6-65ef-45bb-98a9-994e61a58f2e-httpd-run\") pod \"glance-default-single-0\" (UID: \"45132dc6-65ef-45bb-98a9-994e61a58f2e\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 01 09:00:01 crc kubenswrapper[4813]: I1201 09:00:01.711413 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/45132dc6-65ef-45bb-98a9-994e61a58f2e-logs\") pod \"glance-default-single-0\" (UID: \"45132dc6-65ef-45bb-98a9-994e61a58f2e\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 01 09:00:01 crc kubenswrapper[4813]: I1201 09:00:01.717327 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/45132dc6-65ef-45bb-98a9-994e61a58f2e-combined-ca-bundle\") pod \"glance-default-single-0\" (UID: \"45132dc6-65ef-45bb-98a9-994e61a58f2e\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 01 09:00:01 crc kubenswrapper[4813]: I1201 09:00:01.717333 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/45132dc6-65ef-45bb-98a9-994e61a58f2e-scripts\") pod \"glance-default-single-0\" (UID: \"45132dc6-65ef-45bb-98a9-994e61a58f2e\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 01 09:00:01 crc kubenswrapper[4813]: I1201 09:00:01.717842 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/45132dc6-65ef-45bb-98a9-994e61a58f2e-internal-tls-certs\") pod \"glance-default-single-0\" (UID: \"45132dc6-65ef-45bb-98a9-994e61a58f2e\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 01 09:00:01 crc kubenswrapper[4813]: I1201 09:00:01.718170 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/45132dc6-65ef-45bb-98a9-994e61a58f2e-config-data\") pod \"glance-default-single-0\" (UID: \"45132dc6-65ef-45bb-98a9-994e61a58f2e\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 01 09:00:01 crc kubenswrapper[4813]: I1201 09:00:01.732754 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/45132dc6-65ef-45bb-98a9-994e61a58f2e-public-tls-certs\") pod \"glance-default-single-0\" (UID: \"45132dc6-65ef-45bb-98a9-994e61a58f2e\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 01 09:00:01 crc kubenswrapper[4813]: I1201 09:00:01.732766 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pwvm9\" (UniqueName: \"kubernetes.io/projected/45132dc6-65ef-45bb-98a9-994e61a58f2e-kube-api-access-pwvm9\") pod \"glance-default-single-0\" (UID: \"45132dc6-65ef-45bb-98a9-994e61a58f2e\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 01 09:00:01 crc kubenswrapper[4813]: I1201 09:00:01.738207 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-single-0\" (UID: \"45132dc6-65ef-45bb-98a9-994e61a58f2e\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 01 09:00:02 crc kubenswrapper[4813]: I1201 09:00:02.226710 4813 generic.go:334] "Generic (PLEG): container finished" podID="b6af51c1-d4e6-4990-8aa1-b75c6bb4147e" containerID="00728d18cf3382077a7569f4e444fb56e355532c3e3ed604a6634a3c0cbc81a9" exitCode=0 Dec 01 09:00:02 crc kubenswrapper[4813]: I1201 09:00:02.226800 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29409660-wk2cc" event={"ID":"b6af51c1-d4e6-4990-8aa1-b75c6bb4147e","Type":"ContainerDied","Data":"00728d18cf3382077a7569f4e444fb56e355532c3e3ed604a6634a3c0cbc81a9"} Dec 01 09:00:02 crc kubenswrapper[4813]: I1201 09:00:02.226863 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Dec 01 09:00:02 crc kubenswrapper[4813]: I1201 09:00:02.243771 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Dec 01 09:00:02 crc kubenswrapper[4813]: I1201 09:00:02.317578 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/45132dc6-65ef-45bb-98a9-994e61a58f2e-combined-ca-bundle\") pod \"45132dc6-65ef-45bb-98a9-994e61a58f2e\" (UID: \"45132dc6-65ef-45bb-98a9-994e61a58f2e\") " Dec 01 09:00:02 crc kubenswrapper[4813]: I1201 09:00:02.317653 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/45132dc6-65ef-45bb-98a9-994e61a58f2e-config-data\") pod \"45132dc6-65ef-45bb-98a9-994e61a58f2e\" (UID: \"45132dc6-65ef-45bb-98a9-994e61a58f2e\") " Dec 01 09:00:02 crc kubenswrapper[4813]: I1201 09:00:02.317710 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/45132dc6-65ef-45bb-98a9-994e61a58f2e-public-tls-certs\") pod \"45132dc6-65ef-45bb-98a9-994e61a58f2e\" (UID: \"45132dc6-65ef-45bb-98a9-994e61a58f2e\") " Dec 01 09:00:02 crc kubenswrapper[4813]: I1201 09:00:02.317743 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"45132dc6-65ef-45bb-98a9-994e61a58f2e\" (UID: \"45132dc6-65ef-45bb-98a9-994e61a58f2e\") " Dec 01 09:00:02 crc kubenswrapper[4813]: I1201 09:00:02.317776 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pwvm9\" (UniqueName: \"kubernetes.io/projected/45132dc6-65ef-45bb-98a9-994e61a58f2e-kube-api-access-pwvm9\") pod \"45132dc6-65ef-45bb-98a9-994e61a58f2e\" (UID: \"45132dc6-65ef-45bb-98a9-994e61a58f2e\") " Dec 01 09:00:02 crc kubenswrapper[4813]: I1201 09:00:02.317851 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/45132dc6-65ef-45bb-98a9-994e61a58f2e-httpd-run\") pod \"45132dc6-65ef-45bb-98a9-994e61a58f2e\" (UID: \"45132dc6-65ef-45bb-98a9-994e61a58f2e\") " Dec 01 09:00:02 crc kubenswrapper[4813]: I1201 09:00:02.317894 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/45132dc6-65ef-45bb-98a9-994e61a58f2e-internal-tls-certs\") pod \"45132dc6-65ef-45bb-98a9-994e61a58f2e\" (UID: \"45132dc6-65ef-45bb-98a9-994e61a58f2e\") " Dec 01 09:00:02 crc kubenswrapper[4813]: I1201 09:00:02.317993 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/45132dc6-65ef-45bb-98a9-994e61a58f2e-scripts\") pod \"45132dc6-65ef-45bb-98a9-994e61a58f2e\" (UID: \"45132dc6-65ef-45bb-98a9-994e61a58f2e\") " Dec 01 09:00:02 crc kubenswrapper[4813]: I1201 09:00:02.318059 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/45132dc6-65ef-45bb-98a9-994e61a58f2e-logs\") pod \"45132dc6-65ef-45bb-98a9-994e61a58f2e\" (UID: \"45132dc6-65ef-45bb-98a9-994e61a58f2e\") " Dec 01 09:00:02 crc kubenswrapper[4813]: I1201 09:00:02.318658 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/45132dc6-65ef-45bb-98a9-994e61a58f2e-logs" (OuterVolumeSpecName: "logs") pod "45132dc6-65ef-45bb-98a9-994e61a58f2e" (UID: "45132dc6-65ef-45bb-98a9-994e61a58f2e"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 09:00:02 crc kubenswrapper[4813]: I1201 09:00:02.318896 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/45132dc6-65ef-45bb-98a9-994e61a58f2e-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "45132dc6-65ef-45bb-98a9-994e61a58f2e" (UID: "45132dc6-65ef-45bb-98a9-994e61a58f2e"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 09:00:02 crc kubenswrapper[4813]: I1201 09:00:02.322253 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage10-crc" (OuterVolumeSpecName: "glance") pod "45132dc6-65ef-45bb-98a9-994e61a58f2e" (UID: "45132dc6-65ef-45bb-98a9-994e61a58f2e"). InnerVolumeSpecName "local-storage10-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 01 09:00:02 crc kubenswrapper[4813]: I1201 09:00:02.322845 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/45132dc6-65ef-45bb-98a9-994e61a58f2e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "45132dc6-65ef-45bb-98a9-994e61a58f2e" (UID: "45132dc6-65ef-45bb-98a9-994e61a58f2e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:00:02 crc kubenswrapper[4813]: I1201 09:00:02.323070 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/45132dc6-65ef-45bb-98a9-994e61a58f2e-kube-api-access-pwvm9" (OuterVolumeSpecName: "kube-api-access-pwvm9") pod "45132dc6-65ef-45bb-98a9-994e61a58f2e" (UID: "45132dc6-65ef-45bb-98a9-994e61a58f2e"). InnerVolumeSpecName "kube-api-access-pwvm9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:00:02 crc kubenswrapper[4813]: I1201 09:00:02.323374 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/45132dc6-65ef-45bb-98a9-994e61a58f2e-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "45132dc6-65ef-45bb-98a9-994e61a58f2e" (UID: "45132dc6-65ef-45bb-98a9-994e61a58f2e"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:00:02 crc kubenswrapper[4813]: I1201 09:00:02.324701 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/45132dc6-65ef-45bb-98a9-994e61a58f2e-scripts" (OuterVolumeSpecName: "scripts") pod "45132dc6-65ef-45bb-98a9-994e61a58f2e" (UID: "45132dc6-65ef-45bb-98a9-994e61a58f2e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:00:02 crc kubenswrapper[4813]: I1201 09:00:02.332115 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/45132dc6-65ef-45bb-98a9-994e61a58f2e-config-data" (OuterVolumeSpecName: "config-data") pod "45132dc6-65ef-45bb-98a9-994e61a58f2e" (UID: "45132dc6-65ef-45bb-98a9-994e61a58f2e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:00:02 crc kubenswrapper[4813]: I1201 09:00:02.332149 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/45132dc6-65ef-45bb-98a9-994e61a58f2e-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "45132dc6-65ef-45bb-98a9-994e61a58f2e" (UID: "45132dc6-65ef-45bb-98a9-994e61a58f2e"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:00:02 crc kubenswrapper[4813]: I1201 09:00:02.419684 4813 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/45132dc6-65ef-45bb-98a9-994e61a58f2e-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 01 09:00:02 crc kubenswrapper[4813]: I1201 09:00:02.419725 4813 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/45132dc6-65ef-45bb-98a9-994e61a58f2e-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 01 09:00:02 crc kubenswrapper[4813]: I1201 09:00:02.419737 4813 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/45132dc6-65ef-45bb-98a9-994e61a58f2e-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 09:00:02 crc kubenswrapper[4813]: I1201 09:00:02.419748 4813 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/45132dc6-65ef-45bb-98a9-994e61a58f2e-logs\") on node \"crc\" DevicePath \"\"" Dec 01 09:00:02 crc kubenswrapper[4813]: I1201 09:00:02.419761 4813 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/45132dc6-65ef-45bb-98a9-994e61a58f2e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 09:00:02 crc kubenswrapper[4813]: I1201 09:00:02.419772 4813 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/45132dc6-65ef-45bb-98a9-994e61a58f2e-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 09:00:02 crc kubenswrapper[4813]: I1201 09:00:02.419783 4813 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/45132dc6-65ef-45bb-98a9-994e61a58f2e-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 01 09:00:02 crc kubenswrapper[4813]: I1201 09:00:02.419824 4813 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" " Dec 01 09:00:02 crc kubenswrapper[4813]: I1201 09:00:02.419838 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pwvm9\" (UniqueName: \"kubernetes.io/projected/45132dc6-65ef-45bb-98a9-994e61a58f2e-kube-api-access-pwvm9\") on node \"crc\" DevicePath \"\"" Dec 01 09:00:02 crc kubenswrapper[4813]: I1201 09:00:02.433188 4813 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage10-crc" (UniqueName: "kubernetes.io/local-volume/local-storage10-crc") on node "crc" Dec 01 09:00:02 crc kubenswrapper[4813]: I1201 09:00:02.520937 4813 reconciler_common.go:293] "Volume detached for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" DevicePath \"\"" Dec 01 09:00:03 crc kubenswrapper[4813]: I1201 09:00:03.234927 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Dec 01 09:00:03 crc kubenswrapper[4813]: I1201 09:00:03.330431 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Dec 01 09:00:03 crc kubenswrapper[4813]: I1201 09:00:03.340377 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Dec 01 09:00:03 crc kubenswrapper[4813]: I1201 09:00:03.358278 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Dec 01 09:00:03 crc kubenswrapper[4813]: I1201 09:00:03.359623 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Dec 01 09:00:03 crc kubenswrapper[4813]: I1201 09:00:03.363792 4813 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-scripts" Dec 01 09:00:03 crc kubenswrapper[4813]: I1201 09:00:03.364030 4813 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"combined-ca-bundle" Dec 01 09:00:03 crc kubenswrapper[4813]: I1201 09:00:03.364097 4813 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"cert-glance-default-internal-svc" Dec 01 09:00:03 crc kubenswrapper[4813]: I1201 09:00:03.364227 4813 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-default-single-config-data" Dec 01 09:00:03 crc kubenswrapper[4813]: I1201 09:00:03.364442 4813 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-glance-dockercfg-srxng" Dec 01 09:00:03 crc kubenswrapper[4813]: I1201 09:00:03.364613 4813 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"cert-glance-default-public-svc" Dec 01 09:00:03 crc kubenswrapper[4813]: I1201 09:00:03.369946 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Dec 01 09:00:03 crc kubenswrapper[4813]: I1201 09:00:03.435754 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ca718568-e6fc-4cce-abe3-41037ed450f9-internal-tls-certs\") pod \"glance-default-single-0\" (UID: \"ca718568-e6fc-4cce-abe3-41037ed450f9\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 01 09:00:03 crc kubenswrapper[4813]: I1201 09:00:03.435801 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ca718568-e6fc-4cce-abe3-41037ed450f9-scripts\") pod \"glance-default-single-0\" (UID: \"ca718568-e6fc-4cce-abe3-41037ed450f9\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 01 09:00:03 crc kubenswrapper[4813]: I1201 09:00:03.435825 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/ca718568-e6fc-4cce-abe3-41037ed450f9-httpd-run\") pod \"glance-default-single-0\" (UID: \"ca718568-e6fc-4cce-abe3-41037ed450f9\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 01 09:00:03 crc kubenswrapper[4813]: I1201 09:00:03.435877 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca718568-e6fc-4cce-abe3-41037ed450f9-combined-ca-bundle\") pod \"glance-default-single-0\" (UID: \"ca718568-e6fc-4cce-abe3-41037ed450f9\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 01 09:00:03 crc kubenswrapper[4813]: I1201 09:00:03.435899 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ca718568-e6fc-4cce-abe3-41037ed450f9-logs\") pod \"glance-default-single-0\" (UID: \"ca718568-e6fc-4cce-abe3-41037ed450f9\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 01 09:00:03 crc kubenswrapper[4813]: I1201 09:00:03.435939 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ca718568-e6fc-4cce-abe3-41037ed450f9-config-data\") pod \"glance-default-single-0\" (UID: \"ca718568-e6fc-4cce-abe3-41037ed450f9\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 01 09:00:03 crc kubenswrapper[4813]: I1201 09:00:03.435964 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ca718568-e6fc-4cce-abe3-41037ed450f9-public-tls-certs\") pod \"glance-default-single-0\" (UID: \"ca718568-e6fc-4cce-abe3-41037ed450f9\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 01 09:00:03 crc kubenswrapper[4813]: I1201 09:00:03.436058 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-single-0\" (UID: \"ca718568-e6fc-4cce-abe3-41037ed450f9\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 01 09:00:03 crc kubenswrapper[4813]: I1201 09:00:03.436077 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-27f52\" (UniqueName: \"kubernetes.io/projected/ca718568-e6fc-4cce-abe3-41037ed450f9-kube-api-access-27f52\") pod \"glance-default-single-0\" (UID: \"ca718568-e6fc-4cce-abe3-41037ed450f9\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 01 09:00:03 crc kubenswrapper[4813]: I1201 09:00:03.538083 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ca718568-e6fc-4cce-abe3-41037ed450f9-internal-tls-certs\") pod \"glance-default-single-0\" (UID: \"ca718568-e6fc-4cce-abe3-41037ed450f9\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 01 09:00:03 crc kubenswrapper[4813]: I1201 09:00:03.538154 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ca718568-e6fc-4cce-abe3-41037ed450f9-scripts\") pod \"glance-default-single-0\" (UID: \"ca718568-e6fc-4cce-abe3-41037ed450f9\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 01 09:00:03 crc kubenswrapper[4813]: I1201 09:00:03.538187 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/ca718568-e6fc-4cce-abe3-41037ed450f9-httpd-run\") pod \"glance-default-single-0\" (UID: \"ca718568-e6fc-4cce-abe3-41037ed450f9\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 01 09:00:03 crc kubenswrapper[4813]: I1201 09:00:03.538217 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca718568-e6fc-4cce-abe3-41037ed450f9-combined-ca-bundle\") pod \"glance-default-single-0\" (UID: \"ca718568-e6fc-4cce-abe3-41037ed450f9\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 01 09:00:03 crc kubenswrapper[4813]: I1201 09:00:03.538239 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ca718568-e6fc-4cce-abe3-41037ed450f9-logs\") pod \"glance-default-single-0\" (UID: \"ca718568-e6fc-4cce-abe3-41037ed450f9\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 01 09:00:03 crc kubenswrapper[4813]: I1201 09:00:03.538271 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ca718568-e6fc-4cce-abe3-41037ed450f9-config-data\") pod \"glance-default-single-0\" (UID: \"ca718568-e6fc-4cce-abe3-41037ed450f9\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 01 09:00:03 crc kubenswrapper[4813]: I1201 09:00:03.538307 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ca718568-e6fc-4cce-abe3-41037ed450f9-public-tls-certs\") pod \"glance-default-single-0\" (UID: \"ca718568-e6fc-4cce-abe3-41037ed450f9\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 01 09:00:03 crc kubenswrapper[4813]: I1201 09:00:03.538329 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-single-0\" (UID: \"ca718568-e6fc-4cce-abe3-41037ed450f9\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 01 09:00:03 crc kubenswrapper[4813]: I1201 09:00:03.538352 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-27f52\" (UniqueName: \"kubernetes.io/projected/ca718568-e6fc-4cce-abe3-41037ed450f9-kube-api-access-27f52\") pod \"glance-default-single-0\" (UID: \"ca718568-e6fc-4cce-abe3-41037ed450f9\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 01 09:00:03 crc kubenswrapper[4813]: I1201 09:00:03.539855 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ca718568-e6fc-4cce-abe3-41037ed450f9-logs\") pod \"glance-default-single-0\" (UID: \"ca718568-e6fc-4cce-abe3-41037ed450f9\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 01 09:00:03 crc kubenswrapper[4813]: I1201 09:00:03.539864 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/ca718568-e6fc-4cce-abe3-41037ed450f9-httpd-run\") pod \"glance-default-single-0\" (UID: \"ca718568-e6fc-4cce-abe3-41037ed450f9\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 01 09:00:03 crc kubenswrapper[4813]: I1201 09:00:03.540280 4813 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-single-0\" (UID: \"ca718568-e6fc-4cce-abe3-41037ed450f9\") device mount path \"/mnt/openstack/pv10\"" pod="glance-kuttl-tests/glance-default-single-0" Dec 01 09:00:03 crc kubenswrapper[4813]: I1201 09:00:03.543395 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ca718568-e6fc-4cce-abe3-41037ed450f9-scripts\") pod \"glance-default-single-0\" (UID: \"ca718568-e6fc-4cce-abe3-41037ed450f9\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 01 09:00:03 crc kubenswrapper[4813]: I1201 09:00:03.544529 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ca718568-e6fc-4cce-abe3-41037ed450f9-public-tls-certs\") pod \"glance-default-single-0\" (UID: \"ca718568-e6fc-4cce-abe3-41037ed450f9\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 01 09:00:03 crc kubenswrapper[4813]: I1201 09:00:03.548716 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca718568-e6fc-4cce-abe3-41037ed450f9-combined-ca-bundle\") pod \"glance-default-single-0\" (UID: \"ca718568-e6fc-4cce-abe3-41037ed450f9\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 01 09:00:03 crc kubenswrapper[4813]: I1201 09:00:03.552055 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ca718568-e6fc-4cce-abe3-41037ed450f9-config-data\") pod \"glance-default-single-0\" (UID: \"ca718568-e6fc-4cce-abe3-41037ed450f9\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 01 09:00:03 crc kubenswrapper[4813]: I1201 09:00:03.558017 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-27f52\" (UniqueName: \"kubernetes.io/projected/ca718568-e6fc-4cce-abe3-41037ed450f9-kube-api-access-27f52\") pod \"glance-default-single-0\" (UID: \"ca718568-e6fc-4cce-abe3-41037ed450f9\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 01 09:00:03 crc kubenswrapper[4813]: I1201 09:00:03.564250 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ca718568-e6fc-4cce-abe3-41037ed450f9-internal-tls-certs\") pod \"glance-default-single-0\" (UID: \"ca718568-e6fc-4cce-abe3-41037ed450f9\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 01 09:00:03 crc kubenswrapper[4813]: I1201 09:00:03.579954 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-single-0\" (UID: \"ca718568-e6fc-4cce-abe3-41037ed450f9\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 01 09:00:03 crc kubenswrapper[4813]: I1201 09:00:03.622340 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29409660-wk2cc" Dec 01 09:00:03 crc kubenswrapper[4813]: I1201 09:00:03.690358 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Dec 01 09:00:03 crc kubenswrapper[4813]: I1201 09:00:03.741486 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b6af51c1-d4e6-4990-8aa1-b75c6bb4147e-secret-volume\") pod \"b6af51c1-d4e6-4990-8aa1-b75c6bb4147e\" (UID: \"b6af51c1-d4e6-4990-8aa1-b75c6bb4147e\") " Dec 01 09:00:03 crc kubenswrapper[4813]: I1201 09:00:03.741544 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-79qd5\" (UniqueName: \"kubernetes.io/projected/b6af51c1-d4e6-4990-8aa1-b75c6bb4147e-kube-api-access-79qd5\") pod \"b6af51c1-d4e6-4990-8aa1-b75c6bb4147e\" (UID: \"b6af51c1-d4e6-4990-8aa1-b75c6bb4147e\") " Dec 01 09:00:03 crc kubenswrapper[4813]: I1201 09:00:03.741609 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b6af51c1-d4e6-4990-8aa1-b75c6bb4147e-config-volume\") pod \"b6af51c1-d4e6-4990-8aa1-b75c6bb4147e\" (UID: \"b6af51c1-d4e6-4990-8aa1-b75c6bb4147e\") " Dec 01 09:00:03 crc kubenswrapper[4813]: I1201 09:00:03.742822 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6af51c1-d4e6-4990-8aa1-b75c6bb4147e-config-volume" (OuterVolumeSpecName: "config-volume") pod "b6af51c1-d4e6-4990-8aa1-b75c6bb4147e" (UID: "b6af51c1-d4e6-4990-8aa1-b75c6bb4147e"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 09:00:03 crc kubenswrapper[4813]: I1201 09:00:03.745721 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6af51c1-d4e6-4990-8aa1-b75c6bb4147e-kube-api-access-79qd5" (OuterVolumeSpecName: "kube-api-access-79qd5") pod "b6af51c1-d4e6-4990-8aa1-b75c6bb4147e" (UID: "b6af51c1-d4e6-4990-8aa1-b75c6bb4147e"). InnerVolumeSpecName "kube-api-access-79qd5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:00:03 crc kubenswrapper[4813]: I1201 09:00:03.746272 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6af51c1-d4e6-4990-8aa1-b75c6bb4147e-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "b6af51c1-d4e6-4990-8aa1-b75c6bb4147e" (UID: "b6af51c1-d4e6-4990-8aa1-b75c6bb4147e"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:00:03 crc kubenswrapper[4813]: I1201 09:00:03.843278 4813 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b6af51c1-d4e6-4990-8aa1-b75c6bb4147e-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 01 09:00:03 crc kubenswrapper[4813]: I1201 09:00:03.843325 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-79qd5\" (UniqueName: \"kubernetes.io/projected/b6af51c1-d4e6-4990-8aa1-b75c6bb4147e-kube-api-access-79qd5\") on node \"crc\" DevicePath \"\"" Dec 01 09:00:03 crc kubenswrapper[4813]: I1201 09:00:03.843336 4813 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b6af51c1-d4e6-4990-8aa1-b75c6bb4147e-config-volume\") on node \"crc\" DevicePath \"\"" Dec 01 09:00:04 crc kubenswrapper[4813]: I1201 09:00:04.115529 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Dec 01 09:00:04 crc kubenswrapper[4813]: I1201 09:00:04.248495 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"ca718568-e6fc-4cce-abe3-41037ed450f9","Type":"ContainerStarted","Data":"34efcfd40ea0a1695dadf0619af499499dd3035c87ef04d76dfdacf44ee9ac0e"} Dec 01 09:00:04 crc kubenswrapper[4813]: I1201 09:00:04.250705 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29409660-wk2cc" event={"ID":"b6af51c1-d4e6-4990-8aa1-b75c6bb4147e","Type":"ContainerDied","Data":"ced223d33f92d694c8922e442239f50db23503cd23fbedc60040883bffc59ac9"} Dec 01 09:00:04 crc kubenswrapper[4813]: I1201 09:00:04.250783 4813 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ced223d33f92d694c8922e442239f50db23503cd23fbedc60040883bffc59ac9" Dec 01 09:00:04 crc kubenswrapper[4813]: I1201 09:00:04.250751 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29409660-wk2cc" Dec 01 09:00:04 crc kubenswrapper[4813]: I1201 09:00:04.404142 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="45132dc6-65ef-45bb-98a9-994e61a58f2e" path="/var/lib/kubelet/pods/45132dc6-65ef-45bb-98a9-994e61a58f2e/volumes" Dec 01 09:00:05 crc kubenswrapper[4813]: I1201 09:00:05.260088 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"ca718568-e6fc-4cce-abe3-41037ed450f9","Type":"ContainerStarted","Data":"f56c5aedb6985186051d283a5dc19ee9555e52778197169b3869af8208485568"} Dec 01 09:00:05 crc kubenswrapper[4813]: I1201 09:00:05.260489 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"ca718568-e6fc-4cce-abe3-41037ed450f9","Type":"ContainerStarted","Data":"9f603c3520d243fd59fab8149ebff6dc69c7ef48950a8cc439f23f6c7592d59a"} Dec 01 09:00:05 crc kubenswrapper[4813]: I1201 09:00:05.283588 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-default-single-0" podStartSLOduration=2.283560705 podStartE2EDuration="2.283560705s" podCreationTimestamp="2025-12-01 09:00:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 09:00:05.27805363 +0000 UTC m=+1189.900975256" watchObservedRunningTime="2025-12-01 09:00:05.283560705 +0000 UTC m=+1189.906482311" Dec 01 09:00:13 crc kubenswrapper[4813]: I1201 09:00:13.690783 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-single-0" Dec 01 09:00:13 crc kubenswrapper[4813]: I1201 09:00:13.691419 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-single-0" Dec 01 09:00:13 crc kubenswrapper[4813]: I1201 09:00:13.720228 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-single-0" Dec 01 09:00:13 crc kubenswrapper[4813]: I1201 09:00:13.761148 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-single-0" Dec 01 09:00:14 crc kubenswrapper[4813]: I1201 09:00:14.355487 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-single-0" Dec 01 09:00:14 crc kubenswrapper[4813]: I1201 09:00:14.355559 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-single-0" Dec 01 09:00:16 crc kubenswrapper[4813]: I1201 09:00:16.356885 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-single-0" Dec 01 09:00:16 crc kubenswrapper[4813]: I1201 09:00:16.379586 4813 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 01 09:00:16 crc kubenswrapper[4813]: I1201 09:00:16.471617 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-single-0" Dec 01 09:00:18 crc kubenswrapper[4813]: I1201 09:00:18.132999 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-db-sync-pkc4b"] Dec 01 09:00:18 crc kubenswrapper[4813]: I1201 09:00:18.139057 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-db-sync-pkc4b"] Dec 01 09:00:18 crc kubenswrapper[4813]: I1201 09:00:18.202360 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glanceb0ca-account-delete-mrcsf"] Dec 01 09:00:18 crc kubenswrapper[4813]: E1201 09:00:18.202643 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b6af51c1-d4e6-4990-8aa1-b75c6bb4147e" containerName="collect-profiles" Dec 01 09:00:18 crc kubenswrapper[4813]: I1201 09:00:18.202659 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="b6af51c1-d4e6-4990-8aa1-b75c6bb4147e" containerName="collect-profiles" Dec 01 09:00:18 crc kubenswrapper[4813]: I1201 09:00:18.202784 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="b6af51c1-d4e6-4990-8aa1-b75c6bb4147e" containerName="collect-profiles" Dec 01 09:00:18 crc kubenswrapper[4813]: I1201 09:00:18.203430 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glanceb0ca-account-delete-mrcsf" Dec 01 09:00:18 crc kubenswrapper[4813]: I1201 09:00:18.213299 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glanceb0ca-account-delete-mrcsf"] Dec 01 09:00:18 crc kubenswrapper[4813]: I1201 09:00:18.275024 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Dec 01 09:00:18 crc kubenswrapper[4813]: I1201 09:00:18.288475 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r8xj7\" (UniqueName: \"kubernetes.io/projected/9ac262ec-4f70-4f59-b396-a2ca1c994007-kube-api-access-r8xj7\") pod \"glanceb0ca-account-delete-mrcsf\" (UID: \"9ac262ec-4f70-4f59-b396-a2ca1c994007\") " pod="glance-kuttl-tests/glanceb0ca-account-delete-mrcsf" Dec 01 09:00:18 crc kubenswrapper[4813]: I1201 09:00:18.288581 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9ac262ec-4f70-4f59-b396-a2ca1c994007-operator-scripts\") pod \"glanceb0ca-account-delete-mrcsf\" (UID: \"9ac262ec-4f70-4f59-b396-a2ca1c994007\") " pod="glance-kuttl-tests/glanceb0ca-account-delete-mrcsf" Dec 01 09:00:18 crc kubenswrapper[4813]: I1201 09:00:18.389469 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r8xj7\" (UniqueName: \"kubernetes.io/projected/9ac262ec-4f70-4f59-b396-a2ca1c994007-kube-api-access-r8xj7\") pod \"glanceb0ca-account-delete-mrcsf\" (UID: \"9ac262ec-4f70-4f59-b396-a2ca1c994007\") " pod="glance-kuttl-tests/glanceb0ca-account-delete-mrcsf" Dec 01 09:00:18 crc kubenswrapper[4813]: I1201 09:00:18.389543 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9ac262ec-4f70-4f59-b396-a2ca1c994007-operator-scripts\") pod \"glanceb0ca-account-delete-mrcsf\" (UID: \"9ac262ec-4f70-4f59-b396-a2ca1c994007\") " pod="glance-kuttl-tests/glanceb0ca-account-delete-mrcsf" Dec 01 09:00:18 crc kubenswrapper[4813]: I1201 09:00:18.390255 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9ac262ec-4f70-4f59-b396-a2ca1c994007-operator-scripts\") pod \"glanceb0ca-account-delete-mrcsf\" (UID: \"9ac262ec-4f70-4f59-b396-a2ca1c994007\") " pod="glance-kuttl-tests/glanceb0ca-account-delete-mrcsf" Dec 01 09:00:18 crc kubenswrapper[4813]: I1201 09:00:18.396171 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-single-0" podUID="ca718568-e6fc-4cce-abe3-41037ed450f9" containerName="glance-log" containerID="cri-o://9f603c3520d243fd59fab8149ebff6dc69c7ef48950a8cc439f23f6c7592d59a" gracePeriod=30 Dec 01 09:00:18 crc kubenswrapper[4813]: I1201 09:00:18.396627 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-single-0" podUID="ca718568-e6fc-4cce-abe3-41037ed450f9" containerName="glance-httpd" containerID="cri-o://f56c5aedb6985186051d283a5dc19ee9555e52778197169b3869af8208485568" gracePeriod=30 Dec 01 09:00:18 crc kubenswrapper[4813]: I1201 09:00:18.406913 4813 prober.go:107] "Probe failed" probeType="Readiness" pod="glance-kuttl-tests/glance-default-single-0" podUID="ca718568-e6fc-4cce-abe3-41037ed450f9" containerName="glance-log" probeResult="failure" output="Get \"https://10.217.0.108:9292/healthcheck\": EOF" Dec 01 09:00:18 crc kubenswrapper[4813]: I1201 09:00:18.407979 4813 prober.go:107] "Probe failed" probeType="Readiness" pod="glance-kuttl-tests/glance-default-single-0" podUID="ca718568-e6fc-4cce-abe3-41037ed450f9" containerName="glance-httpd" probeResult="failure" output="Get \"https://10.217.0.108:9292/healthcheck\": EOF" Dec 01 09:00:18 crc kubenswrapper[4813]: I1201 09:00:18.409661 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7487c91-a8ac-4c0a-ac6e-6263890da1f2" path="/var/lib/kubelet/pods/e7487c91-a8ac-4c0a-ac6e-6263890da1f2/volumes" Dec 01 09:00:18 crc kubenswrapper[4813]: I1201 09:00:18.413129 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r8xj7\" (UniqueName: \"kubernetes.io/projected/9ac262ec-4f70-4f59-b396-a2ca1c994007-kube-api-access-r8xj7\") pod \"glanceb0ca-account-delete-mrcsf\" (UID: \"9ac262ec-4f70-4f59-b396-a2ca1c994007\") " pod="glance-kuttl-tests/glanceb0ca-account-delete-mrcsf" Dec 01 09:00:18 crc kubenswrapper[4813]: I1201 09:00:18.517568 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glanceb0ca-account-delete-mrcsf" Dec 01 09:00:18 crc kubenswrapper[4813]: I1201 09:00:18.933164 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glanceb0ca-account-delete-mrcsf"] Dec 01 09:00:18 crc kubenswrapper[4813]: W1201 09:00:18.939141 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9ac262ec_4f70_4f59_b396_a2ca1c994007.slice/crio-f05187ed7c941266b337778eaf15a9aa042c9ddfd392c9c787109d581c3640d8 WatchSource:0}: Error finding container f05187ed7c941266b337778eaf15a9aa042c9ddfd392c9c787109d581c3640d8: Status 404 returned error can't find the container with id f05187ed7c941266b337778eaf15a9aa042c9ddfd392c9c787109d581c3640d8 Dec 01 09:00:19 crc kubenswrapper[4813]: I1201 09:00:19.423038 4813 generic.go:334] "Generic (PLEG): container finished" podID="9ac262ec-4f70-4f59-b396-a2ca1c994007" containerID="6e67bbc17ad4e709ff5bed996102f6ebc16522e36b77fd9d2ecf70d4d72df3a6" exitCode=0 Dec 01 09:00:19 crc kubenswrapper[4813]: I1201 09:00:19.423164 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glanceb0ca-account-delete-mrcsf" event={"ID":"9ac262ec-4f70-4f59-b396-a2ca1c994007","Type":"ContainerDied","Data":"6e67bbc17ad4e709ff5bed996102f6ebc16522e36b77fd9d2ecf70d4d72df3a6"} Dec 01 09:00:19 crc kubenswrapper[4813]: I1201 09:00:19.423258 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glanceb0ca-account-delete-mrcsf" event={"ID":"9ac262ec-4f70-4f59-b396-a2ca1c994007","Type":"ContainerStarted","Data":"f05187ed7c941266b337778eaf15a9aa042c9ddfd392c9c787109d581c3640d8"} Dec 01 09:00:19 crc kubenswrapper[4813]: I1201 09:00:19.424816 4813 generic.go:334] "Generic (PLEG): container finished" podID="ca718568-e6fc-4cce-abe3-41037ed450f9" containerID="9f603c3520d243fd59fab8149ebff6dc69c7ef48950a8cc439f23f6c7592d59a" exitCode=143 Dec 01 09:00:19 crc kubenswrapper[4813]: I1201 09:00:19.424847 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"ca718568-e6fc-4cce-abe3-41037ed450f9","Type":"ContainerDied","Data":"9f603c3520d243fd59fab8149ebff6dc69c7ef48950a8cc439f23f6c7592d59a"} Dec 01 09:00:20 crc kubenswrapper[4813]: I1201 09:00:20.771841 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glanceb0ca-account-delete-mrcsf" Dec 01 09:00:20 crc kubenswrapper[4813]: I1201 09:00:20.935613 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9ac262ec-4f70-4f59-b396-a2ca1c994007-operator-scripts\") pod \"9ac262ec-4f70-4f59-b396-a2ca1c994007\" (UID: \"9ac262ec-4f70-4f59-b396-a2ca1c994007\") " Dec 01 09:00:20 crc kubenswrapper[4813]: I1201 09:00:20.936064 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r8xj7\" (UniqueName: \"kubernetes.io/projected/9ac262ec-4f70-4f59-b396-a2ca1c994007-kube-api-access-r8xj7\") pod \"9ac262ec-4f70-4f59-b396-a2ca1c994007\" (UID: \"9ac262ec-4f70-4f59-b396-a2ca1c994007\") " Dec 01 09:00:20 crc kubenswrapper[4813]: I1201 09:00:20.936623 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9ac262ec-4f70-4f59-b396-a2ca1c994007-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "9ac262ec-4f70-4f59-b396-a2ca1c994007" (UID: "9ac262ec-4f70-4f59-b396-a2ca1c994007"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 09:00:20 crc kubenswrapper[4813]: I1201 09:00:20.943751 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9ac262ec-4f70-4f59-b396-a2ca1c994007-kube-api-access-r8xj7" (OuterVolumeSpecName: "kube-api-access-r8xj7") pod "9ac262ec-4f70-4f59-b396-a2ca1c994007" (UID: "9ac262ec-4f70-4f59-b396-a2ca1c994007"). InnerVolumeSpecName "kube-api-access-r8xj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:00:21 crc kubenswrapper[4813]: I1201 09:00:21.037690 4813 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9ac262ec-4f70-4f59-b396-a2ca1c994007-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 09:00:21 crc kubenswrapper[4813]: I1201 09:00:21.037724 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r8xj7\" (UniqueName: \"kubernetes.io/projected/9ac262ec-4f70-4f59-b396-a2ca1c994007-kube-api-access-r8xj7\") on node \"crc\" DevicePath \"\"" Dec 01 09:00:21 crc kubenswrapper[4813]: I1201 09:00:21.444096 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glanceb0ca-account-delete-mrcsf" event={"ID":"9ac262ec-4f70-4f59-b396-a2ca1c994007","Type":"ContainerDied","Data":"f05187ed7c941266b337778eaf15a9aa042c9ddfd392c9c787109d581c3640d8"} Dec 01 09:00:21 crc kubenswrapper[4813]: I1201 09:00:21.444136 4813 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f05187ed7c941266b337778eaf15a9aa042c9ddfd392c9c787109d581c3640d8" Dec 01 09:00:21 crc kubenswrapper[4813]: I1201 09:00:21.444377 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glanceb0ca-account-delete-mrcsf" Dec 01 09:00:22 crc kubenswrapper[4813]: I1201 09:00:22.461369 4813 generic.go:334] "Generic (PLEG): container finished" podID="ca718568-e6fc-4cce-abe3-41037ed450f9" containerID="f56c5aedb6985186051d283a5dc19ee9555e52778197169b3869af8208485568" exitCode=0 Dec 01 09:00:22 crc kubenswrapper[4813]: I1201 09:00:22.461517 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"ca718568-e6fc-4cce-abe3-41037ed450f9","Type":"ContainerDied","Data":"f56c5aedb6985186051d283a5dc19ee9555e52778197169b3869af8208485568"} Dec 01 09:00:22 crc kubenswrapper[4813]: I1201 09:00:22.851586 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Dec 01 09:00:22 crc kubenswrapper[4813]: I1201 09:00:22.893845 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ca718568-e6fc-4cce-abe3-41037ed450f9-scripts\") pod \"ca718568-e6fc-4cce-abe3-41037ed450f9\" (UID: \"ca718568-e6fc-4cce-abe3-41037ed450f9\") " Dec 01 09:00:22 crc kubenswrapper[4813]: I1201 09:00:22.893919 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ca718568-e6fc-4cce-abe3-41037ed450f9-logs\") pod \"ca718568-e6fc-4cce-abe3-41037ed450f9\" (UID: \"ca718568-e6fc-4cce-abe3-41037ed450f9\") " Dec 01 09:00:22 crc kubenswrapper[4813]: I1201 09:00:22.893945 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ca718568-e6fc-4cce-abe3-41037ed450f9-config-data\") pod \"ca718568-e6fc-4cce-abe3-41037ed450f9\" (UID: \"ca718568-e6fc-4cce-abe3-41037ed450f9\") " Dec 01 09:00:22 crc kubenswrapper[4813]: I1201 09:00:22.893997 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-27f52\" (UniqueName: \"kubernetes.io/projected/ca718568-e6fc-4cce-abe3-41037ed450f9-kube-api-access-27f52\") pod \"ca718568-e6fc-4cce-abe3-41037ed450f9\" (UID: \"ca718568-e6fc-4cce-abe3-41037ed450f9\") " Dec 01 09:00:22 crc kubenswrapper[4813]: I1201 09:00:22.894016 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca718568-e6fc-4cce-abe3-41037ed450f9-combined-ca-bundle\") pod \"ca718568-e6fc-4cce-abe3-41037ed450f9\" (UID: \"ca718568-e6fc-4cce-abe3-41037ed450f9\") " Dec 01 09:00:22 crc kubenswrapper[4813]: I1201 09:00:22.894049 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ca718568-e6fc-4cce-abe3-41037ed450f9-public-tls-certs\") pod \"ca718568-e6fc-4cce-abe3-41037ed450f9\" (UID: \"ca718568-e6fc-4cce-abe3-41037ed450f9\") " Dec 01 09:00:22 crc kubenswrapper[4813]: I1201 09:00:22.895018 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ca718568-e6fc-4cce-abe3-41037ed450f9-logs" (OuterVolumeSpecName: "logs") pod "ca718568-e6fc-4cce-abe3-41037ed450f9" (UID: "ca718568-e6fc-4cce-abe3-41037ed450f9"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 09:00:22 crc kubenswrapper[4813]: I1201 09:00:22.895362 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/ca718568-e6fc-4cce-abe3-41037ed450f9-httpd-run\") pod \"ca718568-e6fc-4cce-abe3-41037ed450f9\" (UID: \"ca718568-e6fc-4cce-abe3-41037ed450f9\") " Dec 01 09:00:22 crc kubenswrapper[4813]: I1201 09:00:22.895402 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ca718568-e6fc-4cce-abe3-41037ed450f9-internal-tls-certs\") pod \"ca718568-e6fc-4cce-abe3-41037ed450f9\" (UID: \"ca718568-e6fc-4cce-abe3-41037ed450f9\") " Dec 01 09:00:22 crc kubenswrapper[4813]: I1201 09:00:22.895435 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"ca718568-e6fc-4cce-abe3-41037ed450f9\" (UID: \"ca718568-e6fc-4cce-abe3-41037ed450f9\") " Dec 01 09:00:22 crc kubenswrapper[4813]: I1201 09:00:22.895658 4813 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ca718568-e6fc-4cce-abe3-41037ed450f9-logs\") on node \"crc\" DevicePath \"\"" Dec 01 09:00:22 crc kubenswrapper[4813]: I1201 09:00:22.896480 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ca718568-e6fc-4cce-abe3-41037ed450f9-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "ca718568-e6fc-4cce-abe3-41037ed450f9" (UID: "ca718568-e6fc-4cce-abe3-41037ed450f9"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 09:00:22 crc kubenswrapper[4813]: I1201 09:00:22.899416 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage10-crc" (OuterVolumeSpecName: "glance") pod "ca718568-e6fc-4cce-abe3-41037ed450f9" (UID: "ca718568-e6fc-4cce-abe3-41037ed450f9"). InnerVolumeSpecName "local-storage10-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 01 09:00:22 crc kubenswrapper[4813]: I1201 09:00:22.906135 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ca718568-e6fc-4cce-abe3-41037ed450f9-scripts" (OuterVolumeSpecName: "scripts") pod "ca718568-e6fc-4cce-abe3-41037ed450f9" (UID: "ca718568-e6fc-4cce-abe3-41037ed450f9"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:00:22 crc kubenswrapper[4813]: I1201 09:00:22.906325 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ca718568-e6fc-4cce-abe3-41037ed450f9-kube-api-access-27f52" (OuterVolumeSpecName: "kube-api-access-27f52") pod "ca718568-e6fc-4cce-abe3-41037ed450f9" (UID: "ca718568-e6fc-4cce-abe3-41037ed450f9"). InnerVolumeSpecName "kube-api-access-27f52". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:00:22 crc kubenswrapper[4813]: I1201 09:00:22.915498 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ca718568-e6fc-4cce-abe3-41037ed450f9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ca718568-e6fc-4cce-abe3-41037ed450f9" (UID: "ca718568-e6fc-4cce-abe3-41037ed450f9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:00:22 crc kubenswrapper[4813]: I1201 09:00:22.931774 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ca718568-e6fc-4cce-abe3-41037ed450f9-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "ca718568-e6fc-4cce-abe3-41037ed450f9" (UID: "ca718568-e6fc-4cce-abe3-41037ed450f9"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:00:22 crc kubenswrapper[4813]: I1201 09:00:22.933776 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ca718568-e6fc-4cce-abe3-41037ed450f9-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "ca718568-e6fc-4cce-abe3-41037ed450f9" (UID: "ca718568-e6fc-4cce-abe3-41037ed450f9"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:00:22 crc kubenswrapper[4813]: I1201 09:00:22.950191 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ca718568-e6fc-4cce-abe3-41037ed450f9-config-data" (OuterVolumeSpecName: "config-data") pod "ca718568-e6fc-4cce-abe3-41037ed450f9" (UID: "ca718568-e6fc-4cce-abe3-41037ed450f9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:00:22 crc kubenswrapper[4813]: I1201 09:00:22.997154 4813 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ca718568-e6fc-4cce-abe3-41037ed450f9-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 01 09:00:22 crc kubenswrapper[4813]: I1201 09:00:22.997336 4813 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" " Dec 01 09:00:22 crc kubenswrapper[4813]: I1201 09:00:22.997353 4813 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ca718568-e6fc-4cce-abe3-41037ed450f9-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 09:00:22 crc kubenswrapper[4813]: I1201 09:00:22.997366 4813 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ca718568-e6fc-4cce-abe3-41037ed450f9-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 09:00:22 crc kubenswrapper[4813]: I1201 09:00:22.997376 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-27f52\" (UniqueName: \"kubernetes.io/projected/ca718568-e6fc-4cce-abe3-41037ed450f9-kube-api-access-27f52\") on node \"crc\" DevicePath \"\"" Dec 01 09:00:22 crc kubenswrapper[4813]: I1201 09:00:22.997387 4813 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca718568-e6fc-4cce-abe3-41037ed450f9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 09:00:22 crc kubenswrapper[4813]: I1201 09:00:22.997396 4813 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ca718568-e6fc-4cce-abe3-41037ed450f9-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 01 09:00:22 crc kubenswrapper[4813]: I1201 09:00:22.997407 4813 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/ca718568-e6fc-4cce-abe3-41037ed450f9-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 01 09:00:23 crc kubenswrapper[4813]: I1201 09:00:23.012152 4813 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage10-crc" (UniqueName: "kubernetes.io/local-volume/local-storage10-crc") on node "crc" Dec 01 09:00:23 crc kubenswrapper[4813]: I1201 09:00:23.098204 4813 reconciler_common.go:293] "Volume detached for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" DevicePath \"\"" Dec 01 09:00:23 crc kubenswrapper[4813]: I1201 09:00:23.251287 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-db-create-cg9jd"] Dec 01 09:00:23 crc kubenswrapper[4813]: I1201 09:00:23.263957 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-db-create-cg9jd"] Dec 01 09:00:23 crc kubenswrapper[4813]: I1201 09:00:23.274263 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-b0ca-account-create-update-7wbnc"] Dec 01 09:00:23 crc kubenswrapper[4813]: I1201 09:00:23.282593 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glanceb0ca-account-delete-mrcsf"] Dec 01 09:00:23 crc kubenswrapper[4813]: I1201 09:00:23.290848 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glanceb0ca-account-delete-mrcsf"] Dec 01 09:00:23 crc kubenswrapper[4813]: I1201 09:00:23.296946 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-b0ca-account-create-update-7wbnc"] Dec 01 09:00:23 crc kubenswrapper[4813]: I1201 09:00:23.472462 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"ca718568-e6fc-4cce-abe3-41037ed450f9","Type":"ContainerDied","Data":"34efcfd40ea0a1695dadf0619af499499dd3035c87ef04d76dfdacf44ee9ac0e"} Dec 01 09:00:23 crc kubenswrapper[4813]: I1201 09:00:23.472615 4813 scope.go:117] "RemoveContainer" containerID="f56c5aedb6985186051d283a5dc19ee9555e52778197169b3869af8208485568" Dec 01 09:00:23 crc kubenswrapper[4813]: I1201 09:00:23.473058 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Dec 01 09:00:23 crc kubenswrapper[4813]: I1201 09:00:23.503472 4813 scope.go:117] "RemoveContainer" containerID="9f603c3520d243fd59fab8149ebff6dc69c7ef48950a8cc439f23f6c7592d59a" Dec 01 09:00:23 crc kubenswrapper[4813]: I1201 09:00:23.549008 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Dec 01 09:00:23 crc kubenswrapper[4813]: I1201 09:00:23.555529 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Dec 01 09:00:24 crc kubenswrapper[4813]: I1201 09:00:24.404546 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2e29ec70-392c-45e5-b8bb-28c604bfbcb9" path="/var/lib/kubelet/pods/2e29ec70-392c-45e5-b8bb-28c604bfbcb9/volumes" Dec 01 09:00:24 crc kubenswrapper[4813]: I1201 09:00:24.405522 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8fd46b3b-b57c-45ff-ad5e-45a10050cfa7" path="/var/lib/kubelet/pods/8fd46b3b-b57c-45ff-ad5e-45a10050cfa7/volumes" Dec 01 09:00:24 crc kubenswrapper[4813]: I1201 09:00:24.406081 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9ac262ec-4f70-4f59-b396-a2ca1c994007" path="/var/lib/kubelet/pods/9ac262ec-4f70-4f59-b396-a2ca1c994007/volumes" Dec 01 09:00:24 crc kubenswrapper[4813]: I1201 09:00:24.407139 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ca718568-e6fc-4cce-abe3-41037ed450f9" path="/var/lib/kubelet/pods/ca718568-e6fc-4cce-abe3-41037ed450f9/volumes" Dec 01 09:00:24 crc kubenswrapper[4813]: I1201 09:00:24.611733 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-db-create-nrhnq"] Dec 01 09:00:24 crc kubenswrapper[4813]: E1201 09:00:24.612348 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ca718568-e6fc-4cce-abe3-41037ed450f9" containerName="glance-httpd" Dec 01 09:00:24 crc kubenswrapper[4813]: I1201 09:00:24.612384 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="ca718568-e6fc-4cce-abe3-41037ed450f9" containerName="glance-httpd" Dec 01 09:00:24 crc kubenswrapper[4813]: E1201 09:00:24.612433 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9ac262ec-4f70-4f59-b396-a2ca1c994007" containerName="mariadb-account-delete" Dec 01 09:00:24 crc kubenswrapper[4813]: I1201 09:00:24.612443 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="9ac262ec-4f70-4f59-b396-a2ca1c994007" containerName="mariadb-account-delete" Dec 01 09:00:24 crc kubenswrapper[4813]: E1201 09:00:24.612465 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ca718568-e6fc-4cce-abe3-41037ed450f9" containerName="glance-log" Dec 01 09:00:24 crc kubenswrapper[4813]: I1201 09:00:24.612476 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="ca718568-e6fc-4cce-abe3-41037ed450f9" containerName="glance-log" Dec 01 09:00:24 crc kubenswrapper[4813]: I1201 09:00:24.612748 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="ca718568-e6fc-4cce-abe3-41037ed450f9" containerName="glance-httpd" Dec 01 09:00:24 crc kubenswrapper[4813]: I1201 09:00:24.612767 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="9ac262ec-4f70-4f59-b396-a2ca1c994007" containerName="mariadb-account-delete" Dec 01 09:00:24 crc kubenswrapper[4813]: I1201 09:00:24.612794 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="ca718568-e6fc-4cce-abe3-41037ed450f9" containerName="glance-log" Dec 01 09:00:24 crc kubenswrapper[4813]: I1201 09:00:24.613652 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-nrhnq" Dec 01 09:00:24 crc kubenswrapper[4813]: I1201 09:00:24.626310 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-532e-account-create-update-knrg7"] Dec 01 09:00:24 crc kubenswrapper[4813]: I1201 09:00:24.627759 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-532e-account-create-update-knrg7" Dec 01 09:00:24 crc kubenswrapper[4813]: I1201 09:00:24.630995 4813 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-db-secret" Dec 01 09:00:24 crc kubenswrapper[4813]: I1201 09:00:24.636947 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-create-nrhnq"] Dec 01 09:00:24 crc kubenswrapper[4813]: I1201 09:00:24.648164 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-532e-account-create-update-knrg7"] Dec 01 09:00:24 crc kubenswrapper[4813]: I1201 09:00:24.726476 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ab555f70-ee92-4aad-9324-0da0fb3770a1-operator-scripts\") pod \"glance-db-create-nrhnq\" (UID: \"ab555f70-ee92-4aad-9324-0da0fb3770a1\") " pod="glance-kuttl-tests/glance-db-create-nrhnq" Dec 01 09:00:24 crc kubenswrapper[4813]: I1201 09:00:24.727123 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w86fk\" (UniqueName: \"kubernetes.io/projected/ab555f70-ee92-4aad-9324-0da0fb3770a1-kube-api-access-w86fk\") pod \"glance-db-create-nrhnq\" (UID: \"ab555f70-ee92-4aad-9324-0da0fb3770a1\") " pod="glance-kuttl-tests/glance-db-create-nrhnq" Dec 01 09:00:24 crc kubenswrapper[4813]: I1201 09:00:24.828562 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ab555f70-ee92-4aad-9324-0da0fb3770a1-operator-scripts\") pod \"glance-db-create-nrhnq\" (UID: \"ab555f70-ee92-4aad-9324-0da0fb3770a1\") " pod="glance-kuttl-tests/glance-db-create-nrhnq" Dec 01 09:00:24 crc kubenswrapper[4813]: I1201 09:00:24.828642 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a53d021c-7061-4105-bed8-be1b1ad0d763-operator-scripts\") pod \"glance-532e-account-create-update-knrg7\" (UID: \"a53d021c-7061-4105-bed8-be1b1ad0d763\") " pod="glance-kuttl-tests/glance-532e-account-create-update-knrg7" Dec 01 09:00:24 crc kubenswrapper[4813]: I1201 09:00:24.828669 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zjrl6\" (UniqueName: \"kubernetes.io/projected/a53d021c-7061-4105-bed8-be1b1ad0d763-kube-api-access-zjrl6\") pod \"glance-532e-account-create-update-knrg7\" (UID: \"a53d021c-7061-4105-bed8-be1b1ad0d763\") " pod="glance-kuttl-tests/glance-532e-account-create-update-knrg7" Dec 01 09:00:24 crc kubenswrapper[4813]: I1201 09:00:24.828749 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w86fk\" (UniqueName: \"kubernetes.io/projected/ab555f70-ee92-4aad-9324-0da0fb3770a1-kube-api-access-w86fk\") pod \"glance-db-create-nrhnq\" (UID: \"ab555f70-ee92-4aad-9324-0da0fb3770a1\") " pod="glance-kuttl-tests/glance-db-create-nrhnq" Dec 01 09:00:24 crc kubenswrapper[4813]: I1201 09:00:24.830278 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ab555f70-ee92-4aad-9324-0da0fb3770a1-operator-scripts\") pod \"glance-db-create-nrhnq\" (UID: \"ab555f70-ee92-4aad-9324-0da0fb3770a1\") " pod="glance-kuttl-tests/glance-db-create-nrhnq" Dec 01 09:00:24 crc kubenswrapper[4813]: I1201 09:00:24.847748 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w86fk\" (UniqueName: \"kubernetes.io/projected/ab555f70-ee92-4aad-9324-0da0fb3770a1-kube-api-access-w86fk\") pod \"glance-db-create-nrhnq\" (UID: \"ab555f70-ee92-4aad-9324-0da0fb3770a1\") " pod="glance-kuttl-tests/glance-db-create-nrhnq" Dec 01 09:00:24 crc kubenswrapper[4813]: I1201 09:00:24.930558 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a53d021c-7061-4105-bed8-be1b1ad0d763-operator-scripts\") pod \"glance-532e-account-create-update-knrg7\" (UID: \"a53d021c-7061-4105-bed8-be1b1ad0d763\") " pod="glance-kuttl-tests/glance-532e-account-create-update-knrg7" Dec 01 09:00:24 crc kubenswrapper[4813]: I1201 09:00:24.930663 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zjrl6\" (UniqueName: \"kubernetes.io/projected/a53d021c-7061-4105-bed8-be1b1ad0d763-kube-api-access-zjrl6\") pod \"glance-532e-account-create-update-knrg7\" (UID: \"a53d021c-7061-4105-bed8-be1b1ad0d763\") " pod="glance-kuttl-tests/glance-532e-account-create-update-knrg7" Dec 01 09:00:24 crc kubenswrapper[4813]: I1201 09:00:24.932166 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a53d021c-7061-4105-bed8-be1b1ad0d763-operator-scripts\") pod \"glance-532e-account-create-update-knrg7\" (UID: \"a53d021c-7061-4105-bed8-be1b1ad0d763\") " pod="glance-kuttl-tests/glance-532e-account-create-update-knrg7" Dec 01 09:00:24 crc kubenswrapper[4813]: I1201 09:00:24.934011 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-nrhnq" Dec 01 09:00:25 crc kubenswrapper[4813]: I1201 09:00:25.010787 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zjrl6\" (UniqueName: \"kubernetes.io/projected/a53d021c-7061-4105-bed8-be1b1ad0d763-kube-api-access-zjrl6\") pod \"glance-532e-account-create-update-knrg7\" (UID: \"a53d021c-7061-4105-bed8-be1b1ad0d763\") " pod="glance-kuttl-tests/glance-532e-account-create-update-knrg7" Dec 01 09:00:25 crc kubenswrapper[4813]: I1201 09:00:25.257003 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-532e-account-create-update-knrg7" Dec 01 09:00:25 crc kubenswrapper[4813]: I1201 09:00:25.643678 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-create-nrhnq"] Dec 01 09:00:25 crc kubenswrapper[4813]: I1201 09:00:25.679372 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-532e-account-create-update-knrg7"] Dec 01 09:00:25 crc kubenswrapper[4813]: W1201 09:00:25.689686 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda53d021c_7061_4105_bed8_be1b1ad0d763.slice/crio-4b9a85f11edaf7e20596c10e328c60e2f4567756cc93fd2f09c5ff9f8689ad84 WatchSource:0}: Error finding container 4b9a85f11edaf7e20596c10e328c60e2f4567756cc93fd2f09c5ff9f8689ad84: Status 404 returned error can't find the container with id 4b9a85f11edaf7e20596c10e328c60e2f4567756cc93fd2f09c5ff9f8689ad84 Dec 01 09:00:26 crc kubenswrapper[4813]: I1201 09:00:26.503214 4813 generic.go:334] "Generic (PLEG): container finished" podID="a53d021c-7061-4105-bed8-be1b1ad0d763" containerID="999cf5344afaf180b4fdc0b4a14b75b0a4918a25cb64e82b2fda77658ac894f7" exitCode=0 Dec 01 09:00:26 crc kubenswrapper[4813]: I1201 09:00:26.503315 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-532e-account-create-update-knrg7" event={"ID":"a53d021c-7061-4105-bed8-be1b1ad0d763","Type":"ContainerDied","Data":"999cf5344afaf180b4fdc0b4a14b75b0a4918a25cb64e82b2fda77658ac894f7"} Dec 01 09:00:26 crc kubenswrapper[4813]: I1201 09:00:26.503544 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-532e-account-create-update-knrg7" event={"ID":"a53d021c-7061-4105-bed8-be1b1ad0d763","Type":"ContainerStarted","Data":"4b9a85f11edaf7e20596c10e328c60e2f4567756cc93fd2f09c5ff9f8689ad84"} Dec 01 09:00:26 crc kubenswrapper[4813]: I1201 09:00:26.507279 4813 generic.go:334] "Generic (PLEG): container finished" podID="ab555f70-ee92-4aad-9324-0da0fb3770a1" containerID="65fc01f686054552418e0984b87e8e822fc48eddd50873431e757f3f28f510f4" exitCode=0 Dec 01 09:00:26 crc kubenswrapper[4813]: I1201 09:00:26.507330 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-create-nrhnq" event={"ID":"ab555f70-ee92-4aad-9324-0da0fb3770a1","Type":"ContainerDied","Data":"65fc01f686054552418e0984b87e8e822fc48eddd50873431e757f3f28f510f4"} Dec 01 09:00:26 crc kubenswrapper[4813]: I1201 09:00:26.507357 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-create-nrhnq" event={"ID":"ab555f70-ee92-4aad-9324-0da0fb3770a1","Type":"ContainerStarted","Data":"7c46a01a1503e102c2ea132654b413a53496b8278ae27826e7e8135afe313b4e"} Dec 01 09:00:27 crc kubenswrapper[4813]: I1201 09:00:27.868998 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-532e-account-create-update-knrg7" Dec 01 09:00:27 crc kubenswrapper[4813]: I1201 09:00:27.874120 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-nrhnq" Dec 01 09:00:27 crc kubenswrapper[4813]: I1201 09:00:27.984404 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ab555f70-ee92-4aad-9324-0da0fb3770a1-operator-scripts\") pod \"ab555f70-ee92-4aad-9324-0da0fb3770a1\" (UID: \"ab555f70-ee92-4aad-9324-0da0fb3770a1\") " Dec 01 09:00:27 crc kubenswrapper[4813]: I1201 09:00:27.984531 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a53d021c-7061-4105-bed8-be1b1ad0d763-operator-scripts\") pod \"a53d021c-7061-4105-bed8-be1b1ad0d763\" (UID: \"a53d021c-7061-4105-bed8-be1b1ad0d763\") " Dec 01 09:00:27 crc kubenswrapper[4813]: I1201 09:00:27.984635 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w86fk\" (UniqueName: \"kubernetes.io/projected/ab555f70-ee92-4aad-9324-0da0fb3770a1-kube-api-access-w86fk\") pod \"ab555f70-ee92-4aad-9324-0da0fb3770a1\" (UID: \"ab555f70-ee92-4aad-9324-0da0fb3770a1\") " Dec 01 09:00:27 crc kubenswrapper[4813]: I1201 09:00:27.984712 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zjrl6\" (UniqueName: \"kubernetes.io/projected/a53d021c-7061-4105-bed8-be1b1ad0d763-kube-api-access-zjrl6\") pod \"a53d021c-7061-4105-bed8-be1b1ad0d763\" (UID: \"a53d021c-7061-4105-bed8-be1b1ad0d763\") " Dec 01 09:00:27 crc kubenswrapper[4813]: I1201 09:00:27.985227 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ab555f70-ee92-4aad-9324-0da0fb3770a1-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "ab555f70-ee92-4aad-9324-0da0fb3770a1" (UID: "ab555f70-ee92-4aad-9324-0da0fb3770a1"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 09:00:27 crc kubenswrapper[4813]: I1201 09:00:27.986330 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a53d021c-7061-4105-bed8-be1b1ad0d763-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "a53d021c-7061-4105-bed8-be1b1ad0d763" (UID: "a53d021c-7061-4105-bed8-be1b1ad0d763"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 09:00:27 crc kubenswrapper[4813]: I1201 09:00:27.992483 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ab555f70-ee92-4aad-9324-0da0fb3770a1-kube-api-access-w86fk" (OuterVolumeSpecName: "kube-api-access-w86fk") pod "ab555f70-ee92-4aad-9324-0da0fb3770a1" (UID: "ab555f70-ee92-4aad-9324-0da0fb3770a1"). InnerVolumeSpecName "kube-api-access-w86fk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:00:27 crc kubenswrapper[4813]: I1201 09:00:27.993315 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a53d021c-7061-4105-bed8-be1b1ad0d763-kube-api-access-zjrl6" (OuterVolumeSpecName: "kube-api-access-zjrl6") pod "a53d021c-7061-4105-bed8-be1b1ad0d763" (UID: "a53d021c-7061-4105-bed8-be1b1ad0d763"). InnerVolumeSpecName "kube-api-access-zjrl6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:00:28 crc kubenswrapper[4813]: I1201 09:00:28.086597 4813 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ab555f70-ee92-4aad-9324-0da0fb3770a1-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 09:00:28 crc kubenswrapper[4813]: I1201 09:00:28.086651 4813 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a53d021c-7061-4105-bed8-be1b1ad0d763-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 09:00:28 crc kubenswrapper[4813]: I1201 09:00:28.086667 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w86fk\" (UniqueName: \"kubernetes.io/projected/ab555f70-ee92-4aad-9324-0da0fb3770a1-kube-api-access-w86fk\") on node \"crc\" DevicePath \"\"" Dec 01 09:00:28 crc kubenswrapper[4813]: I1201 09:00:28.086685 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zjrl6\" (UniqueName: \"kubernetes.io/projected/a53d021c-7061-4105-bed8-be1b1ad0d763-kube-api-access-zjrl6\") on node \"crc\" DevicePath \"\"" Dec 01 09:00:28 crc kubenswrapper[4813]: I1201 09:00:28.531632 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-nrhnq" Dec 01 09:00:28 crc kubenswrapper[4813]: I1201 09:00:28.531631 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-create-nrhnq" event={"ID":"ab555f70-ee92-4aad-9324-0da0fb3770a1","Type":"ContainerDied","Data":"7c46a01a1503e102c2ea132654b413a53496b8278ae27826e7e8135afe313b4e"} Dec 01 09:00:28 crc kubenswrapper[4813]: I1201 09:00:28.531825 4813 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7c46a01a1503e102c2ea132654b413a53496b8278ae27826e7e8135afe313b4e" Dec 01 09:00:28 crc kubenswrapper[4813]: I1201 09:00:28.533572 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-532e-account-create-update-knrg7" event={"ID":"a53d021c-7061-4105-bed8-be1b1ad0d763","Type":"ContainerDied","Data":"4b9a85f11edaf7e20596c10e328c60e2f4567756cc93fd2f09c5ff9f8689ad84"} Dec 01 09:00:28 crc kubenswrapper[4813]: I1201 09:00:28.533606 4813 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4b9a85f11edaf7e20596c10e328c60e2f4567756cc93fd2f09c5ff9f8689ad84" Dec 01 09:00:28 crc kubenswrapper[4813]: I1201 09:00:28.533653 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-532e-account-create-update-knrg7" Dec 01 09:00:29 crc kubenswrapper[4813]: I1201 09:00:29.772534 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-db-sync-75cp2"] Dec 01 09:00:29 crc kubenswrapper[4813]: E1201 09:00:29.773227 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a53d021c-7061-4105-bed8-be1b1ad0d763" containerName="mariadb-account-create-update" Dec 01 09:00:29 crc kubenswrapper[4813]: I1201 09:00:29.773263 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="a53d021c-7061-4105-bed8-be1b1ad0d763" containerName="mariadb-account-create-update" Dec 01 09:00:29 crc kubenswrapper[4813]: E1201 09:00:29.773315 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ab555f70-ee92-4aad-9324-0da0fb3770a1" containerName="mariadb-database-create" Dec 01 09:00:29 crc kubenswrapper[4813]: I1201 09:00:29.773333 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="ab555f70-ee92-4aad-9324-0da0fb3770a1" containerName="mariadb-database-create" Dec 01 09:00:29 crc kubenswrapper[4813]: I1201 09:00:29.773658 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="a53d021c-7061-4105-bed8-be1b1ad0d763" containerName="mariadb-account-create-update" Dec 01 09:00:29 crc kubenswrapper[4813]: I1201 09:00:29.773717 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="ab555f70-ee92-4aad-9324-0da0fb3770a1" containerName="mariadb-database-create" Dec 01 09:00:29 crc kubenswrapper[4813]: I1201 09:00:29.774852 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-75cp2" Dec 01 09:00:29 crc kubenswrapper[4813]: I1201 09:00:29.777733 4813 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-glance-dockercfg-hrdhr" Dec 01 09:00:29 crc kubenswrapper[4813]: I1201 09:00:29.777872 4813 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-config-data" Dec 01 09:00:29 crc kubenswrapper[4813]: I1201 09:00:29.788999 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-sync-75cp2"] Dec 01 09:00:29 crc kubenswrapper[4813]: I1201 09:00:29.914088 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/20498728-2d3d-4736-b3f7-a2f32d3a1c0c-db-sync-config-data\") pod \"glance-db-sync-75cp2\" (UID: \"20498728-2d3d-4736-b3f7-a2f32d3a1c0c\") " pod="glance-kuttl-tests/glance-db-sync-75cp2" Dec 01 09:00:29 crc kubenswrapper[4813]: I1201 09:00:29.914156 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m6gnq\" (UniqueName: \"kubernetes.io/projected/20498728-2d3d-4736-b3f7-a2f32d3a1c0c-kube-api-access-m6gnq\") pod \"glance-db-sync-75cp2\" (UID: \"20498728-2d3d-4736-b3f7-a2f32d3a1c0c\") " pod="glance-kuttl-tests/glance-db-sync-75cp2" Dec 01 09:00:29 crc kubenswrapper[4813]: I1201 09:00:29.914269 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/20498728-2d3d-4736-b3f7-a2f32d3a1c0c-config-data\") pod \"glance-db-sync-75cp2\" (UID: \"20498728-2d3d-4736-b3f7-a2f32d3a1c0c\") " pod="glance-kuttl-tests/glance-db-sync-75cp2" Dec 01 09:00:30 crc kubenswrapper[4813]: I1201 09:00:30.015669 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/20498728-2d3d-4736-b3f7-a2f32d3a1c0c-config-data\") pod \"glance-db-sync-75cp2\" (UID: \"20498728-2d3d-4736-b3f7-a2f32d3a1c0c\") " pod="glance-kuttl-tests/glance-db-sync-75cp2" Dec 01 09:00:30 crc kubenswrapper[4813]: I1201 09:00:30.015790 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/20498728-2d3d-4736-b3f7-a2f32d3a1c0c-db-sync-config-data\") pod \"glance-db-sync-75cp2\" (UID: \"20498728-2d3d-4736-b3f7-a2f32d3a1c0c\") " pod="glance-kuttl-tests/glance-db-sync-75cp2" Dec 01 09:00:30 crc kubenswrapper[4813]: I1201 09:00:30.015819 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m6gnq\" (UniqueName: \"kubernetes.io/projected/20498728-2d3d-4736-b3f7-a2f32d3a1c0c-kube-api-access-m6gnq\") pod \"glance-db-sync-75cp2\" (UID: \"20498728-2d3d-4736-b3f7-a2f32d3a1c0c\") " pod="glance-kuttl-tests/glance-db-sync-75cp2" Dec 01 09:00:30 crc kubenswrapper[4813]: I1201 09:00:30.022220 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/20498728-2d3d-4736-b3f7-a2f32d3a1c0c-config-data\") pod \"glance-db-sync-75cp2\" (UID: \"20498728-2d3d-4736-b3f7-a2f32d3a1c0c\") " pod="glance-kuttl-tests/glance-db-sync-75cp2" Dec 01 09:00:30 crc kubenswrapper[4813]: I1201 09:00:30.022276 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/20498728-2d3d-4736-b3f7-a2f32d3a1c0c-db-sync-config-data\") pod \"glance-db-sync-75cp2\" (UID: \"20498728-2d3d-4736-b3f7-a2f32d3a1c0c\") " pod="glance-kuttl-tests/glance-db-sync-75cp2" Dec 01 09:00:30 crc kubenswrapper[4813]: I1201 09:00:30.052341 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m6gnq\" (UniqueName: \"kubernetes.io/projected/20498728-2d3d-4736-b3f7-a2f32d3a1c0c-kube-api-access-m6gnq\") pod \"glance-db-sync-75cp2\" (UID: \"20498728-2d3d-4736-b3f7-a2f32d3a1c0c\") " pod="glance-kuttl-tests/glance-db-sync-75cp2" Dec 01 09:00:30 crc kubenswrapper[4813]: I1201 09:00:30.113662 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-75cp2" Dec 01 09:00:30 crc kubenswrapper[4813]: I1201 09:00:30.559400 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-sync-75cp2"] Dec 01 09:00:30 crc kubenswrapper[4813]: W1201 09:00:30.563992 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod20498728_2d3d_4736_b3f7_a2f32d3a1c0c.slice/crio-3a2d3559023fec971b24cbd35f5e208ab921c35674faac41509efe1d05e48489 WatchSource:0}: Error finding container 3a2d3559023fec971b24cbd35f5e208ab921c35674faac41509efe1d05e48489: Status 404 returned error can't find the container with id 3a2d3559023fec971b24cbd35f5e208ab921c35674faac41509efe1d05e48489 Dec 01 09:00:31 crc kubenswrapper[4813]: I1201 09:00:31.558074 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-75cp2" event={"ID":"20498728-2d3d-4736-b3f7-a2f32d3a1c0c","Type":"ContainerStarted","Data":"90721cbc8882119c577ead880faf6ef839f46c442f67c31345fe35b86e07362f"} Dec 01 09:00:31 crc kubenswrapper[4813]: I1201 09:00:31.558449 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-75cp2" event={"ID":"20498728-2d3d-4736-b3f7-a2f32d3a1c0c","Type":"ContainerStarted","Data":"3a2d3559023fec971b24cbd35f5e208ab921c35674faac41509efe1d05e48489"} Dec 01 09:00:31 crc kubenswrapper[4813]: I1201 09:00:31.576313 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-db-sync-75cp2" podStartSLOduration=2.576265062 podStartE2EDuration="2.576265062s" podCreationTimestamp="2025-12-01 09:00:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 09:00:31.571508398 +0000 UTC m=+1216.194429984" watchObservedRunningTime="2025-12-01 09:00:31.576265062 +0000 UTC m=+1216.199186648" Dec 01 09:00:34 crc kubenswrapper[4813]: I1201 09:00:34.586017 4813 generic.go:334] "Generic (PLEG): container finished" podID="20498728-2d3d-4736-b3f7-a2f32d3a1c0c" containerID="90721cbc8882119c577ead880faf6ef839f46c442f67c31345fe35b86e07362f" exitCode=0 Dec 01 09:00:34 crc kubenswrapper[4813]: I1201 09:00:34.586090 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-75cp2" event={"ID":"20498728-2d3d-4736-b3f7-a2f32d3a1c0c","Type":"ContainerDied","Data":"90721cbc8882119c577ead880faf6ef839f46c442f67c31345fe35b86e07362f"} Dec 01 09:00:35 crc kubenswrapper[4813]: I1201 09:00:35.901299 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-75cp2" Dec 01 09:00:36 crc kubenswrapper[4813]: I1201 09:00:36.020819 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m6gnq\" (UniqueName: \"kubernetes.io/projected/20498728-2d3d-4736-b3f7-a2f32d3a1c0c-kube-api-access-m6gnq\") pod \"20498728-2d3d-4736-b3f7-a2f32d3a1c0c\" (UID: \"20498728-2d3d-4736-b3f7-a2f32d3a1c0c\") " Dec 01 09:00:36 crc kubenswrapper[4813]: I1201 09:00:36.021324 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/20498728-2d3d-4736-b3f7-a2f32d3a1c0c-config-data\") pod \"20498728-2d3d-4736-b3f7-a2f32d3a1c0c\" (UID: \"20498728-2d3d-4736-b3f7-a2f32d3a1c0c\") " Dec 01 09:00:36 crc kubenswrapper[4813]: I1201 09:00:36.021377 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/20498728-2d3d-4736-b3f7-a2f32d3a1c0c-db-sync-config-data\") pod \"20498728-2d3d-4736-b3f7-a2f32d3a1c0c\" (UID: \"20498728-2d3d-4736-b3f7-a2f32d3a1c0c\") " Dec 01 09:00:36 crc kubenswrapper[4813]: I1201 09:00:36.026083 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20498728-2d3d-4736-b3f7-a2f32d3a1c0c-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "20498728-2d3d-4736-b3f7-a2f32d3a1c0c" (UID: "20498728-2d3d-4736-b3f7-a2f32d3a1c0c"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:00:36 crc kubenswrapper[4813]: I1201 09:00:36.026182 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20498728-2d3d-4736-b3f7-a2f32d3a1c0c-kube-api-access-m6gnq" (OuterVolumeSpecName: "kube-api-access-m6gnq") pod "20498728-2d3d-4736-b3f7-a2f32d3a1c0c" (UID: "20498728-2d3d-4736-b3f7-a2f32d3a1c0c"). InnerVolumeSpecName "kube-api-access-m6gnq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:00:36 crc kubenswrapper[4813]: I1201 09:00:36.054897 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20498728-2d3d-4736-b3f7-a2f32d3a1c0c-config-data" (OuterVolumeSpecName: "config-data") pod "20498728-2d3d-4736-b3f7-a2f32d3a1c0c" (UID: "20498728-2d3d-4736-b3f7-a2f32d3a1c0c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:00:36 crc kubenswrapper[4813]: I1201 09:00:36.122549 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m6gnq\" (UniqueName: \"kubernetes.io/projected/20498728-2d3d-4736-b3f7-a2f32d3a1c0c-kube-api-access-m6gnq\") on node \"crc\" DevicePath \"\"" Dec 01 09:00:36 crc kubenswrapper[4813]: I1201 09:00:36.122583 4813 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/20498728-2d3d-4736-b3f7-a2f32d3a1c0c-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 09:00:36 crc kubenswrapper[4813]: I1201 09:00:36.122592 4813 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/20498728-2d3d-4736-b3f7-a2f32d3a1c0c-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 09:00:36 crc kubenswrapper[4813]: I1201 09:00:36.605000 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-75cp2" event={"ID":"20498728-2d3d-4736-b3f7-a2f32d3a1c0c","Type":"ContainerDied","Data":"3a2d3559023fec971b24cbd35f5e208ab921c35674faac41509efe1d05e48489"} Dec 01 09:00:36 crc kubenswrapper[4813]: I1201 09:00:36.605423 4813 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3a2d3559023fec971b24cbd35f5e208ab921c35674faac41509efe1d05e48489" Dec 01 09:00:36 crc kubenswrapper[4813]: I1201 09:00:36.605283 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-75cp2" Dec 01 09:00:37 crc kubenswrapper[4813]: I1201 09:00:37.667291 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-external-api-0"] Dec 01 09:00:37 crc kubenswrapper[4813]: E1201 09:00:37.667648 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="20498728-2d3d-4736-b3f7-a2f32d3a1c0c" containerName="glance-db-sync" Dec 01 09:00:37 crc kubenswrapper[4813]: I1201 09:00:37.667665 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="20498728-2d3d-4736-b3f7-a2f32d3a1c0c" containerName="glance-db-sync" Dec 01 09:00:37 crc kubenswrapper[4813]: I1201 09:00:37.671600 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="20498728-2d3d-4736-b3f7-a2f32d3a1c0c" containerName="glance-db-sync" Dec 01 09:00:37 crc kubenswrapper[4813]: I1201 09:00:37.678004 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-0" Dec 01 09:00:37 crc kubenswrapper[4813]: I1201 09:00:37.687426 4813 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-scripts" Dec 01 09:00:37 crc kubenswrapper[4813]: I1201 09:00:37.687687 4813 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-default-external-config-data" Dec 01 09:00:37 crc kubenswrapper[4813]: I1201 09:00:37.687924 4813 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-glance-dockercfg-hrdhr" Dec 01 09:00:37 crc kubenswrapper[4813]: I1201 09:00:37.689204 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-0"] Dec 01 09:00:37 crc kubenswrapper[4813]: I1201 09:00:37.864922 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/e738e940-82c7-481a-8a7d-069215ae99c7-dev\") pod \"glance-default-external-api-0\" (UID: \"e738e940-82c7-481a-8a7d-069215ae99c7\") " pod="glance-kuttl-tests/glance-default-external-api-0" Dec 01 09:00:37 crc kubenswrapper[4813]: I1201 09:00:37.864998 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/e738e940-82c7-481a-8a7d-069215ae99c7-etc-iscsi\") pod \"glance-default-external-api-0\" (UID: \"e738e940-82c7-481a-8a7d-069215ae99c7\") " pod="glance-kuttl-tests/glance-default-external-api-0" Dec 01 09:00:37 crc kubenswrapper[4813]: I1201 09:00:37.865021 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/e738e940-82c7-481a-8a7d-069215ae99c7-lib-modules\") pod \"glance-default-external-api-0\" (UID: \"e738e940-82c7-481a-8a7d-069215ae99c7\") " pod="glance-kuttl-tests/glance-default-external-api-0" Dec 01 09:00:37 crc kubenswrapper[4813]: I1201 09:00:37.865041 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/e738e940-82c7-481a-8a7d-069215ae99c7-run\") pod \"glance-default-external-api-0\" (UID: \"e738e940-82c7-481a-8a7d-069215ae99c7\") " pod="glance-kuttl-tests/glance-default-external-api-0" Dec 01 09:00:37 crc kubenswrapper[4813]: I1201 09:00:37.865069 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage13-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage13-crc\") pod \"glance-default-external-api-0\" (UID: \"e738e940-82c7-481a-8a7d-069215ae99c7\") " pod="glance-kuttl-tests/glance-default-external-api-0" Dec 01 09:00:37 crc kubenswrapper[4813]: I1201 09:00:37.865097 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/e738e940-82c7-481a-8a7d-069215ae99c7-sys\") pod \"glance-default-external-api-0\" (UID: \"e738e940-82c7-481a-8a7d-069215ae99c7\") " pod="glance-kuttl-tests/glance-default-external-api-0" Dec 01 09:00:37 crc kubenswrapper[4813]: I1201 09:00:37.865131 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9787f\" (UniqueName: \"kubernetes.io/projected/e738e940-82c7-481a-8a7d-069215ae99c7-kube-api-access-9787f\") pod \"glance-default-external-api-0\" (UID: \"e738e940-82c7-481a-8a7d-069215ae99c7\") " pod="glance-kuttl-tests/glance-default-external-api-0" Dec 01 09:00:37 crc kubenswrapper[4813]: I1201 09:00:37.865150 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e738e940-82c7-481a-8a7d-069215ae99c7-logs\") pod \"glance-default-external-api-0\" (UID: \"e738e940-82c7-481a-8a7d-069215ae99c7\") " pod="glance-kuttl-tests/glance-default-external-api-0" Dec 01 09:00:37 crc kubenswrapper[4813]: I1201 09:00:37.865168 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-external-api-0\" (UID: \"e738e940-82c7-481a-8a7d-069215ae99c7\") " pod="glance-kuttl-tests/glance-default-external-api-0" Dec 01 09:00:37 crc kubenswrapper[4813]: I1201 09:00:37.865365 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e738e940-82c7-481a-8a7d-069215ae99c7-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"e738e940-82c7-481a-8a7d-069215ae99c7\") " pod="glance-kuttl-tests/glance-default-external-api-0" Dec 01 09:00:37 crc kubenswrapper[4813]: I1201 09:00:37.865416 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/e738e940-82c7-481a-8a7d-069215ae99c7-var-locks-brick\") pod \"glance-default-external-api-0\" (UID: \"e738e940-82c7-481a-8a7d-069215ae99c7\") " pod="glance-kuttl-tests/glance-default-external-api-0" Dec 01 09:00:37 crc kubenswrapper[4813]: I1201 09:00:37.865433 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e738e940-82c7-481a-8a7d-069215ae99c7-scripts\") pod \"glance-default-external-api-0\" (UID: \"e738e940-82c7-481a-8a7d-069215ae99c7\") " pod="glance-kuttl-tests/glance-default-external-api-0" Dec 01 09:00:37 crc kubenswrapper[4813]: I1201 09:00:37.865447 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/e738e940-82c7-481a-8a7d-069215ae99c7-etc-nvme\") pod \"glance-default-external-api-0\" (UID: \"e738e940-82c7-481a-8a7d-069215ae99c7\") " pod="glance-kuttl-tests/glance-default-external-api-0" Dec 01 09:00:37 crc kubenswrapper[4813]: I1201 09:00:37.865467 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e738e940-82c7-481a-8a7d-069215ae99c7-config-data\") pod \"glance-default-external-api-0\" (UID: \"e738e940-82c7-481a-8a7d-069215ae99c7\") " pod="glance-kuttl-tests/glance-default-external-api-0" Dec 01 09:00:37 crc kubenswrapper[4813]: I1201 09:00:37.966707 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage13-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage13-crc\") pod \"glance-default-external-api-0\" (UID: \"e738e940-82c7-481a-8a7d-069215ae99c7\") " pod="glance-kuttl-tests/glance-default-external-api-0" Dec 01 09:00:37 crc kubenswrapper[4813]: I1201 09:00:37.966767 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/e738e940-82c7-481a-8a7d-069215ae99c7-sys\") pod \"glance-default-external-api-0\" (UID: \"e738e940-82c7-481a-8a7d-069215ae99c7\") " pod="glance-kuttl-tests/glance-default-external-api-0" Dec 01 09:00:37 crc kubenswrapper[4813]: I1201 09:00:37.966808 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9787f\" (UniqueName: \"kubernetes.io/projected/e738e940-82c7-481a-8a7d-069215ae99c7-kube-api-access-9787f\") pod \"glance-default-external-api-0\" (UID: \"e738e940-82c7-481a-8a7d-069215ae99c7\") " pod="glance-kuttl-tests/glance-default-external-api-0" Dec 01 09:00:37 crc kubenswrapper[4813]: I1201 09:00:37.966826 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e738e940-82c7-481a-8a7d-069215ae99c7-logs\") pod \"glance-default-external-api-0\" (UID: \"e738e940-82c7-481a-8a7d-069215ae99c7\") " pod="glance-kuttl-tests/glance-default-external-api-0" Dec 01 09:00:37 crc kubenswrapper[4813]: I1201 09:00:37.966846 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-external-api-0\" (UID: \"e738e940-82c7-481a-8a7d-069215ae99c7\") " pod="glance-kuttl-tests/glance-default-external-api-0" Dec 01 09:00:37 crc kubenswrapper[4813]: I1201 09:00:37.966875 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e738e940-82c7-481a-8a7d-069215ae99c7-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"e738e940-82c7-481a-8a7d-069215ae99c7\") " pod="glance-kuttl-tests/glance-default-external-api-0" Dec 01 09:00:37 crc kubenswrapper[4813]: I1201 09:00:37.966899 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/e738e940-82c7-481a-8a7d-069215ae99c7-var-locks-brick\") pod \"glance-default-external-api-0\" (UID: \"e738e940-82c7-481a-8a7d-069215ae99c7\") " pod="glance-kuttl-tests/glance-default-external-api-0" Dec 01 09:00:37 crc kubenswrapper[4813]: I1201 09:00:37.966914 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e738e940-82c7-481a-8a7d-069215ae99c7-scripts\") pod \"glance-default-external-api-0\" (UID: \"e738e940-82c7-481a-8a7d-069215ae99c7\") " pod="glance-kuttl-tests/glance-default-external-api-0" Dec 01 09:00:37 crc kubenswrapper[4813]: I1201 09:00:37.966936 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/e738e940-82c7-481a-8a7d-069215ae99c7-etc-nvme\") pod \"glance-default-external-api-0\" (UID: \"e738e940-82c7-481a-8a7d-069215ae99c7\") " pod="glance-kuttl-tests/glance-default-external-api-0" Dec 01 09:00:37 crc kubenswrapper[4813]: I1201 09:00:37.966956 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e738e940-82c7-481a-8a7d-069215ae99c7-config-data\") pod \"glance-default-external-api-0\" (UID: \"e738e940-82c7-481a-8a7d-069215ae99c7\") " pod="glance-kuttl-tests/glance-default-external-api-0" Dec 01 09:00:37 crc kubenswrapper[4813]: I1201 09:00:37.966946 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/e738e940-82c7-481a-8a7d-069215ae99c7-sys\") pod \"glance-default-external-api-0\" (UID: \"e738e940-82c7-481a-8a7d-069215ae99c7\") " pod="glance-kuttl-tests/glance-default-external-api-0" Dec 01 09:00:37 crc kubenswrapper[4813]: I1201 09:00:37.967015 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/e738e940-82c7-481a-8a7d-069215ae99c7-dev\") pod \"glance-default-external-api-0\" (UID: \"e738e940-82c7-481a-8a7d-069215ae99c7\") " pod="glance-kuttl-tests/glance-default-external-api-0" Dec 01 09:00:37 crc kubenswrapper[4813]: I1201 09:00:37.967042 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/e738e940-82c7-481a-8a7d-069215ae99c7-etc-iscsi\") pod \"glance-default-external-api-0\" (UID: \"e738e940-82c7-481a-8a7d-069215ae99c7\") " pod="glance-kuttl-tests/glance-default-external-api-0" Dec 01 09:00:37 crc kubenswrapper[4813]: I1201 09:00:37.967061 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/e738e940-82c7-481a-8a7d-069215ae99c7-lib-modules\") pod \"glance-default-external-api-0\" (UID: \"e738e940-82c7-481a-8a7d-069215ae99c7\") " pod="glance-kuttl-tests/glance-default-external-api-0" Dec 01 09:00:37 crc kubenswrapper[4813]: I1201 09:00:37.967081 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/e738e940-82c7-481a-8a7d-069215ae99c7-run\") pod \"glance-default-external-api-0\" (UID: \"e738e940-82c7-481a-8a7d-069215ae99c7\") " pod="glance-kuttl-tests/glance-default-external-api-0" Dec 01 09:00:37 crc kubenswrapper[4813]: I1201 09:00:37.967172 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/e738e940-82c7-481a-8a7d-069215ae99c7-run\") pod \"glance-default-external-api-0\" (UID: \"e738e940-82c7-481a-8a7d-069215ae99c7\") " pod="glance-kuttl-tests/glance-default-external-api-0" Dec 01 09:00:37 crc kubenswrapper[4813]: I1201 09:00:37.967330 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/e738e940-82c7-481a-8a7d-069215ae99c7-etc-nvme\") pod \"glance-default-external-api-0\" (UID: \"e738e940-82c7-481a-8a7d-069215ae99c7\") " pod="glance-kuttl-tests/glance-default-external-api-0" Dec 01 09:00:37 crc kubenswrapper[4813]: I1201 09:00:37.967408 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/e738e940-82c7-481a-8a7d-069215ae99c7-var-locks-brick\") pod \"glance-default-external-api-0\" (UID: \"e738e940-82c7-481a-8a7d-069215ae99c7\") " pod="glance-kuttl-tests/glance-default-external-api-0" Dec 01 09:00:37 crc kubenswrapper[4813]: I1201 09:00:37.967402 4813 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage13-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage13-crc\") pod \"glance-default-external-api-0\" (UID: \"e738e940-82c7-481a-8a7d-069215ae99c7\") device mount path \"/mnt/openstack/pv13\"" pod="glance-kuttl-tests/glance-default-external-api-0" Dec 01 09:00:37 crc kubenswrapper[4813]: I1201 09:00:37.967431 4813 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-external-api-0\" (UID: \"e738e940-82c7-481a-8a7d-069215ae99c7\") device mount path \"/mnt/openstack/pv12\"" pod="glance-kuttl-tests/glance-default-external-api-0" Dec 01 09:00:37 crc kubenswrapper[4813]: I1201 09:00:37.967560 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e738e940-82c7-481a-8a7d-069215ae99c7-logs\") pod \"glance-default-external-api-0\" (UID: \"e738e940-82c7-481a-8a7d-069215ae99c7\") " pod="glance-kuttl-tests/glance-default-external-api-0" Dec 01 09:00:37 crc kubenswrapper[4813]: I1201 09:00:37.967636 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/e738e940-82c7-481a-8a7d-069215ae99c7-dev\") pod \"glance-default-external-api-0\" (UID: \"e738e940-82c7-481a-8a7d-069215ae99c7\") " pod="glance-kuttl-tests/glance-default-external-api-0" Dec 01 09:00:37 crc kubenswrapper[4813]: I1201 09:00:37.967738 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/e738e940-82c7-481a-8a7d-069215ae99c7-etc-iscsi\") pod \"glance-default-external-api-0\" (UID: \"e738e940-82c7-481a-8a7d-069215ae99c7\") " pod="glance-kuttl-tests/glance-default-external-api-0" Dec 01 09:00:37 crc kubenswrapper[4813]: I1201 09:00:37.967761 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/e738e940-82c7-481a-8a7d-069215ae99c7-lib-modules\") pod \"glance-default-external-api-0\" (UID: \"e738e940-82c7-481a-8a7d-069215ae99c7\") " pod="glance-kuttl-tests/glance-default-external-api-0" Dec 01 09:00:37 crc kubenswrapper[4813]: I1201 09:00:37.967759 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e738e940-82c7-481a-8a7d-069215ae99c7-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"e738e940-82c7-481a-8a7d-069215ae99c7\") " pod="glance-kuttl-tests/glance-default-external-api-0" Dec 01 09:00:37 crc kubenswrapper[4813]: I1201 09:00:37.973174 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e738e940-82c7-481a-8a7d-069215ae99c7-config-data\") pod \"glance-default-external-api-0\" (UID: \"e738e940-82c7-481a-8a7d-069215ae99c7\") " pod="glance-kuttl-tests/glance-default-external-api-0" Dec 01 09:00:37 crc kubenswrapper[4813]: I1201 09:00:37.975181 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e738e940-82c7-481a-8a7d-069215ae99c7-scripts\") pod \"glance-default-external-api-0\" (UID: \"e738e940-82c7-481a-8a7d-069215ae99c7\") " pod="glance-kuttl-tests/glance-default-external-api-0" Dec 01 09:00:37 crc kubenswrapper[4813]: I1201 09:00:37.985038 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9787f\" (UniqueName: \"kubernetes.io/projected/e738e940-82c7-481a-8a7d-069215ae99c7-kube-api-access-9787f\") pod \"glance-default-external-api-0\" (UID: \"e738e940-82c7-481a-8a7d-069215ae99c7\") " pod="glance-kuttl-tests/glance-default-external-api-0" Dec 01 09:00:37 crc kubenswrapper[4813]: I1201 09:00:37.991666 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-external-api-0\" (UID: \"e738e940-82c7-481a-8a7d-069215ae99c7\") " pod="glance-kuttl-tests/glance-default-external-api-0" Dec 01 09:00:37 crc kubenswrapper[4813]: I1201 09:00:37.998166 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage13-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage13-crc\") pod \"glance-default-external-api-0\" (UID: \"e738e940-82c7-481a-8a7d-069215ae99c7\") " pod="glance-kuttl-tests/glance-default-external-api-0" Dec 01 09:00:38 crc kubenswrapper[4813]: I1201 09:00:38.004205 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-0" Dec 01 09:00:38 crc kubenswrapper[4813]: I1201 09:00:38.256492 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Dec 01 09:00:38 crc kubenswrapper[4813]: I1201 09:00:38.258577 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:00:38 crc kubenswrapper[4813]: I1201 09:00:38.261925 4813 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-default-internal-config-data" Dec 01 09:00:38 crc kubenswrapper[4813]: I1201 09:00:38.275328 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Dec 01 09:00:38 crc kubenswrapper[4813]: I1201 09:00:38.373217 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/c2d1fc05-984a-4216-b9f5-befc2acb69db-etc-nvme\") pod \"glance-default-internal-api-0\" (UID: \"c2d1fc05-984a-4216-b9f5-befc2acb69db\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:00:38 crc kubenswrapper[4813]: I1201 09:00:38.373546 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/c2d1fc05-984a-4216-b9f5-befc2acb69db-dev\") pod \"glance-default-internal-api-0\" (UID: \"c2d1fc05-984a-4216-b9f5-befc2acb69db\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:00:38 crc kubenswrapper[4813]: I1201 09:00:38.373570 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/c2d1fc05-984a-4216-b9f5-befc2acb69db-var-locks-brick\") pod \"glance-default-internal-api-0\" (UID: \"c2d1fc05-984a-4216-b9f5-befc2acb69db\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:00:38 crc kubenswrapper[4813]: I1201 09:00:38.373592 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-djj2r\" (UniqueName: \"kubernetes.io/projected/c2d1fc05-984a-4216-b9f5-befc2acb69db-kube-api-access-djj2r\") pod \"glance-default-internal-api-0\" (UID: \"c2d1fc05-984a-4216-b9f5-befc2acb69db\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:00:38 crc kubenswrapper[4813]: I1201 09:00:38.373610 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/c2d1fc05-984a-4216-b9f5-befc2acb69db-sys\") pod \"glance-default-internal-api-0\" (UID: \"c2d1fc05-984a-4216-b9f5-befc2acb69db\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:00:38 crc kubenswrapper[4813]: I1201 09:00:38.373638 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/c2d1fc05-984a-4216-b9f5-befc2acb69db-run\") pod \"glance-default-internal-api-0\" (UID: \"c2d1fc05-984a-4216-b9f5-befc2acb69db\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:00:38 crc kubenswrapper[4813]: I1201 09:00:38.373782 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-internal-api-0\" (UID: \"c2d1fc05-984a-4216-b9f5-befc2acb69db\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:00:38 crc kubenswrapper[4813]: I1201 09:00:38.373847 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage17-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage17-crc\") pod \"glance-default-internal-api-0\" (UID: \"c2d1fc05-984a-4216-b9f5-befc2acb69db\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:00:38 crc kubenswrapper[4813]: I1201 09:00:38.373903 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c2d1fc05-984a-4216-b9f5-befc2acb69db-scripts\") pod \"glance-default-internal-api-0\" (UID: \"c2d1fc05-984a-4216-b9f5-befc2acb69db\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:00:38 crc kubenswrapper[4813]: I1201 09:00:38.373936 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c2d1fc05-984a-4216-b9f5-befc2acb69db-config-data\") pod \"glance-default-internal-api-0\" (UID: \"c2d1fc05-984a-4216-b9f5-befc2acb69db\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:00:38 crc kubenswrapper[4813]: I1201 09:00:38.373990 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/c2d1fc05-984a-4216-b9f5-befc2acb69db-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"c2d1fc05-984a-4216-b9f5-befc2acb69db\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:00:38 crc kubenswrapper[4813]: I1201 09:00:38.374066 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/c2d1fc05-984a-4216-b9f5-befc2acb69db-etc-iscsi\") pod \"glance-default-internal-api-0\" (UID: \"c2d1fc05-984a-4216-b9f5-befc2acb69db\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:00:38 crc kubenswrapper[4813]: I1201 09:00:38.374089 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c2d1fc05-984a-4216-b9f5-befc2acb69db-logs\") pod \"glance-default-internal-api-0\" (UID: \"c2d1fc05-984a-4216-b9f5-befc2acb69db\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:00:38 crc kubenswrapper[4813]: I1201 09:00:38.374141 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/c2d1fc05-984a-4216-b9f5-befc2acb69db-lib-modules\") pod \"glance-default-internal-api-0\" (UID: \"c2d1fc05-984a-4216-b9f5-befc2acb69db\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:00:38 crc kubenswrapper[4813]: I1201 09:00:38.458239 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-0"] Dec 01 09:00:38 crc kubenswrapper[4813]: I1201 09:00:38.476270 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/c2d1fc05-984a-4216-b9f5-befc2acb69db-etc-nvme\") pod \"glance-default-internal-api-0\" (UID: \"c2d1fc05-984a-4216-b9f5-befc2acb69db\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:00:38 crc kubenswrapper[4813]: I1201 09:00:38.476349 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/c2d1fc05-984a-4216-b9f5-befc2acb69db-dev\") pod \"glance-default-internal-api-0\" (UID: \"c2d1fc05-984a-4216-b9f5-befc2acb69db\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:00:38 crc kubenswrapper[4813]: I1201 09:00:38.476396 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/c2d1fc05-984a-4216-b9f5-befc2acb69db-var-locks-brick\") pod \"glance-default-internal-api-0\" (UID: \"c2d1fc05-984a-4216-b9f5-befc2acb69db\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:00:38 crc kubenswrapper[4813]: I1201 09:00:38.476414 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/c2d1fc05-984a-4216-b9f5-befc2acb69db-etc-nvme\") pod \"glance-default-internal-api-0\" (UID: \"c2d1fc05-984a-4216-b9f5-befc2acb69db\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:00:38 crc kubenswrapper[4813]: I1201 09:00:38.476451 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-djj2r\" (UniqueName: \"kubernetes.io/projected/c2d1fc05-984a-4216-b9f5-befc2acb69db-kube-api-access-djj2r\") pod \"glance-default-internal-api-0\" (UID: \"c2d1fc05-984a-4216-b9f5-befc2acb69db\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:00:38 crc kubenswrapper[4813]: I1201 09:00:38.476502 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/c2d1fc05-984a-4216-b9f5-befc2acb69db-sys\") pod \"glance-default-internal-api-0\" (UID: \"c2d1fc05-984a-4216-b9f5-befc2acb69db\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:00:38 crc kubenswrapper[4813]: I1201 09:00:38.476527 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/c2d1fc05-984a-4216-b9f5-befc2acb69db-dev\") pod \"glance-default-internal-api-0\" (UID: \"c2d1fc05-984a-4216-b9f5-befc2acb69db\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:00:38 crc kubenswrapper[4813]: I1201 09:00:38.476547 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/c2d1fc05-984a-4216-b9f5-befc2acb69db-run\") pod \"glance-default-internal-api-0\" (UID: \"c2d1fc05-984a-4216-b9f5-befc2acb69db\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:00:38 crc kubenswrapper[4813]: I1201 09:00:38.476567 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/c2d1fc05-984a-4216-b9f5-befc2acb69db-sys\") pod \"glance-default-internal-api-0\" (UID: \"c2d1fc05-984a-4216-b9f5-befc2acb69db\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:00:38 crc kubenswrapper[4813]: I1201 09:00:38.476583 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-internal-api-0\" (UID: \"c2d1fc05-984a-4216-b9f5-befc2acb69db\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:00:38 crc kubenswrapper[4813]: I1201 09:00:38.476674 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage17-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage17-crc\") pod \"glance-default-internal-api-0\" (UID: \"c2d1fc05-984a-4216-b9f5-befc2acb69db\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:00:38 crc kubenswrapper[4813]: I1201 09:00:38.476704 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c2d1fc05-984a-4216-b9f5-befc2acb69db-scripts\") pod \"glance-default-internal-api-0\" (UID: \"c2d1fc05-984a-4216-b9f5-befc2acb69db\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:00:38 crc kubenswrapper[4813]: I1201 09:00:38.476726 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c2d1fc05-984a-4216-b9f5-befc2acb69db-config-data\") pod \"glance-default-internal-api-0\" (UID: \"c2d1fc05-984a-4216-b9f5-befc2acb69db\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:00:38 crc kubenswrapper[4813]: I1201 09:00:38.476747 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/c2d1fc05-984a-4216-b9f5-befc2acb69db-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"c2d1fc05-984a-4216-b9f5-befc2acb69db\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:00:38 crc kubenswrapper[4813]: I1201 09:00:38.476781 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/c2d1fc05-984a-4216-b9f5-befc2acb69db-etc-iscsi\") pod \"glance-default-internal-api-0\" (UID: \"c2d1fc05-984a-4216-b9f5-befc2acb69db\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:00:38 crc kubenswrapper[4813]: I1201 09:00:38.476808 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c2d1fc05-984a-4216-b9f5-befc2acb69db-logs\") pod \"glance-default-internal-api-0\" (UID: \"c2d1fc05-984a-4216-b9f5-befc2acb69db\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:00:38 crc kubenswrapper[4813]: I1201 09:00:38.476839 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/c2d1fc05-984a-4216-b9f5-befc2acb69db-lib-modules\") pod \"glance-default-internal-api-0\" (UID: \"c2d1fc05-984a-4216-b9f5-befc2acb69db\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:00:38 crc kubenswrapper[4813]: I1201 09:00:38.477107 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/c2d1fc05-984a-4216-b9f5-befc2acb69db-etc-iscsi\") pod \"glance-default-internal-api-0\" (UID: \"c2d1fc05-984a-4216-b9f5-befc2acb69db\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:00:38 crc kubenswrapper[4813]: I1201 09:00:38.477576 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/c2d1fc05-984a-4216-b9f5-befc2acb69db-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"c2d1fc05-984a-4216-b9f5-befc2acb69db\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:00:38 crc kubenswrapper[4813]: I1201 09:00:38.477610 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c2d1fc05-984a-4216-b9f5-befc2acb69db-logs\") pod \"glance-default-internal-api-0\" (UID: \"c2d1fc05-984a-4216-b9f5-befc2acb69db\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:00:38 crc kubenswrapper[4813]: I1201 09:00:38.477614 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/c2d1fc05-984a-4216-b9f5-befc2acb69db-lib-modules\") pod \"glance-default-internal-api-0\" (UID: \"c2d1fc05-984a-4216-b9f5-befc2acb69db\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:00:38 crc kubenswrapper[4813]: I1201 09:00:38.477629 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/c2d1fc05-984a-4216-b9f5-befc2acb69db-run\") pod \"glance-default-internal-api-0\" (UID: \"c2d1fc05-984a-4216-b9f5-befc2acb69db\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:00:38 crc kubenswrapper[4813]: I1201 09:00:38.476501 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/c2d1fc05-984a-4216-b9f5-befc2acb69db-var-locks-brick\") pod \"glance-default-internal-api-0\" (UID: \"c2d1fc05-984a-4216-b9f5-befc2acb69db\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:00:38 crc kubenswrapper[4813]: I1201 09:00:38.478034 4813 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage17-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage17-crc\") pod \"glance-default-internal-api-0\" (UID: \"c2d1fc05-984a-4216-b9f5-befc2acb69db\") device mount path \"/mnt/openstack/pv17\"" pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:00:38 crc kubenswrapper[4813]: I1201 09:00:38.478065 4813 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-internal-api-0\" (UID: \"c2d1fc05-984a-4216-b9f5-befc2acb69db\") device mount path \"/mnt/openstack/pv01\"" pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:00:38 crc kubenswrapper[4813]: I1201 09:00:38.484234 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c2d1fc05-984a-4216-b9f5-befc2acb69db-scripts\") pod \"glance-default-internal-api-0\" (UID: \"c2d1fc05-984a-4216-b9f5-befc2acb69db\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:00:38 crc kubenswrapper[4813]: I1201 09:00:38.484425 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c2d1fc05-984a-4216-b9f5-befc2acb69db-config-data\") pod \"glance-default-internal-api-0\" (UID: \"c2d1fc05-984a-4216-b9f5-befc2acb69db\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:00:38 crc kubenswrapper[4813]: I1201 09:00:38.500354 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-djj2r\" (UniqueName: \"kubernetes.io/projected/c2d1fc05-984a-4216-b9f5-befc2acb69db-kube-api-access-djj2r\") pod \"glance-default-internal-api-0\" (UID: \"c2d1fc05-984a-4216-b9f5-befc2acb69db\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:00:38 crc kubenswrapper[4813]: I1201 09:00:38.502674 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage17-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage17-crc\") pod \"glance-default-internal-api-0\" (UID: \"c2d1fc05-984a-4216-b9f5-befc2acb69db\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:00:38 crc kubenswrapper[4813]: I1201 09:00:38.509922 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-internal-api-0\" (UID: \"c2d1fc05-984a-4216-b9f5-befc2acb69db\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:00:38 crc kubenswrapper[4813]: I1201 09:00:38.578282 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:00:38 crc kubenswrapper[4813]: I1201 09:00:38.633115 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Dec 01 09:00:38 crc kubenswrapper[4813]: I1201 09:00:38.642549 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-0" event={"ID":"e738e940-82c7-481a-8a7d-069215ae99c7","Type":"ContainerStarted","Data":"c7f8843d64fb50ae7e49ac4a05f42b1eac5e18fac3aeb36b3a1e4e47e9457f20"} Dec 01 09:00:38 crc kubenswrapper[4813]: I1201 09:00:38.642595 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-0" event={"ID":"e738e940-82c7-481a-8a7d-069215ae99c7","Type":"ContainerStarted","Data":"299e57f1bdbe70875c7a62d154f86a2a573d6cf0d878700c05def95d4a925c71"} Dec 01 09:00:38 crc kubenswrapper[4813]: I1201 09:00:38.932082 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Dec 01 09:00:38 crc kubenswrapper[4813]: W1201 09:00:38.935918 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc2d1fc05_984a_4216_b9f5_befc2acb69db.slice/crio-5d60339f0374ce44dab9a713f80d49cec7c1bd2bab44fdccf475652e79854f1c WatchSource:0}: Error finding container 5d60339f0374ce44dab9a713f80d49cec7c1bd2bab44fdccf475652e79854f1c: Status 404 returned error can't find the container with id 5d60339f0374ce44dab9a713f80d49cec7c1bd2bab44fdccf475652e79854f1c Dec 01 09:00:39 crc kubenswrapper[4813]: I1201 09:00:39.657860 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"c2d1fc05-984a-4216-b9f5-befc2acb69db","Type":"ContainerStarted","Data":"9cd9e2a2e5957ab1447d39e837d5ad9e6577d09b9acf973b948ddb61cd0d004e"} Dec 01 09:00:39 crc kubenswrapper[4813]: I1201 09:00:39.658042 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-internal-api-0" podUID="c2d1fc05-984a-4216-b9f5-befc2acb69db" containerName="glance-log" containerID="cri-o://82204e399295940f7898ab7236a95fe5ab0cf118e7275d729cf375d3f2aae2f8" gracePeriod=30 Dec 01 09:00:39 crc kubenswrapper[4813]: I1201 09:00:39.658085 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-internal-api-0" podUID="c2d1fc05-984a-4216-b9f5-befc2acb69db" containerName="glance-api" containerID="cri-o://9cd9e2a2e5957ab1447d39e837d5ad9e6577d09b9acf973b948ddb61cd0d004e" gracePeriod=30 Dec 01 09:00:39 crc kubenswrapper[4813]: I1201 09:00:39.658111 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-internal-api-0" podUID="c2d1fc05-984a-4216-b9f5-befc2acb69db" containerName="glance-httpd" containerID="cri-o://b66651a85c9e7d2664792d33c59df62688ba872ef0c03fde20d3a6e69e08b9b4" gracePeriod=30 Dec 01 09:00:39 crc kubenswrapper[4813]: I1201 09:00:39.660430 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"c2d1fc05-984a-4216-b9f5-befc2acb69db","Type":"ContainerStarted","Data":"b66651a85c9e7d2664792d33c59df62688ba872ef0c03fde20d3a6e69e08b9b4"} Dec 01 09:00:39 crc kubenswrapper[4813]: I1201 09:00:39.660639 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"c2d1fc05-984a-4216-b9f5-befc2acb69db","Type":"ContainerStarted","Data":"82204e399295940f7898ab7236a95fe5ab0cf118e7275d729cf375d3f2aae2f8"} Dec 01 09:00:39 crc kubenswrapper[4813]: I1201 09:00:39.660652 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"c2d1fc05-984a-4216-b9f5-befc2acb69db","Type":"ContainerStarted","Data":"5d60339f0374ce44dab9a713f80d49cec7c1bd2bab44fdccf475652e79854f1c"} Dec 01 09:00:39 crc kubenswrapper[4813]: I1201 09:00:39.663955 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-0" event={"ID":"e738e940-82c7-481a-8a7d-069215ae99c7","Type":"ContainerStarted","Data":"c6d074d29e895601fa04da4eca3bc4fca8102be4e3f3ac5b7fa6de4021e4f833"} Dec 01 09:00:39 crc kubenswrapper[4813]: I1201 09:00:39.664003 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-0" event={"ID":"e738e940-82c7-481a-8a7d-069215ae99c7","Type":"ContainerStarted","Data":"e3c4d4cd844c6982eeee8cf67db9ce5ebdfe25059f786d79a084320b82fe71d6"} Dec 01 09:00:39 crc kubenswrapper[4813]: I1201 09:00:39.700685 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-default-internal-api-0" podStartSLOduration=2.700661891 podStartE2EDuration="2.700661891s" podCreationTimestamp="2025-12-01 09:00:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 09:00:39.697146692 +0000 UTC m=+1224.320068298" watchObservedRunningTime="2025-12-01 09:00:39.700661891 +0000 UTC m=+1224.323583477" Dec 01 09:00:39 crc kubenswrapper[4813]: I1201 09:00:39.728631 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-default-external-api-0" podStartSLOduration=2.728581604 podStartE2EDuration="2.728581604s" podCreationTimestamp="2025-12-01 09:00:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 09:00:39.720554489 +0000 UTC m=+1224.343476075" watchObservedRunningTime="2025-12-01 09:00:39.728581604 +0000 UTC m=+1224.351503200" Dec 01 09:00:39 crc kubenswrapper[4813]: W1201 09:00:39.882003 4813 watcher.go:93] Error while processing event ("/sys/fs/cgroup/user.slice/user-0.slice/session-c12.scope": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/user.slice/user-0.slice/session-c12.scope: no such file or directory Dec 01 09:00:39 crc kubenswrapper[4813]: W1201 09:00:39.882527 4813 watcher.go:93] Error while processing event ("/sys/fs/cgroup/user.slice/user-0.slice/session-c13.scope": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/user.slice/user-0.slice/session-c13.scope: no such file or directory Dec 01 09:00:40 crc kubenswrapper[4813]: I1201 09:00:40.103557 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:00:40 crc kubenswrapper[4813]: I1201 09:00:40.209127 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-djj2r\" (UniqueName: \"kubernetes.io/projected/c2d1fc05-984a-4216-b9f5-befc2acb69db-kube-api-access-djj2r\") pod \"c2d1fc05-984a-4216-b9f5-befc2acb69db\" (UID: \"c2d1fc05-984a-4216-b9f5-befc2acb69db\") " Dec 01 09:00:40 crc kubenswrapper[4813]: I1201 09:00:40.209247 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/c2d1fc05-984a-4216-b9f5-befc2acb69db-etc-nvme\") pod \"c2d1fc05-984a-4216-b9f5-befc2acb69db\" (UID: \"c2d1fc05-984a-4216-b9f5-befc2acb69db\") " Dec 01 09:00:40 crc kubenswrapper[4813]: I1201 09:00:40.209274 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/c2d1fc05-984a-4216-b9f5-befc2acb69db-sys\") pod \"c2d1fc05-984a-4216-b9f5-befc2acb69db\" (UID: \"c2d1fc05-984a-4216-b9f5-befc2acb69db\") " Dec 01 09:00:40 crc kubenswrapper[4813]: I1201 09:00:40.209318 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c2d1fc05-984a-4216-b9f5-befc2acb69db-config-data\") pod \"c2d1fc05-984a-4216-b9f5-befc2acb69db\" (UID: \"c2d1fc05-984a-4216-b9f5-befc2acb69db\") " Dec 01 09:00:40 crc kubenswrapper[4813]: I1201 09:00:40.209344 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/c2d1fc05-984a-4216-b9f5-befc2acb69db-lib-modules\") pod \"c2d1fc05-984a-4216-b9f5-befc2acb69db\" (UID: \"c2d1fc05-984a-4216-b9f5-befc2acb69db\") " Dec 01 09:00:40 crc kubenswrapper[4813]: I1201 09:00:40.209370 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance-cache\" (UniqueName: \"kubernetes.io/local-volume/local-storage17-crc\") pod \"c2d1fc05-984a-4216-b9f5-befc2acb69db\" (UID: \"c2d1fc05-984a-4216-b9f5-befc2acb69db\") " Dec 01 09:00:40 crc kubenswrapper[4813]: I1201 09:00:40.209412 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/c2d1fc05-984a-4216-b9f5-befc2acb69db-etc-iscsi\") pod \"c2d1fc05-984a-4216-b9f5-befc2acb69db\" (UID: \"c2d1fc05-984a-4216-b9f5-befc2acb69db\") " Dec 01 09:00:40 crc kubenswrapper[4813]: I1201 09:00:40.209457 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/c2d1fc05-984a-4216-b9f5-befc2acb69db-run\") pod \"c2d1fc05-984a-4216-b9f5-befc2acb69db\" (UID: \"c2d1fc05-984a-4216-b9f5-befc2acb69db\") " Dec 01 09:00:40 crc kubenswrapper[4813]: I1201 09:00:40.209505 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c2d1fc05-984a-4216-b9f5-befc2acb69db-scripts\") pod \"c2d1fc05-984a-4216-b9f5-befc2acb69db\" (UID: \"c2d1fc05-984a-4216-b9f5-befc2acb69db\") " Dec 01 09:00:40 crc kubenswrapper[4813]: I1201 09:00:40.209541 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/c2d1fc05-984a-4216-b9f5-befc2acb69db-dev\") pod \"c2d1fc05-984a-4216-b9f5-befc2acb69db\" (UID: \"c2d1fc05-984a-4216-b9f5-befc2acb69db\") " Dec 01 09:00:40 crc kubenswrapper[4813]: I1201 09:00:40.209588 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/c2d1fc05-984a-4216-b9f5-befc2acb69db-httpd-run\") pod \"c2d1fc05-984a-4216-b9f5-befc2acb69db\" (UID: \"c2d1fc05-984a-4216-b9f5-befc2acb69db\") " Dec 01 09:00:40 crc kubenswrapper[4813]: I1201 09:00:40.209626 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/c2d1fc05-984a-4216-b9f5-befc2acb69db-var-locks-brick\") pod \"c2d1fc05-984a-4216-b9f5-befc2acb69db\" (UID: \"c2d1fc05-984a-4216-b9f5-befc2acb69db\") " Dec 01 09:00:40 crc kubenswrapper[4813]: I1201 09:00:40.209651 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c2d1fc05-984a-4216-b9f5-befc2acb69db-logs\") pod \"c2d1fc05-984a-4216-b9f5-befc2acb69db\" (UID: \"c2d1fc05-984a-4216-b9f5-befc2acb69db\") " Dec 01 09:00:40 crc kubenswrapper[4813]: I1201 09:00:40.209687 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"c2d1fc05-984a-4216-b9f5-befc2acb69db\" (UID: \"c2d1fc05-984a-4216-b9f5-befc2acb69db\") " Dec 01 09:00:40 crc kubenswrapper[4813]: I1201 09:00:40.209911 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c2d1fc05-984a-4216-b9f5-befc2acb69db-etc-iscsi" (OuterVolumeSpecName: "etc-iscsi") pod "c2d1fc05-984a-4216-b9f5-befc2acb69db" (UID: "c2d1fc05-984a-4216-b9f5-befc2acb69db"). InnerVolumeSpecName "etc-iscsi". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 09:00:40 crc kubenswrapper[4813]: I1201 09:00:40.209986 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c2d1fc05-984a-4216-b9f5-befc2acb69db-etc-nvme" (OuterVolumeSpecName: "etc-nvme") pod "c2d1fc05-984a-4216-b9f5-befc2acb69db" (UID: "c2d1fc05-984a-4216-b9f5-befc2acb69db"). InnerVolumeSpecName "etc-nvme". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 09:00:40 crc kubenswrapper[4813]: I1201 09:00:40.210039 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c2d1fc05-984a-4216-b9f5-befc2acb69db-run" (OuterVolumeSpecName: "run") pod "c2d1fc05-984a-4216-b9f5-befc2acb69db" (UID: "c2d1fc05-984a-4216-b9f5-befc2acb69db"). InnerVolumeSpecName "run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 09:00:40 crc kubenswrapper[4813]: I1201 09:00:40.210216 4813 reconciler_common.go:293] "Volume detached for volume \"run\" (UniqueName: \"kubernetes.io/host-path/c2d1fc05-984a-4216-b9f5-befc2acb69db-run\") on node \"crc\" DevicePath \"\"" Dec 01 09:00:40 crc kubenswrapper[4813]: I1201 09:00:40.210234 4813 reconciler_common.go:293] "Volume detached for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/c2d1fc05-984a-4216-b9f5-befc2acb69db-etc-nvme\") on node \"crc\" DevicePath \"\"" Dec 01 09:00:40 crc kubenswrapper[4813]: I1201 09:00:40.210246 4813 reconciler_common.go:293] "Volume detached for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/c2d1fc05-984a-4216-b9f5-befc2acb69db-etc-iscsi\") on node \"crc\" DevicePath \"\"" Dec 01 09:00:40 crc kubenswrapper[4813]: I1201 09:00:40.210276 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c2d1fc05-984a-4216-b9f5-befc2acb69db-sys" (OuterVolumeSpecName: "sys") pod "c2d1fc05-984a-4216-b9f5-befc2acb69db" (UID: "c2d1fc05-984a-4216-b9f5-befc2acb69db"). InnerVolumeSpecName "sys". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 09:00:40 crc kubenswrapper[4813]: I1201 09:00:40.210484 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c2d1fc05-984a-4216-b9f5-befc2acb69db-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "c2d1fc05-984a-4216-b9f5-befc2acb69db" (UID: "c2d1fc05-984a-4216-b9f5-befc2acb69db"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 09:00:40 crc kubenswrapper[4813]: I1201 09:00:40.210526 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c2d1fc05-984a-4216-b9f5-befc2acb69db-var-locks-brick" (OuterVolumeSpecName: "var-locks-brick") pod "c2d1fc05-984a-4216-b9f5-befc2acb69db" (UID: "c2d1fc05-984a-4216-b9f5-befc2acb69db"). InnerVolumeSpecName "var-locks-brick". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 09:00:40 crc kubenswrapper[4813]: I1201 09:00:40.210554 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c2d1fc05-984a-4216-b9f5-befc2acb69db-dev" (OuterVolumeSpecName: "dev") pod "c2d1fc05-984a-4216-b9f5-befc2acb69db" (UID: "c2d1fc05-984a-4216-b9f5-befc2acb69db"). InnerVolumeSpecName "dev". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 09:00:40 crc kubenswrapper[4813]: I1201 09:00:40.210822 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c2d1fc05-984a-4216-b9f5-befc2acb69db-logs" (OuterVolumeSpecName: "logs") pod "c2d1fc05-984a-4216-b9f5-befc2acb69db" (UID: "c2d1fc05-984a-4216-b9f5-befc2acb69db"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 09:00:40 crc kubenswrapper[4813]: I1201 09:00:40.210894 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c2d1fc05-984a-4216-b9f5-befc2acb69db-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "c2d1fc05-984a-4216-b9f5-befc2acb69db" (UID: "c2d1fc05-984a-4216-b9f5-befc2acb69db"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 09:00:40 crc kubenswrapper[4813]: I1201 09:00:40.214592 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c2d1fc05-984a-4216-b9f5-befc2acb69db-scripts" (OuterVolumeSpecName: "scripts") pod "c2d1fc05-984a-4216-b9f5-befc2acb69db" (UID: "c2d1fc05-984a-4216-b9f5-befc2acb69db"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:00:40 crc kubenswrapper[4813]: I1201 09:00:40.215122 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c2d1fc05-984a-4216-b9f5-befc2acb69db-kube-api-access-djj2r" (OuterVolumeSpecName: "kube-api-access-djj2r") pod "c2d1fc05-984a-4216-b9f5-befc2acb69db" (UID: "c2d1fc05-984a-4216-b9f5-befc2acb69db"). InnerVolumeSpecName "kube-api-access-djj2r". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:00:40 crc kubenswrapper[4813]: I1201 09:00:40.215192 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage17-crc" (OuterVolumeSpecName: "glance-cache") pod "c2d1fc05-984a-4216-b9f5-befc2acb69db" (UID: "c2d1fc05-984a-4216-b9f5-befc2acb69db"). InnerVolumeSpecName "local-storage17-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 01 09:00:40 crc kubenswrapper[4813]: I1201 09:00:40.222038 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage01-crc" (OuterVolumeSpecName: "glance") pod "c2d1fc05-984a-4216-b9f5-befc2acb69db" (UID: "c2d1fc05-984a-4216-b9f5-befc2acb69db"). InnerVolumeSpecName "local-storage01-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 01 09:00:40 crc kubenswrapper[4813]: I1201 09:00:40.283384 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c2d1fc05-984a-4216-b9f5-befc2acb69db-config-data" (OuterVolumeSpecName: "config-data") pod "c2d1fc05-984a-4216-b9f5-befc2acb69db" (UID: "c2d1fc05-984a-4216-b9f5-befc2acb69db"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:00:40 crc kubenswrapper[4813]: I1201 09:00:40.311525 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-djj2r\" (UniqueName: \"kubernetes.io/projected/c2d1fc05-984a-4216-b9f5-befc2acb69db-kube-api-access-djj2r\") on node \"crc\" DevicePath \"\"" Dec 01 09:00:40 crc kubenswrapper[4813]: I1201 09:00:40.311568 4813 reconciler_common.go:293] "Volume detached for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/c2d1fc05-984a-4216-b9f5-befc2acb69db-sys\") on node \"crc\" DevicePath \"\"" Dec 01 09:00:40 crc kubenswrapper[4813]: I1201 09:00:40.311604 4813 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c2d1fc05-984a-4216-b9f5-befc2acb69db-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 09:00:40 crc kubenswrapper[4813]: I1201 09:00:40.311614 4813 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/c2d1fc05-984a-4216-b9f5-befc2acb69db-lib-modules\") on node \"crc\" DevicePath \"\"" Dec 01 09:00:40 crc kubenswrapper[4813]: I1201 09:00:40.311658 4813 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage17-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage17-crc\") on node \"crc\" " Dec 01 09:00:40 crc kubenswrapper[4813]: I1201 09:00:40.311668 4813 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c2d1fc05-984a-4216-b9f5-befc2acb69db-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 09:00:40 crc kubenswrapper[4813]: I1201 09:00:40.311677 4813 reconciler_common.go:293] "Volume detached for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/c2d1fc05-984a-4216-b9f5-befc2acb69db-dev\") on node \"crc\" DevicePath \"\"" Dec 01 09:00:40 crc kubenswrapper[4813]: I1201 09:00:40.311685 4813 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/c2d1fc05-984a-4216-b9f5-befc2acb69db-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 01 09:00:40 crc kubenswrapper[4813]: I1201 09:00:40.311694 4813 reconciler_common.go:293] "Volume detached for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/c2d1fc05-984a-4216-b9f5-befc2acb69db-var-locks-brick\") on node \"crc\" DevicePath \"\"" Dec 01 09:00:40 crc kubenswrapper[4813]: I1201 09:00:40.311703 4813 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c2d1fc05-984a-4216-b9f5-befc2acb69db-logs\") on node \"crc\" DevicePath \"\"" Dec 01 09:00:40 crc kubenswrapper[4813]: I1201 09:00:40.311715 4813 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" " Dec 01 09:00:40 crc kubenswrapper[4813]: I1201 09:00:40.324748 4813 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage01-crc" (UniqueName: "kubernetes.io/local-volume/local-storage01-crc") on node "crc" Dec 01 09:00:40 crc kubenswrapper[4813]: I1201 09:00:40.324796 4813 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage17-crc" (UniqueName: "kubernetes.io/local-volume/local-storage17-crc") on node "crc" Dec 01 09:00:40 crc kubenswrapper[4813]: I1201 09:00:40.413259 4813 reconciler_common.go:293] "Volume detached for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" DevicePath \"\"" Dec 01 09:00:40 crc kubenswrapper[4813]: I1201 09:00:40.413312 4813 reconciler_common.go:293] "Volume detached for volume \"local-storage17-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage17-crc\") on node \"crc\" DevicePath \"\"" Dec 01 09:00:40 crc kubenswrapper[4813]: I1201 09:00:40.676366 4813 generic.go:334] "Generic (PLEG): container finished" podID="c2d1fc05-984a-4216-b9f5-befc2acb69db" containerID="9cd9e2a2e5957ab1447d39e837d5ad9e6577d09b9acf973b948ddb61cd0d004e" exitCode=143 Dec 01 09:00:40 crc kubenswrapper[4813]: I1201 09:00:40.676425 4813 generic.go:334] "Generic (PLEG): container finished" podID="c2d1fc05-984a-4216-b9f5-befc2acb69db" containerID="b66651a85c9e7d2664792d33c59df62688ba872ef0c03fde20d3a6e69e08b9b4" exitCode=143 Dec 01 09:00:40 crc kubenswrapper[4813]: I1201 09:00:40.676454 4813 generic.go:334] "Generic (PLEG): container finished" podID="c2d1fc05-984a-4216-b9f5-befc2acb69db" containerID="82204e399295940f7898ab7236a95fe5ab0cf118e7275d729cf375d3f2aae2f8" exitCode=143 Dec 01 09:00:40 crc kubenswrapper[4813]: I1201 09:00:40.676443 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"c2d1fc05-984a-4216-b9f5-befc2acb69db","Type":"ContainerDied","Data":"9cd9e2a2e5957ab1447d39e837d5ad9e6577d09b9acf973b948ddb61cd0d004e"} Dec 01 09:00:40 crc kubenswrapper[4813]: I1201 09:00:40.676515 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"c2d1fc05-984a-4216-b9f5-befc2acb69db","Type":"ContainerDied","Data":"b66651a85c9e7d2664792d33c59df62688ba872ef0c03fde20d3a6e69e08b9b4"} Dec 01 09:00:40 crc kubenswrapper[4813]: I1201 09:00:40.676529 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"c2d1fc05-984a-4216-b9f5-befc2acb69db","Type":"ContainerDied","Data":"82204e399295940f7898ab7236a95fe5ab0cf118e7275d729cf375d3f2aae2f8"} Dec 01 09:00:40 crc kubenswrapper[4813]: I1201 09:00:40.676554 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"c2d1fc05-984a-4216-b9f5-befc2acb69db","Type":"ContainerDied","Data":"5d60339f0374ce44dab9a713f80d49cec7c1bd2bab44fdccf475652e79854f1c"} Dec 01 09:00:40 crc kubenswrapper[4813]: I1201 09:00:40.676575 4813 scope.go:117] "RemoveContainer" containerID="9cd9e2a2e5957ab1447d39e837d5ad9e6577d09b9acf973b948ddb61cd0d004e" Dec 01 09:00:40 crc kubenswrapper[4813]: I1201 09:00:40.677129 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:00:40 crc kubenswrapper[4813]: I1201 09:00:40.711336 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Dec 01 09:00:40 crc kubenswrapper[4813]: I1201 09:00:40.714416 4813 scope.go:117] "RemoveContainer" containerID="b66651a85c9e7d2664792d33c59df62688ba872ef0c03fde20d3a6e69e08b9b4" Dec 01 09:00:40 crc kubenswrapper[4813]: I1201 09:00:40.728108 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Dec 01 09:00:40 crc kubenswrapper[4813]: I1201 09:00:40.735824 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Dec 01 09:00:40 crc kubenswrapper[4813]: E1201 09:00:40.736216 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c2d1fc05-984a-4216-b9f5-befc2acb69db" containerName="glance-httpd" Dec 01 09:00:40 crc kubenswrapper[4813]: I1201 09:00:40.736233 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="c2d1fc05-984a-4216-b9f5-befc2acb69db" containerName="glance-httpd" Dec 01 09:00:40 crc kubenswrapper[4813]: E1201 09:00:40.736308 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c2d1fc05-984a-4216-b9f5-befc2acb69db" containerName="glance-api" Dec 01 09:00:40 crc kubenswrapper[4813]: I1201 09:00:40.736315 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="c2d1fc05-984a-4216-b9f5-befc2acb69db" containerName="glance-api" Dec 01 09:00:40 crc kubenswrapper[4813]: E1201 09:00:40.736333 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c2d1fc05-984a-4216-b9f5-befc2acb69db" containerName="glance-log" Dec 01 09:00:40 crc kubenswrapper[4813]: I1201 09:00:40.736339 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="c2d1fc05-984a-4216-b9f5-befc2acb69db" containerName="glance-log" Dec 01 09:00:40 crc kubenswrapper[4813]: I1201 09:00:40.736461 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="c2d1fc05-984a-4216-b9f5-befc2acb69db" containerName="glance-api" Dec 01 09:00:40 crc kubenswrapper[4813]: I1201 09:00:40.736479 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="c2d1fc05-984a-4216-b9f5-befc2acb69db" containerName="glance-httpd" Dec 01 09:00:40 crc kubenswrapper[4813]: I1201 09:00:40.736493 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="c2d1fc05-984a-4216-b9f5-befc2acb69db" containerName="glance-log" Dec 01 09:00:40 crc kubenswrapper[4813]: I1201 09:00:40.737492 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:00:40 crc kubenswrapper[4813]: I1201 09:00:40.739934 4813 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-default-internal-config-data" Dec 01 09:00:40 crc kubenswrapper[4813]: I1201 09:00:40.754002 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Dec 01 09:00:40 crc kubenswrapper[4813]: I1201 09:00:40.759824 4813 scope.go:117] "RemoveContainer" containerID="82204e399295940f7898ab7236a95fe5ab0cf118e7275d729cf375d3f2aae2f8" Dec 01 09:00:40 crc kubenswrapper[4813]: I1201 09:00:40.785105 4813 scope.go:117] "RemoveContainer" containerID="9cd9e2a2e5957ab1447d39e837d5ad9e6577d09b9acf973b948ddb61cd0d004e" Dec 01 09:00:40 crc kubenswrapper[4813]: E1201 09:00:40.787600 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9cd9e2a2e5957ab1447d39e837d5ad9e6577d09b9acf973b948ddb61cd0d004e\": container with ID starting with 9cd9e2a2e5957ab1447d39e837d5ad9e6577d09b9acf973b948ddb61cd0d004e not found: ID does not exist" containerID="9cd9e2a2e5957ab1447d39e837d5ad9e6577d09b9acf973b948ddb61cd0d004e" Dec 01 09:00:40 crc kubenswrapper[4813]: I1201 09:00:40.787702 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9cd9e2a2e5957ab1447d39e837d5ad9e6577d09b9acf973b948ddb61cd0d004e"} err="failed to get container status \"9cd9e2a2e5957ab1447d39e837d5ad9e6577d09b9acf973b948ddb61cd0d004e\": rpc error: code = NotFound desc = could not find container \"9cd9e2a2e5957ab1447d39e837d5ad9e6577d09b9acf973b948ddb61cd0d004e\": container with ID starting with 9cd9e2a2e5957ab1447d39e837d5ad9e6577d09b9acf973b948ddb61cd0d004e not found: ID does not exist" Dec 01 09:00:40 crc kubenswrapper[4813]: I1201 09:00:40.787789 4813 scope.go:117] "RemoveContainer" containerID="b66651a85c9e7d2664792d33c59df62688ba872ef0c03fde20d3a6e69e08b9b4" Dec 01 09:00:40 crc kubenswrapper[4813]: E1201 09:00:40.788252 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b66651a85c9e7d2664792d33c59df62688ba872ef0c03fde20d3a6e69e08b9b4\": container with ID starting with b66651a85c9e7d2664792d33c59df62688ba872ef0c03fde20d3a6e69e08b9b4 not found: ID does not exist" containerID="b66651a85c9e7d2664792d33c59df62688ba872ef0c03fde20d3a6e69e08b9b4" Dec 01 09:00:40 crc kubenswrapper[4813]: I1201 09:00:40.788316 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b66651a85c9e7d2664792d33c59df62688ba872ef0c03fde20d3a6e69e08b9b4"} err="failed to get container status \"b66651a85c9e7d2664792d33c59df62688ba872ef0c03fde20d3a6e69e08b9b4\": rpc error: code = NotFound desc = could not find container \"b66651a85c9e7d2664792d33c59df62688ba872ef0c03fde20d3a6e69e08b9b4\": container with ID starting with b66651a85c9e7d2664792d33c59df62688ba872ef0c03fde20d3a6e69e08b9b4 not found: ID does not exist" Dec 01 09:00:40 crc kubenswrapper[4813]: I1201 09:00:40.788350 4813 scope.go:117] "RemoveContainer" containerID="82204e399295940f7898ab7236a95fe5ab0cf118e7275d729cf375d3f2aae2f8" Dec 01 09:00:40 crc kubenswrapper[4813]: E1201 09:00:40.788634 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"82204e399295940f7898ab7236a95fe5ab0cf118e7275d729cf375d3f2aae2f8\": container with ID starting with 82204e399295940f7898ab7236a95fe5ab0cf118e7275d729cf375d3f2aae2f8 not found: ID does not exist" containerID="82204e399295940f7898ab7236a95fe5ab0cf118e7275d729cf375d3f2aae2f8" Dec 01 09:00:40 crc kubenswrapper[4813]: I1201 09:00:40.788673 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"82204e399295940f7898ab7236a95fe5ab0cf118e7275d729cf375d3f2aae2f8"} err="failed to get container status \"82204e399295940f7898ab7236a95fe5ab0cf118e7275d729cf375d3f2aae2f8\": rpc error: code = NotFound desc = could not find container \"82204e399295940f7898ab7236a95fe5ab0cf118e7275d729cf375d3f2aae2f8\": container with ID starting with 82204e399295940f7898ab7236a95fe5ab0cf118e7275d729cf375d3f2aae2f8 not found: ID does not exist" Dec 01 09:00:40 crc kubenswrapper[4813]: I1201 09:00:40.788706 4813 scope.go:117] "RemoveContainer" containerID="9cd9e2a2e5957ab1447d39e837d5ad9e6577d09b9acf973b948ddb61cd0d004e" Dec 01 09:00:40 crc kubenswrapper[4813]: I1201 09:00:40.789026 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9cd9e2a2e5957ab1447d39e837d5ad9e6577d09b9acf973b948ddb61cd0d004e"} err="failed to get container status \"9cd9e2a2e5957ab1447d39e837d5ad9e6577d09b9acf973b948ddb61cd0d004e\": rpc error: code = NotFound desc = could not find container \"9cd9e2a2e5957ab1447d39e837d5ad9e6577d09b9acf973b948ddb61cd0d004e\": container with ID starting with 9cd9e2a2e5957ab1447d39e837d5ad9e6577d09b9acf973b948ddb61cd0d004e not found: ID does not exist" Dec 01 09:00:40 crc kubenswrapper[4813]: I1201 09:00:40.789066 4813 scope.go:117] "RemoveContainer" containerID="b66651a85c9e7d2664792d33c59df62688ba872ef0c03fde20d3a6e69e08b9b4" Dec 01 09:00:40 crc kubenswrapper[4813]: I1201 09:00:40.789518 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b66651a85c9e7d2664792d33c59df62688ba872ef0c03fde20d3a6e69e08b9b4"} err="failed to get container status \"b66651a85c9e7d2664792d33c59df62688ba872ef0c03fde20d3a6e69e08b9b4\": rpc error: code = NotFound desc = could not find container \"b66651a85c9e7d2664792d33c59df62688ba872ef0c03fde20d3a6e69e08b9b4\": container with ID starting with b66651a85c9e7d2664792d33c59df62688ba872ef0c03fde20d3a6e69e08b9b4 not found: ID does not exist" Dec 01 09:00:40 crc kubenswrapper[4813]: I1201 09:00:40.789545 4813 scope.go:117] "RemoveContainer" containerID="82204e399295940f7898ab7236a95fe5ab0cf118e7275d729cf375d3f2aae2f8" Dec 01 09:00:40 crc kubenswrapper[4813]: I1201 09:00:40.789794 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"82204e399295940f7898ab7236a95fe5ab0cf118e7275d729cf375d3f2aae2f8"} err="failed to get container status \"82204e399295940f7898ab7236a95fe5ab0cf118e7275d729cf375d3f2aae2f8\": rpc error: code = NotFound desc = could not find container \"82204e399295940f7898ab7236a95fe5ab0cf118e7275d729cf375d3f2aae2f8\": container with ID starting with 82204e399295940f7898ab7236a95fe5ab0cf118e7275d729cf375d3f2aae2f8 not found: ID does not exist" Dec 01 09:00:40 crc kubenswrapper[4813]: I1201 09:00:40.789836 4813 scope.go:117] "RemoveContainer" containerID="9cd9e2a2e5957ab1447d39e837d5ad9e6577d09b9acf973b948ddb61cd0d004e" Dec 01 09:00:40 crc kubenswrapper[4813]: I1201 09:00:40.790294 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9cd9e2a2e5957ab1447d39e837d5ad9e6577d09b9acf973b948ddb61cd0d004e"} err="failed to get container status \"9cd9e2a2e5957ab1447d39e837d5ad9e6577d09b9acf973b948ddb61cd0d004e\": rpc error: code = NotFound desc = could not find container \"9cd9e2a2e5957ab1447d39e837d5ad9e6577d09b9acf973b948ddb61cd0d004e\": container with ID starting with 9cd9e2a2e5957ab1447d39e837d5ad9e6577d09b9acf973b948ddb61cd0d004e not found: ID does not exist" Dec 01 09:00:40 crc kubenswrapper[4813]: I1201 09:00:40.790394 4813 scope.go:117] "RemoveContainer" containerID="b66651a85c9e7d2664792d33c59df62688ba872ef0c03fde20d3a6e69e08b9b4" Dec 01 09:00:40 crc kubenswrapper[4813]: I1201 09:00:40.790677 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b66651a85c9e7d2664792d33c59df62688ba872ef0c03fde20d3a6e69e08b9b4"} err="failed to get container status \"b66651a85c9e7d2664792d33c59df62688ba872ef0c03fde20d3a6e69e08b9b4\": rpc error: code = NotFound desc = could not find container \"b66651a85c9e7d2664792d33c59df62688ba872ef0c03fde20d3a6e69e08b9b4\": container with ID starting with b66651a85c9e7d2664792d33c59df62688ba872ef0c03fde20d3a6e69e08b9b4 not found: ID does not exist" Dec 01 09:00:40 crc kubenswrapper[4813]: I1201 09:00:40.790709 4813 scope.go:117] "RemoveContainer" containerID="82204e399295940f7898ab7236a95fe5ab0cf118e7275d729cf375d3f2aae2f8" Dec 01 09:00:40 crc kubenswrapper[4813]: I1201 09:00:40.791320 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"82204e399295940f7898ab7236a95fe5ab0cf118e7275d729cf375d3f2aae2f8"} err="failed to get container status \"82204e399295940f7898ab7236a95fe5ab0cf118e7275d729cf375d3f2aae2f8\": rpc error: code = NotFound desc = could not find container \"82204e399295940f7898ab7236a95fe5ab0cf118e7275d729cf375d3f2aae2f8\": container with ID starting with 82204e399295940f7898ab7236a95fe5ab0cf118e7275d729cf375d3f2aae2f8 not found: ID does not exist" Dec 01 09:00:40 crc kubenswrapper[4813]: I1201 09:00:40.819816 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/5e438749-e7d5-4ca0-98e1-6d3319286720-etc-iscsi\") pod \"glance-default-internal-api-0\" (UID: \"5e438749-e7d5-4ca0-98e1-6d3319286720\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:00:40 crc kubenswrapper[4813]: I1201 09:00:40.819874 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/5e438749-e7d5-4ca0-98e1-6d3319286720-var-locks-brick\") pod \"glance-default-internal-api-0\" (UID: \"5e438749-e7d5-4ca0-98e1-6d3319286720\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:00:40 crc kubenswrapper[4813]: I1201 09:00:40.820081 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/5e438749-e7d5-4ca0-98e1-6d3319286720-run\") pod \"glance-default-internal-api-0\" (UID: \"5e438749-e7d5-4ca0-98e1-6d3319286720\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:00:40 crc kubenswrapper[4813]: I1201 09:00:40.820179 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/5e438749-e7d5-4ca0-98e1-6d3319286720-sys\") pod \"glance-default-internal-api-0\" (UID: \"5e438749-e7d5-4ca0-98e1-6d3319286720\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:00:40 crc kubenswrapper[4813]: I1201 09:00:40.820371 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/5e438749-e7d5-4ca0-98e1-6d3319286720-etc-nvme\") pod \"glance-default-internal-api-0\" (UID: \"5e438749-e7d5-4ca0-98e1-6d3319286720\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:00:40 crc kubenswrapper[4813]: I1201 09:00:40.820402 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5e438749-e7d5-4ca0-98e1-6d3319286720-scripts\") pod \"glance-default-internal-api-0\" (UID: \"5e438749-e7d5-4ca0-98e1-6d3319286720\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:00:40 crc kubenswrapper[4813]: I1201 09:00:40.820469 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-internal-api-0\" (UID: \"5e438749-e7d5-4ca0-98e1-6d3319286720\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:00:40 crc kubenswrapper[4813]: I1201 09:00:40.820520 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage17-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage17-crc\") pod \"glance-default-internal-api-0\" (UID: \"5e438749-e7d5-4ca0-98e1-6d3319286720\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:00:40 crc kubenswrapper[4813]: I1201 09:00:40.820564 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s2gqr\" (UniqueName: \"kubernetes.io/projected/5e438749-e7d5-4ca0-98e1-6d3319286720-kube-api-access-s2gqr\") pod \"glance-default-internal-api-0\" (UID: \"5e438749-e7d5-4ca0-98e1-6d3319286720\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:00:40 crc kubenswrapper[4813]: I1201 09:00:40.820594 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5e438749-e7d5-4ca0-98e1-6d3319286720-config-data\") pod \"glance-default-internal-api-0\" (UID: \"5e438749-e7d5-4ca0-98e1-6d3319286720\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:00:40 crc kubenswrapper[4813]: I1201 09:00:40.820616 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5e438749-e7d5-4ca0-98e1-6d3319286720-logs\") pod \"glance-default-internal-api-0\" (UID: \"5e438749-e7d5-4ca0-98e1-6d3319286720\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:00:40 crc kubenswrapper[4813]: I1201 09:00:40.820680 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/5e438749-e7d5-4ca0-98e1-6d3319286720-lib-modules\") pod \"glance-default-internal-api-0\" (UID: \"5e438749-e7d5-4ca0-98e1-6d3319286720\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:00:40 crc kubenswrapper[4813]: I1201 09:00:40.820735 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5e438749-e7d5-4ca0-98e1-6d3319286720-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"5e438749-e7d5-4ca0-98e1-6d3319286720\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:00:40 crc kubenswrapper[4813]: I1201 09:00:40.820774 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/5e438749-e7d5-4ca0-98e1-6d3319286720-dev\") pod \"glance-default-internal-api-0\" (UID: \"5e438749-e7d5-4ca0-98e1-6d3319286720\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:00:40 crc kubenswrapper[4813]: I1201 09:00:40.922316 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/5e438749-e7d5-4ca0-98e1-6d3319286720-etc-nvme\") pod \"glance-default-internal-api-0\" (UID: \"5e438749-e7d5-4ca0-98e1-6d3319286720\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:00:40 crc kubenswrapper[4813]: I1201 09:00:40.922670 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5e438749-e7d5-4ca0-98e1-6d3319286720-scripts\") pod \"glance-default-internal-api-0\" (UID: \"5e438749-e7d5-4ca0-98e1-6d3319286720\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:00:40 crc kubenswrapper[4813]: I1201 09:00:40.922704 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-internal-api-0\" (UID: \"5e438749-e7d5-4ca0-98e1-6d3319286720\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:00:40 crc kubenswrapper[4813]: I1201 09:00:40.922726 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage17-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage17-crc\") pod \"glance-default-internal-api-0\" (UID: \"5e438749-e7d5-4ca0-98e1-6d3319286720\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:00:40 crc kubenswrapper[4813]: I1201 09:00:40.922750 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2gqr\" (UniqueName: \"kubernetes.io/projected/5e438749-e7d5-4ca0-98e1-6d3319286720-kube-api-access-s2gqr\") pod \"glance-default-internal-api-0\" (UID: \"5e438749-e7d5-4ca0-98e1-6d3319286720\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:00:40 crc kubenswrapper[4813]: I1201 09:00:40.922769 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5e438749-e7d5-4ca0-98e1-6d3319286720-config-data\") pod \"glance-default-internal-api-0\" (UID: \"5e438749-e7d5-4ca0-98e1-6d3319286720\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:00:40 crc kubenswrapper[4813]: I1201 09:00:40.922784 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5e438749-e7d5-4ca0-98e1-6d3319286720-logs\") pod \"glance-default-internal-api-0\" (UID: \"5e438749-e7d5-4ca0-98e1-6d3319286720\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:00:40 crc kubenswrapper[4813]: I1201 09:00:40.922809 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/5e438749-e7d5-4ca0-98e1-6d3319286720-lib-modules\") pod \"glance-default-internal-api-0\" (UID: \"5e438749-e7d5-4ca0-98e1-6d3319286720\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:00:40 crc kubenswrapper[4813]: I1201 09:00:40.922834 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5e438749-e7d5-4ca0-98e1-6d3319286720-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"5e438749-e7d5-4ca0-98e1-6d3319286720\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:00:40 crc kubenswrapper[4813]: I1201 09:00:40.922852 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/5e438749-e7d5-4ca0-98e1-6d3319286720-dev\") pod \"glance-default-internal-api-0\" (UID: \"5e438749-e7d5-4ca0-98e1-6d3319286720\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:00:40 crc kubenswrapper[4813]: I1201 09:00:40.922882 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/5e438749-e7d5-4ca0-98e1-6d3319286720-etc-iscsi\") pod \"glance-default-internal-api-0\" (UID: \"5e438749-e7d5-4ca0-98e1-6d3319286720\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:00:40 crc kubenswrapper[4813]: I1201 09:00:40.922904 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/5e438749-e7d5-4ca0-98e1-6d3319286720-var-locks-brick\") pod \"glance-default-internal-api-0\" (UID: \"5e438749-e7d5-4ca0-98e1-6d3319286720\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:00:40 crc kubenswrapper[4813]: I1201 09:00:40.922922 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/5e438749-e7d5-4ca0-98e1-6d3319286720-run\") pod \"glance-default-internal-api-0\" (UID: \"5e438749-e7d5-4ca0-98e1-6d3319286720\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:00:40 crc kubenswrapper[4813]: I1201 09:00:40.922943 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/5e438749-e7d5-4ca0-98e1-6d3319286720-sys\") pod \"glance-default-internal-api-0\" (UID: \"5e438749-e7d5-4ca0-98e1-6d3319286720\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:00:40 crc kubenswrapper[4813]: I1201 09:00:40.923048 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/5e438749-e7d5-4ca0-98e1-6d3319286720-sys\") pod \"glance-default-internal-api-0\" (UID: \"5e438749-e7d5-4ca0-98e1-6d3319286720\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:00:40 crc kubenswrapper[4813]: I1201 09:00:40.922451 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/5e438749-e7d5-4ca0-98e1-6d3319286720-etc-nvme\") pod \"glance-default-internal-api-0\" (UID: \"5e438749-e7d5-4ca0-98e1-6d3319286720\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:00:40 crc kubenswrapper[4813]: I1201 09:00:40.923592 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/5e438749-e7d5-4ca0-98e1-6d3319286720-lib-modules\") pod \"glance-default-internal-api-0\" (UID: \"5e438749-e7d5-4ca0-98e1-6d3319286720\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:00:40 crc kubenswrapper[4813]: I1201 09:00:40.923706 4813 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-internal-api-0\" (UID: \"5e438749-e7d5-4ca0-98e1-6d3319286720\") device mount path \"/mnt/openstack/pv01\"" pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:00:40 crc kubenswrapper[4813]: I1201 09:00:40.924070 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/5e438749-e7d5-4ca0-98e1-6d3319286720-etc-iscsi\") pod \"glance-default-internal-api-0\" (UID: \"5e438749-e7d5-4ca0-98e1-6d3319286720\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:00:40 crc kubenswrapper[4813]: I1201 09:00:40.924131 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/5e438749-e7d5-4ca0-98e1-6d3319286720-dev\") pod \"glance-default-internal-api-0\" (UID: \"5e438749-e7d5-4ca0-98e1-6d3319286720\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:00:40 crc kubenswrapper[4813]: I1201 09:00:40.924445 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/5e438749-e7d5-4ca0-98e1-6d3319286720-var-locks-brick\") pod \"glance-default-internal-api-0\" (UID: \"5e438749-e7d5-4ca0-98e1-6d3319286720\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:00:40 crc kubenswrapper[4813]: I1201 09:00:40.924472 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/5e438749-e7d5-4ca0-98e1-6d3319286720-run\") pod \"glance-default-internal-api-0\" (UID: \"5e438749-e7d5-4ca0-98e1-6d3319286720\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:00:40 crc kubenswrapper[4813]: I1201 09:00:40.924600 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5e438749-e7d5-4ca0-98e1-6d3319286720-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"5e438749-e7d5-4ca0-98e1-6d3319286720\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:00:40 crc kubenswrapper[4813]: I1201 09:00:40.925995 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5e438749-e7d5-4ca0-98e1-6d3319286720-logs\") pod \"glance-default-internal-api-0\" (UID: \"5e438749-e7d5-4ca0-98e1-6d3319286720\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:00:40 crc kubenswrapper[4813]: I1201 09:00:40.926204 4813 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage17-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage17-crc\") pod \"glance-default-internal-api-0\" (UID: \"5e438749-e7d5-4ca0-98e1-6d3319286720\") device mount path \"/mnt/openstack/pv17\"" pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:00:40 crc kubenswrapper[4813]: I1201 09:00:40.928845 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5e438749-e7d5-4ca0-98e1-6d3319286720-config-data\") pod \"glance-default-internal-api-0\" (UID: \"5e438749-e7d5-4ca0-98e1-6d3319286720\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:00:40 crc kubenswrapper[4813]: I1201 09:00:40.942211 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2gqr\" (UniqueName: \"kubernetes.io/projected/5e438749-e7d5-4ca0-98e1-6d3319286720-kube-api-access-s2gqr\") pod \"glance-default-internal-api-0\" (UID: \"5e438749-e7d5-4ca0-98e1-6d3319286720\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:00:40 crc kubenswrapper[4813]: I1201 09:00:40.945088 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5e438749-e7d5-4ca0-98e1-6d3319286720-scripts\") pod \"glance-default-internal-api-0\" (UID: \"5e438749-e7d5-4ca0-98e1-6d3319286720\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:00:40 crc kubenswrapper[4813]: I1201 09:00:40.945131 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-internal-api-0\" (UID: \"5e438749-e7d5-4ca0-98e1-6d3319286720\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:00:40 crc kubenswrapper[4813]: I1201 09:00:40.947999 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage17-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage17-crc\") pod \"glance-default-internal-api-0\" (UID: \"5e438749-e7d5-4ca0-98e1-6d3319286720\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:00:41 crc kubenswrapper[4813]: I1201 09:00:41.063902 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:00:41 crc kubenswrapper[4813]: I1201 09:00:41.492530 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Dec 01 09:00:41 crc kubenswrapper[4813]: I1201 09:00:41.686086 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"5e438749-e7d5-4ca0-98e1-6d3319286720","Type":"ContainerStarted","Data":"dcb779b0095ccf2057b4fa7f52ea855e49228c7d0c78ae4beaafc5c8e1f860ba"} Dec 01 09:00:41 crc kubenswrapper[4813]: I1201 09:00:41.686385 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"5e438749-e7d5-4ca0-98e1-6d3319286720","Type":"ContainerStarted","Data":"2307171a9b2640da2b9a255b325670d959a506be54a9a7da454bee9bc5d3e61d"} Dec 01 09:00:42 crc kubenswrapper[4813]: I1201 09:00:42.405679 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c2d1fc05-984a-4216-b9f5-befc2acb69db" path="/var/lib/kubelet/pods/c2d1fc05-984a-4216-b9f5-befc2acb69db/volumes" Dec 01 09:00:42 crc kubenswrapper[4813]: I1201 09:00:42.702727 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"5e438749-e7d5-4ca0-98e1-6d3319286720","Type":"ContainerStarted","Data":"6604eac2164f12eccf3a2895e8fbaf218ce287738434acbba0a5cadfe7fcc82f"} Dec 01 09:00:42 crc kubenswrapper[4813]: I1201 09:00:42.703706 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"5e438749-e7d5-4ca0-98e1-6d3319286720","Type":"ContainerStarted","Data":"0fba4493026f7e0bb5c77eac41f7dd9e9be04fb93081af5767525ccad569ecc3"} Dec 01 09:00:42 crc kubenswrapper[4813]: I1201 09:00:42.761411 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-default-internal-api-0" podStartSLOduration=2.761378025 podStartE2EDuration="2.761378025s" podCreationTimestamp="2025-12-01 09:00:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 09:00:42.742205747 +0000 UTC m=+1227.365127363" watchObservedRunningTime="2025-12-01 09:00:42.761378025 +0000 UTC m=+1227.384299651" Dec 01 09:00:47 crc kubenswrapper[4813]: I1201 09:00:47.222370 4813 patch_prober.go:28] interesting pod/machine-config-daemon-9wn4q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 09:00:47 crc kubenswrapper[4813]: I1201 09:00:47.223066 4813 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9wn4q" podUID="3d423402-e08f-42e9-b60c-6eb9bf2f6d48" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 09:00:48 crc kubenswrapper[4813]: I1201 09:00:48.005357 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-external-api-0" Dec 01 09:00:48 crc kubenswrapper[4813]: I1201 09:00:48.005742 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-external-api-0" Dec 01 09:00:48 crc kubenswrapper[4813]: I1201 09:00:48.005767 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-external-api-0" Dec 01 09:00:48 crc kubenswrapper[4813]: I1201 09:00:48.035911 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-external-api-0" Dec 01 09:00:48 crc kubenswrapper[4813]: I1201 09:00:48.037729 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-external-api-0" Dec 01 09:00:48 crc kubenswrapper[4813]: I1201 09:00:48.054665 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-external-api-0" Dec 01 09:00:48 crc kubenswrapper[4813]: I1201 09:00:48.759875 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-external-api-0" Dec 01 09:00:48 crc kubenswrapper[4813]: I1201 09:00:48.759927 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-external-api-0" Dec 01 09:00:48 crc kubenswrapper[4813]: I1201 09:00:48.759941 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-external-api-0" Dec 01 09:00:48 crc kubenswrapper[4813]: I1201 09:00:48.777594 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-external-api-0" Dec 01 09:00:48 crc kubenswrapper[4813]: I1201 09:00:48.779527 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-external-api-0" Dec 01 09:00:48 crc kubenswrapper[4813]: I1201 09:00:48.787587 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-external-api-0" Dec 01 09:00:51 crc kubenswrapper[4813]: I1201 09:00:51.065348 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:00:51 crc kubenswrapper[4813]: I1201 09:00:51.065834 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:00:51 crc kubenswrapper[4813]: I1201 09:00:51.065850 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:00:51 crc kubenswrapper[4813]: I1201 09:00:51.103541 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:00:51 crc kubenswrapper[4813]: I1201 09:00:51.113924 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:00:51 crc kubenswrapper[4813]: I1201 09:00:51.121250 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:00:51 crc kubenswrapper[4813]: I1201 09:00:51.790137 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:00:51 crc kubenswrapper[4813]: I1201 09:00:51.790561 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:00:51 crc kubenswrapper[4813]: I1201 09:00:51.790582 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:00:51 crc kubenswrapper[4813]: I1201 09:00:51.811231 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:00:51 crc kubenswrapper[4813]: I1201 09:00:51.812331 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:00:51 crc kubenswrapper[4813]: I1201 09:00:51.819201 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:00:54 crc kubenswrapper[4813]: I1201 09:00:54.320902 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-external-api-2"] Dec 01 09:00:54 crc kubenswrapper[4813]: I1201 09:00:54.323865 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-2" Dec 01 09:00:54 crc kubenswrapper[4813]: I1201 09:00:54.328143 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-external-api-1"] Dec 01 09:00:54 crc kubenswrapper[4813]: I1201 09:00:54.330206 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-1" Dec 01 09:00:54 crc kubenswrapper[4813]: I1201 09:00:54.354049 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-1"] Dec 01 09:00:54 crc kubenswrapper[4813]: I1201 09:00:54.386178 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-2"] Dec 01 09:00:54 crc kubenswrapper[4813]: I1201 09:00:54.436160 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-1"] Dec 01 09:00:54 crc kubenswrapper[4813]: I1201 09:00:54.437806 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-1" Dec 01 09:00:54 crc kubenswrapper[4813]: I1201 09:00:54.449445 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-2"] Dec 01 09:00:54 crc kubenswrapper[4813]: I1201 09:00:54.451951 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-2" Dec 01 09:00:54 crc kubenswrapper[4813]: I1201 09:00:54.456643 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-2"] Dec 01 09:00:54 crc kubenswrapper[4813]: I1201 09:00:54.470369 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-1"] Dec 01 09:00:54 crc kubenswrapper[4813]: I1201 09:00:54.480863 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/423d0710-1fe1-4cd3-99a8-3c32c7f08ee3-etc-iscsi\") pod \"glance-default-external-api-2\" (UID: \"423d0710-1fe1-4cd3-99a8-3c32c7f08ee3\") " pod="glance-kuttl-tests/glance-default-external-api-2" Dec 01 09:00:54 crc kubenswrapper[4813]: I1201 09:00:54.480941 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/423d0710-1fe1-4cd3-99a8-3c32c7f08ee3-dev\") pod \"glance-default-external-api-2\" (UID: \"423d0710-1fe1-4cd3-99a8-3c32c7f08ee3\") " pod="glance-kuttl-tests/glance-default-external-api-2" Dec 01 09:00:54 crc kubenswrapper[4813]: I1201 09:00:54.481135 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/423d0710-1fe1-4cd3-99a8-3c32c7f08ee3-sys\") pod \"glance-default-external-api-2\" (UID: \"423d0710-1fe1-4cd3-99a8-3c32c7f08ee3\") " pod="glance-kuttl-tests/glance-default-external-api-2" Dec 01 09:00:54 crc kubenswrapper[4813]: I1201 09:00:54.481195 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/04117ae2-817e-41a3-9d1e-608290bf79c8-config-data\") pod \"glance-default-external-api-1\" (UID: \"04117ae2-817e-41a3-9d1e-608290bf79c8\") " pod="glance-kuttl-tests/glance-default-external-api-1" Dec 01 09:00:54 crc kubenswrapper[4813]: I1201 09:00:54.481272 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/04117ae2-817e-41a3-9d1e-608290bf79c8-var-locks-brick\") pod \"glance-default-external-api-1\" (UID: \"04117ae2-817e-41a3-9d1e-608290bf79c8\") " pod="glance-kuttl-tests/glance-default-external-api-1" Dec 01 09:00:54 crc kubenswrapper[4813]: I1201 09:00:54.481327 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/423d0710-1fe1-4cd3-99a8-3c32c7f08ee3-scripts\") pod \"glance-default-external-api-2\" (UID: \"423d0710-1fe1-4cd3-99a8-3c32c7f08ee3\") " pod="glance-kuttl-tests/glance-default-external-api-2" Dec 01 09:00:54 crc kubenswrapper[4813]: I1201 09:00:54.481364 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kf7jh\" (UniqueName: \"kubernetes.io/projected/423d0710-1fe1-4cd3-99a8-3c32c7f08ee3-kube-api-access-kf7jh\") pod \"glance-default-external-api-2\" (UID: \"423d0710-1fe1-4cd3-99a8-3c32c7f08ee3\") " pod="glance-kuttl-tests/glance-default-external-api-2" Dec 01 09:00:54 crc kubenswrapper[4813]: I1201 09:00:54.481411 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-external-api-1\" (UID: \"04117ae2-817e-41a3-9d1e-608290bf79c8\") " pod="glance-kuttl-tests/glance-default-external-api-1" Dec 01 09:00:54 crc kubenswrapper[4813]: I1201 09:00:54.481447 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/423d0710-1fe1-4cd3-99a8-3c32c7f08ee3-lib-modules\") pod \"glance-default-external-api-2\" (UID: \"423d0710-1fe1-4cd3-99a8-3c32c7f08ee3\") " pod="glance-kuttl-tests/glance-default-external-api-2" Dec 01 09:00:54 crc kubenswrapper[4813]: I1201 09:00:54.481637 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/04117ae2-817e-41a3-9d1e-608290bf79c8-httpd-run\") pod \"glance-default-external-api-1\" (UID: \"04117ae2-817e-41a3-9d1e-608290bf79c8\") " pod="glance-kuttl-tests/glance-default-external-api-1" Dec 01 09:00:54 crc kubenswrapper[4813]: I1201 09:00:54.481889 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage14-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage14-crc\") pod \"glance-default-external-api-2\" (UID: \"423d0710-1fe1-4cd3-99a8-3c32c7f08ee3\") " pod="glance-kuttl-tests/glance-default-external-api-2" Dec 01 09:00:54 crc kubenswrapper[4813]: I1201 09:00:54.482136 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/04117ae2-817e-41a3-9d1e-608290bf79c8-run\") pod \"glance-default-external-api-1\" (UID: \"04117ae2-817e-41a3-9d1e-608290bf79c8\") " pod="glance-kuttl-tests/glance-default-external-api-1" Dec 01 09:00:54 crc kubenswrapper[4813]: I1201 09:00:54.482193 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/04117ae2-817e-41a3-9d1e-608290bf79c8-logs\") pod \"glance-default-external-api-1\" (UID: \"04117ae2-817e-41a3-9d1e-608290bf79c8\") " pod="glance-kuttl-tests/glance-default-external-api-1" Dec 01 09:00:54 crc kubenswrapper[4813]: I1201 09:00:54.482272 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/04117ae2-817e-41a3-9d1e-608290bf79c8-scripts\") pod \"glance-default-external-api-1\" (UID: \"04117ae2-817e-41a3-9d1e-608290bf79c8\") " pod="glance-kuttl-tests/glance-default-external-api-1" Dec 01 09:00:54 crc kubenswrapper[4813]: I1201 09:00:54.482358 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/04117ae2-817e-41a3-9d1e-608290bf79c8-dev\") pod \"glance-default-external-api-1\" (UID: \"04117ae2-817e-41a3-9d1e-608290bf79c8\") " pod="glance-kuttl-tests/glance-default-external-api-1" Dec 01 09:00:54 crc kubenswrapper[4813]: I1201 09:00:54.482430 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/04117ae2-817e-41a3-9d1e-608290bf79c8-lib-modules\") pod \"glance-default-external-api-1\" (UID: \"04117ae2-817e-41a3-9d1e-608290bf79c8\") " pod="glance-kuttl-tests/glance-default-external-api-1" Dec 01 09:00:54 crc kubenswrapper[4813]: I1201 09:00:54.482502 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-external-api-2\" (UID: \"423d0710-1fe1-4cd3-99a8-3c32c7f08ee3\") " pod="glance-kuttl-tests/glance-default-external-api-2" Dec 01 09:00:54 crc kubenswrapper[4813]: I1201 09:00:54.482546 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/423d0710-1fe1-4cd3-99a8-3c32c7f08ee3-httpd-run\") pod \"glance-default-external-api-2\" (UID: \"423d0710-1fe1-4cd3-99a8-3c32c7f08ee3\") " pod="glance-kuttl-tests/glance-default-external-api-2" Dec 01 09:00:54 crc kubenswrapper[4813]: I1201 09:00:54.482617 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-external-api-1\" (UID: \"04117ae2-817e-41a3-9d1e-608290bf79c8\") " pod="glance-kuttl-tests/glance-default-external-api-1" Dec 01 09:00:54 crc kubenswrapper[4813]: I1201 09:00:54.482913 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/04117ae2-817e-41a3-9d1e-608290bf79c8-etc-iscsi\") pod \"glance-default-external-api-1\" (UID: \"04117ae2-817e-41a3-9d1e-608290bf79c8\") " pod="glance-kuttl-tests/glance-default-external-api-1" Dec 01 09:00:54 crc kubenswrapper[4813]: I1201 09:00:54.484725 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/423d0710-1fe1-4cd3-99a8-3c32c7f08ee3-logs\") pod \"glance-default-external-api-2\" (UID: \"423d0710-1fe1-4cd3-99a8-3c32c7f08ee3\") " pod="glance-kuttl-tests/glance-default-external-api-2" Dec 01 09:00:54 crc kubenswrapper[4813]: I1201 09:00:54.484819 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/423d0710-1fe1-4cd3-99a8-3c32c7f08ee3-config-data\") pod \"glance-default-external-api-2\" (UID: \"423d0710-1fe1-4cd3-99a8-3c32c7f08ee3\") " pod="glance-kuttl-tests/glance-default-external-api-2" Dec 01 09:00:54 crc kubenswrapper[4813]: I1201 09:00:54.484859 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/423d0710-1fe1-4cd3-99a8-3c32c7f08ee3-var-locks-brick\") pod \"glance-default-external-api-2\" (UID: \"423d0710-1fe1-4cd3-99a8-3c32c7f08ee3\") " pod="glance-kuttl-tests/glance-default-external-api-2" Dec 01 09:00:54 crc kubenswrapper[4813]: I1201 09:00:54.485074 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/04117ae2-817e-41a3-9d1e-608290bf79c8-sys\") pod \"glance-default-external-api-1\" (UID: \"04117ae2-817e-41a3-9d1e-608290bf79c8\") " pod="glance-kuttl-tests/glance-default-external-api-1" Dec 01 09:00:54 crc kubenswrapper[4813]: I1201 09:00:54.485188 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/423d0710-1fe1-4cd3-99a8-3c32c7f08ee3-etc-nvme\") pod \"glance-default-external-api-2\" (UID: \"423d0710-1fe1-4cd3-99a8-3c32c7f08ee3\") " pod="glance-kuttl-tests/glance-default-external-api-2" Dec 01 09:00:54 crc kubenswrapper[4813]: I1201 09:00:54.485282 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/04117ae2-817e-41a3-9d1e-608290bf79c8-etc-nvme\") pod \"glance-default-external-api-1\" (UID: \"04117ae2-817e-41a3-9d1e-608290bf79c8\") " pod="glance-kuttl-tests/glance-default-external-api-1" Dec 01 09:00:54 crc kubenswrapper[4813]: I1201 09:00:54.485358 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/423d0710-1fe1-4cd3-99a8-3c32c7f08ee3-run\") pod \"glance-default-external-api-2\" (UID: \"423d0710-1fe1-4cd3-99a8-3c32c7f08ee3\") " pod="glance-kuttl-tests/glance-default-external-api-2" Dec 01 09:00:54 crc kubenswrapper[4813]: I1201 09:00:54.485419 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2d8d9\" (UniqueName: \"kubernetes.io/projected/04117ae2-817e-41a3-9d1e-608290bf79c8-kube-api-access-2d8d9\") pod \"glance-default-external-api-1\" (UID: \"04117ae2-817e-41a3-9d1e-608290bf79c8\") " pod="glance-kuttl-tests/glance-default-external-api-1" Dec 01 09:00:54 crc kubenswrapper[4813]: I1201 09:00:54.586849 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/423d0710-1fe1-4cd3-99a8-3c32c7f08ee3-run\") pod \"glance-default-external-api-2\" (UID: \"423d0710-1fe1-4cd3-99a8-3c32c7f08ee3\") " pod="glance-kuttl-tests/glance-default-external-api-2" Dec 01 09:00:54 crc kubenswrapper[4813]: I1201 09:00:54.586909 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2d8d9\" (UniqueName: \"kubernetes.io/projected/04117ae2-817e-41a3-9d1e-608290bf79c8-kube-api-access-2d8d9\") pod \"glance-default-external-api-1\" (UID: \"04117ae2-817e-41a3-9d1e-608290bf79c8\") " pod="glance-kuttl-tests/glance-default-external-api-1" Dec 01 09:00:54 crc kubenswrapper[4813]: I1201 09:00:54.586943 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lxn6d\" (UniqueName: \"kubernetes.io/projected/0aa6a534-c6a3-4a53-9e0e-053abec1ad18-kube-api-access-lxn6d\") pod \"glance-default-internal-api-1\" (UID: \"0aa6a534-c6a3-4a53-9e0e-053abec1ad18\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Dec 01 09:00:54 crc kubenswrapper[4813]: I1201 09:00:54.586986 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/423d0710-1fe1-4cd3-99a8-3c32c7f08ee3-etc-iscsi\") pod \"glance-default-external-api-2\" (UID: \"423d0710-1fe1-4cd3-99a8-3c32c7f08ee3\") " pod="glance-kuttl-tests/glance-default-external-api-2" Dec 01 09:00:54 crc kubenswrapper[4813]: I1201 09:00:54.587004 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a1ad44fa-8926-4242-8d07-1be670bb316a-scripts\") pod \"glance-default-internal-api-2\" (UID: \"a1ad44fa-8926-4242-8d07-1be670bb316a\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Dec 01 09:00:54 crc kubenswrapper[4813]: I1201 09:00:54.587024 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0aa6a534-c6a3-4a53-9e0e-053abec1ad18-logs\") pod \"glance-default-internal-api-1\" (UID: \"0aa6a534-c6a3-4a53-9e0e-053abec1ad18\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Dec 01 09:00:54 crc kubenswrapper[4813]: I1201 09:00:54.587042 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/0aa6a534-c6a3-4a53-9e0e-053abec1ad18-var-locks-brick\") pod \"glance-default-internal-api-1\" (UID: \"0aa6a534-c6a3-4a53-9e0e-053abec1ad18\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Dec 01 09:00:54 crc kubenswrapper[4813]: I1201 09:00:54.587048 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/423d0710-1fe1-4cd3-99a8-3c32c7f08ee3-run\") pod \"glance-default-external-api-2\" (UID: \"423d0710-1fe1-4cd3-99a8-3c32c7f08ee3\") " pod="glance-kuttl-tests/glance-default-external-api-2" Dec 01 09:00:54 crc kubenswrapper[4813]: I1201 09:00:54.587088 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/423d0710-1fe1-4cd3-99a8-3c32c7f08ee3-dev\") pod \"glance-default-external-api-2\" (UID: \"423d0710-1fe1-4cd3-99a8-3c32c7f08ee3\") " pod="glance-kuttl-tests/glance-default-external-api-2" Dec 01 09:00:54 crc kubenswrapper[4813]: I1201 09:00:54.587063 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/423d0710-1fe1-4cd3-99a8-3c32c7f08ee3-dev\") pod \"glance-default-external-api-2\" (UID: \"423d0710-1fe1-4cd3-99a8-3c32c7f08ee3\") " pod="glance-kuttl-tests/glance-default-external-api-2" Dec 01 09:00:54 crc kubenswrapper[4813]: I1201 09:00:54.587189 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-internal-api-2\" (UID: \"a1ad44fa-8926-4242-8d07-1be670bb316a\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Dec 01 09:00:54 crc kubenswrapper[4813]: I1201 09:00:54.587235 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/423d0710-1fe1-4cd3-99a8-3c32c7f08ee3-etc-iscsi\") pod \"glance-default-external-api-2\" (UID: \"423d0710-1fe1-4cd3-99a8-3c32c7f08ee3\") " pod="glance-kuttl-tests/glance-default-external-api-2" Dec 01 09:00:54 crc kubenswrapper[4813]: I1201 09:00:54.587423 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/0aa6a534-c6a3-4a53-9e0e-053abec1ad18-etc-iscsi\") pod \"glance-default-internal-api-1\" (UID: \"0aa6a534-c6a3-4a53-9e0e-053abec1ad18\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Dec 01 09:00:54 crc kubenswrapper[4813]: I1201 09:00:54.587502 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a1ad44fa-8926-4242-8d07-1be670bb316a-config-data\") pod \"glance-default-internal-api-2\" (UID: \"a1ad44fa-8926-4242-8d07-1be670bb316a\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Dec 01 09:00:54 crc kubenswrapper[4813]: I1201 09:00:54.587545 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/423d0710-1fe1-4cd3-99a8-3c32c7f08ee3-sys\") pod \"glance-default-external-api-2\" (UID: \"423d0710-1fe1-4cd3-99a8-3c32c7f08ee3\") " pod="glance-kuttl-tests/glance-default-external-api-2" Dec 01 09:00:54 crc kubenswrapper[4813]: I1201 09:00:54.587568 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/a1ad44fa-8926-4242-8d07-1be670bb316a-run\") pod \"glance-default-internal-api-2\" (UID: \"a1ad44fa-8926-4242-8d07-1be670bb316a\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Dec 01 09:00:54 crc kubenswrapper[4813]: I1201 09:00:54.587599 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/04117ae2-817e-41a3-9d1e-608290bf79c8-config-data\") pod \"glance-default-external-api-1\" (UID: \"04117ae2-817e-41a3-9d1e-608290bf79c8\") " pod="glance-kuttl-tests/glance-default-external-api-1" Dec 01 09:00:54 crc kubenswrapper[4813]: I1201 09:00:54.587623 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/0aa6a534-c6a3-4a53-9e0e-053abec1ad18-run\") pod \"glance-default-internal-api-1\" (UID: \"0aa6a534-c6a3-4a53-9e0e-053abec1ad18\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Dec 01 09:00:54 crc kubenswrapper[4813]: I1201 09:00:54.587632 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/423d0710-1fe1-4cd3-99a8-3c32c7f08ee3-sys\") pod \"glance-default-external-api-2\" (UID: \"423d0710-1fe1-4cd3-99a8-3c32c7f08ee3\") " pod="glance-kuttl-tests/glance-default-external-api-2" Dec 01 09:00:54 crc kubenswrapper[4813]: I1201 09:00:54.587676 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/a1ad44fa-8926-4242-8d07-1be670bb316a-dev\") pod \"glance-default-internal-api-2\" (UID: \"a1ad44fa-8926-4242-8d07-1be670bb316a\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Dec 01 09:00:54 crc kubenswrapper[4813]: I1201 09:00:54.587718 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/04117ae2-817e-41a3-9d1e-608290bf79c8-var-locks-brick\") pod \"glance-default-external-api-1\" (UID: \"04117ae2-817e-41a3-9d1e-608290bf79c8\") " pod="glance-kuttl-tests/glance-default-external-api-1" Dec 01 09:00:54 crc kubenswrapper[4813]: I1201 09:00:54.587738 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/a1ad44fa-8926-4242-8d07-1be670bb316a-lib-modules\") pod \"glance-default-internal-api-2\" (UID: \"a1ad44fa-8926-4242-8d07-1be670bb316a\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Dec 01 09:00:54 crc kubenswrapper[4813]: I1201 09:00:54.587777 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kf7jh\" (UniqueName: \"kubernetes.io/projected/423d0710-1fe1-4cd3-99a8-3c32c7f08ee3-kube-api-access-kf7jh\") pod \"glance-default-external-api-2\" (UID: \"423d0710-1fe1-4cd3-99a8-3c32c7f08ee3\") " pod="glance-kuttl-tests/glance-default-external-api-2" Dec 01 09:00:54 crc kubenswrapper[4813]: I1201 09:00:54.587800 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/423d0710-1fe1-4cd3-99a8-3c32c7f08ee3-scripts\") pod \"glance-default-external-api-2\" (UID: \"423d0710-1fe1-4cd3-99a8-3c32c7f08ee3\") " pod="glance-kuttl-tests/glance-default-external-api-2" Dec 01 09:00:54 crc kubenswrapper[4813]: I1201 09:00:54.587827 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/0aa6a534-c6a3-4a53-9e0e-053abec1ad18-sys\") pod \"glance-default-internal-api-1\" (UID: \"0aa6a534-c6a3-4a53-9e0e-053abec1ad18\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Dec 01 09:00:54 crc kubenswrapper[4813]: I1201 09:00:54.587890 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-external-api-1\" (UID: \"04117ae2-817e-41a3-9d1e-608290bf79c8\") " pod="glance-kuttl-tests/glance-default-external-api-1" Dec 01 09:00:54 crc kubenswrapper[4813]: I1201 09:00:54.587917 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/423d0710-1fe1-4cd3-99a8-3c32c7f08ee3-lib-modules\") pod \"glance-default-external-api-2\" (UID: \"423d0710-1fe1-4cd3-99a8-3c32c7f08ee3\") " pod="glance-kuttl-tests/glance-default-external-api-2" Dec 01 09:00:54 crc kubenswrapper[4813]: I1201 09:00:54.587933 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/a1ad44fa-8926-4242-8d07-1be670bb316a-etc-nvme\") pod \"glance-default-internal-api-2\" (UID: \"a1ad44fa-8926-4242-8d07-1be670bb316a\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Dec 01 09:00:54 crc kubenswrapper[4813]: I1201 09:00:54.587995 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/04117ae2-817e-41a3-9d1e-608290bf79c8-httpd-run\") pod \"glance-default-external-api-1\" (UID: \"04117ae2-817e-41a3-9d1e-608290bf79c8\") " pod="glance-kuttl-tests/glance-default-external-api-1" Dec 01 09:00:54 crc kubenswrapper[4813]: I1201 09:00:54.588020 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage14-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage14-crc\") pod \"glance-default-external-api-2\" (UID: \"423d0710-1fe1-4cd3-99a8-3c32c7f08ee3\") " pod="glance-kuttl-tests/glance-default-external-api-2" Dec 01 09:00:54 crc kubenswrapper[4813]: I1201 09:00:54.588111 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/04117ae2-817e-41a3-9d1e-608290bf79c8-var-locks-brick\") pod \"glance-default-external-api-1\" (UID: \"04117ae2-817e-41a3-9d1e-608290bf79c8\") " pod="glance-kuttl-tests/glance-default-external-api-1" Dec 01 09:00:54 crc kubenswrapper[4813]: I1201 09:00:54.588234 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/423d0710-1fe1-4cd3-99a8-3c32c7f08ee3-lib-modules\") pod \"glance-default-external-api-2\" (UID: \"423d0710-1fe1-4cd3-99a8-3c32c7f08ee3\") " pod="glance-kuttl-tests/glance-default-external-api-2" Dec 01 09:00:54 crc kubenswrapper[4813]: I1201 09:00:54.588290 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/04117ae2-817e-41a3-9d1e-608290bf79c8-run\") pod \"glance-default-external-api-1\" (UID: \"04117ae2-817e-41a3-9d1e-608290bf79c8\") " pod="glance-kuttl-tests/glance-default-external-api-1" Dec 01 09:00:54 crc kubenswrapper[4813]: I1201 09:00:54.588315 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/0aa6a534-c6a3-4a53-9e0e-053abec1ad18-dev\") pod \"glance-default-internal-api-1\" (UID: \"0aa6a534-c6a3-4a53-9e0e-053abec1ad18\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Dec 01 09:00:54 crc kubenswrapper[4813]: I1201 09:00:54.588338 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/04117ae2-817e-41a3-9d1e-608290bf79c8-logs\") pod \"glance-default-external-api-1\" (UID: \"04117ae2-817e-41a3-9d1e-608290bf79c8\") " pod="glance-kuttl-tests/glance-default-external-api-1" Dec 01 09:00:54 crc kubenswrapper[4813]: I1201 09:00:54.588334 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/04117ae2-817e-41a3-9d1e-608290bf79c8-run\") pod \"glance-default-external-api-1\" (UID: \"04117ae2-817e-41a3-9d1e-608290bf79c8\") " pod="glance-kuttl-tests/glance-default-external-api-1" Dec 01 09:00:54 crc kubenswrapper[4813]: I1201 09:00:54.588360 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/04117ae2-817e-41a3-9d1e-608290bf79c8-scripts\") pod \"glance-default-external-api-1\" (UID: \"04117ae2-817e-41a3-9d1e-608290bf79c8\") " pod="glance-kuttl-tests/glance-default-external-api-1" Dec 01 09:00:54 crc kubenswrapper[4813]: I1201 09:00:54.588514 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/a1ad44fa-8926-4242-8d07-1be670bb316a-etc-iscsi\") pod \"glance-default-internal-api-2\" (UID: \"a1ad44fa-8926-4242-8d07-1be670bb316a\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Dec 01 09:00:54 crc kubenswrapper[4813]: I1201 09:00:54.588530 4813 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage14-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage14-crc\") pod \"glance-default-external-api-2\" (UID: \"423d0710-1fe1-4cd3-99a8-3c32c7f08ee3\") device mount path \"/mnt/openstack/pv14\"" pod="glance-kuttl-tests/glance-default-external-api-2" Dec 01 09:00:54 crc kubenswrapper[4813]: I1201 09:00:54.588577 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/04117ae2-817e-41a3-9d1e-608290bf79c8-dev\") pod \"glance-default-external-api-1\" (UID: \"04117ae2-817e-41a3-9d1e-608290bf79c8\") " pod="glance-kuttl-tests/glance-default-external-api-1" Dec 01 09:00:54 crc kubenswrapper[4813]: I1201 09:00:54.588568 4813 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-external-api-1\" (UID: \"04117ae2-817e-41a3-9d1e-608290bf79c8\") device mount path \"/mnt/openstack/pv04\"" pod="glance-kuttl-tests/glance-default-external-api-1" Dec 01 09:00:54 crc kubenswrapper[4813]: I1201 09:00:54.588608 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/04117ae2-817e-41a3-9d1e-608290bf79c8-lib-modules\") pod \"glance-default-external-api-1\" (UID: \"04117ae2-817e-41a3-9d1e-608290bf79c8\") " pod="glance-kuttl-tests/glance-default-external-api-1" Dec 01 09:00:54 crc kubenswrapper[4813]: I1201 09:00:54.588634 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/0aa6a534-c6a3-4a53-9e0e-053abec1ad18-etc-nvme\") pod \"glance-default-internal-api-1\" (UID: \"0aa6a534-c6a3-4a53-9e0e-053abec1ad18\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Dec 01 09:00:54 crc kubenswrapper[4813]: I1201 09:00:54.588658 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-internal-api-1\" (UID: \"0aa6a534-c6a3-4a53-9e0e-053abec1ad18\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Dec 01 09:00:54 crc kubenswrapper[4813]: I1201 09:00:54.588692 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/04117ae2-817e-41a3-9d1e-608290bf79c8-dev\") pod \"glance-default-external-api-1\" (UID: \"04117ae2-817e-41a3-9d1e-608290bf79c8\") " pod="glance-kuttl-tests/glance-default-external-api-1" Dec 01 09:00:54 crc kubenswrapper[4813]: I1201 09:00:54.588741 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/04117ae2-817e-41a3-9d1e-608290bf79c8-lib-modules\") pod \"glance-default-external-api-1\" (UID: \"04117ae2-817e-41a3-9d1e-608290bf79c8\") " pod="glance-kuttl-tests/glance-default-external-api-1" Dec 01 09:00:54 crc kubenswrapper[4813]: I1201 09:00:54.588787 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-external-api-2\" (UID: \"423d0710-1fe1-4cd3-99a8-3c32c7f08ee3\") " pod="glance-kuttl-tests/glance-default-external-api-2" Dec 01 09:00:54 crc kubenswrapper[4813]: I1201 09:00:54.588818 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/423d0710-1fe1-4cd3-99a8-3c32c7f08ee3-httpd-run\") pod \"glance-default-external-api-2\" (UID: \"423d0710-1fe1-4cd3-99a8-3c32c7f08ee3\") " pod="glance-kuttl-tests/glance-default-external-api-2" Dec 01 09:00:54 crc kubenswrapper[4813]: I1201 09:00:54.588850 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0aa6a534-c6a3-4a53-9e0e-053abec1ad18-scripts\") pod \"glance-default-internal-api-1\" (UID: \"0aa6a534-c6a3-4a53-9e0e-053abec1ad18\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Dec 01 09:00:54 crc kubenswrapper[4813]: I1201 09:00:54.588875 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-external-api-1\" (UID: \"04117ae2-817e-41a3-9d1e-608290bf79c8\") " pod="glance-kuttl-tests/glance-default-external-api-1" Dec 01 09:00:54 crc kubenswrapper[4813]: I1201 09:00:54.588926 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/04117ae2-817e-41a3-9d1e-608290bf79c8-etc-iscsi\") pod \"glance-default-external-api-1\" (UID: \"04117ae2-817e-41a3-9d1e-608290bf79c8\") " pod="glance-kuttl-tests/glance-default-external-api-1" Dec 01 09:00:54 crc kubenswrapper[4813]: I1201 09:00:54.588985 4813 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-external-api-2\" (UID: \"423d0710-1fe1-4cd3-99a8-3c32c7f08ee3\") device mount path \"/mnt/openstack/pv07\"" pod="glance-kuttl-tests/glance-default-external-api-2" Dec 01 09:00:54 crc kubenswrapper[4813]: I1201 09:00:54.589036 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/423d0710-1fe1-4cd3-99a8-3c32c7f08ee3-logs\") pod \"glance-default-external-api-2\" (UID: \"423d0710-1fe1-4cd3-99a8-3c32c7f08ee3\") " pod="glance-kuttl-tests/glance-default-external-api-2" Dec 01 09:00:54 crc kubenswrapper[4813]: I1201 09:00:54.589068 4813 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-external-api-1\" (UID: \"04117ae2-817e-41a3-9d1e-608290bf79c8\") device mount path \"/mnt/openstack/pv03\"" pod="glance-kuttl-tests/glance-default-external-api-1" Dec 01 09:00:54 crc kubenswrapper[4813]: I1201 09:00:54.589086 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/423d0710-1fe1-4cd3-99a8-3c32c7f08ee3-config-data\") pod \"glance-default-external-api-2\" (UID: \"423d0710-1fe1-4cd3-99a8-3c32c7f08ee3\") " pod="glance-kuttl-tests/glance-default-external-api-2" Dec 01 09:00:54 crc kubenswrapper[4813]: I1201 09:00:54.589042 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/04117ae2-817e-41a3-9d1e-608290bf79c8-etc-iscsi\") pod \"glance-default-external-api-1\" (UID: \"04117ae2-817e-41a3-9d1e-608290bf79c8\") " pod="glance-kuttl-tests/glance-default-external-api-1" Dec 01 09:00:54 crc kubenswrapper[4813]: I1201 09:00:54.589143 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/a1ad44fa-8926-4242-8d07-1be670bb316a-var-locks-brick\") pod \"glance-default-internal-api-2\" (UID: \"a1ad44fa-8926-4242-8d07-1be670bb316a\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Dec 01 09:00:54 crc kubenswrapper[4813]: I1201 09:00:54.589186 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/0aa6a534-c6a3-4a53-9e0e-053abec1ad18-httpd-run\") pod \"glance-default-internal-api-1\" (UID: \"0aa6a534-c6a3-4a53-9e0e-053abec1ad18\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Dec 01 09:00:54 crc kubenswrapper[4813]: I1201 09:00:54.589232 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/423d0710-1fe1-4cd3-99a8-3c32c7f08ee3-var-locks-brick\") pod \"glance-default-external-api-2\" (UID: \"423d0710-1fe1-4cd3-99a8-3c32c7f08ee3\") " pod="glance-kuttl-tests/glance-default-external-api-2" Dec 01 09:00:54 crc kubenswrapper[4813]: I1201 09:00:54.589283 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ng4mg\" (UniqueName: \"kubernetes.io/projected/a1ad44fa-8926-4242-8d07-1be670bb316a-kube-api-access-ng4mg\") pod \"glance-default-internal-api-2\" (UID: \"a1ad44fa-8926-4242-8d07-1be670bb316a\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Dec 01 09:00:54 crc kubenswrapper[4813]: I1201 09:00:54.589322 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/a1ad44fa-8926-4242-8d07-1be670bb316a-httpd-run\") pod \"glance-default-internal-api-2\" (UID: \"a1ad44fa-8926-4242-8d07-1be670bb316a\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Dec 01 09:00:54 crc kubenswrapper[4813]: I1201 09:00:54.589401 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/04117ae2-817e-41a3-9d1e-608290bf79c8-sys\") pod \"glance-default-external-api-1\" (UID: \"04117ae2-817e-41a3-9d1e-608290bf79c8\") " pod="glance-kuttl-tests/glance-default-external-api-1" Dec 01 09:00:54 crc kubenswrapper[4813]: I1201 09:00:54.589421 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/423d0710-1fe1-4cd3-99a8-3c32c7f08ee3-httpd-run\") pod \"glance-default-external-api-2\" (UID: \"423d0710-1fe1-4cd3-99a8-3c32c7f08ee3\") " pod="glance-kuttl-tests/glance-default-external-api-2" Dec 01 09:00:54 crc kubenswrapper[4813]: I1201 09:00:54.589442 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a1ad44fa-8926-4242-8d07-1be670bb316a-logs\") pod \"glance-default-internal-api-2\" (UID: \"a1ad44fa-8926-4242-8d07-1be670bb316a\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Dec 01 09:00:54 crc kubenswrapper[4813]: I1201 09:00:54.589521 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/423d0710-1fe1-4cd3-99a8-3c32c7f08ee3-var-locks-brick\") pod \"glance-default-external-api-2\" (UID: \"423d0710-1fe1-4cd3-99a8-3c32c7f08ee3\") " pod="glance-kuttl-tests/glance-default-external-api-2" Dec 01 09:00:54 crc kubenswrapper[4813]: I1201 09:00:54.589544 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage19-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage19-crc\") pod \"glance-default-internal-api-2\" (UID: \"a1ad44fa-8926-4242-8d07-1be670bb316a\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Dec 01 09:00:54 crc kubenswrapper[4813]: I1201 09:00:54.589593 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage18-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage18-crc\") pod \"glance-default-internal-api-1\" (UID: \"0aa6a534-c6a3-4a53-9e0e-053abec1ad18\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Dec 01 09:00:54 crc kubenswrapper[4813]: I1201 09:00:54.589618 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/04117ae2-817e-41a3-9d1e-608290bf79c8-sys\") pod \"glance-default-external-api-1\" (UID: \"04117ae2-817e-41a3-9d1e-608290bf79c8\") " pod="glance-kuttl-tests/glance-default-external-api-1" Dec 01 09:00:54 crc kubenswrapper[4813]: I1201 09:00:54.589749 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/04117ae2-817e-41a3-9d1e-608290bf79c8-httpd-run\") pod \"glance-default-external-api-1\" (UID: \"04117ae2-817e-41a3-9d1e-608290bf79c8\") " pod="glance-kuttl-tests/glance-default-external-api-1" Dec 01 09:00:54 crc kubenswrapper[4813]: I1201 09:00:54.589788 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/423d0710-1fe1-4cd3-99a8-3c32c7f08ee3-etc-nvme\") pod \"glance-default-external-api-2\" (UID: \"423d0710-1fe1-4cd3-99a8-3c32c7f08ee3\") " pod="glance-kuttl-tests/glance-default-external-api-2" Dec 01 09:00:54 crc kubenswrapper[4813]: I1201 09:00:54.589836 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0aa6a534-c6a3-4a53-9e0e-053abec1ad18-config-data\") pod \"glance-default-internal-api-1\" (UID: \"0aa6a534-c6a3-4a53-9e0e-053abec1ad18\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Dec 01 09:00:54 crc kubenswrapper[4813]: I1201 09:00:54.589887 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/a1ad44fa-8926-4242-8d07-1be670bb316a-sys\") pod \"glance-default-internal-api-2\" (UID: \"a1ad44fa-8926-4242-8d07-1be670bb316a\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Dec 01 09:00:54 crc kubenswrapper[4813]: I1201 09:00:54.589907 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/423d0710-1fe1-4cd3-99a8-3c32c7f08ee3-etc-nvme\") pod \"glance-default-external-api-2\" (UID: \"423d0710-1fe1-4cd3-99a8-3c32c7f08ee3\") " pod="glance-kuttl-tests/glance-default-external-api-2" Dec 01 09:00:54 crc kubenswrapper[4813]: I1201 09:00:54.589943 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/04117ae2-817e-41a3-9d1e-608290bf79c8-etc-nvme\") pod \"glance-default-external-api-1\" (UID: \"04117ae2-817e-41a3-9d1e-608290bf79c8\") " pod="glance-kuttl-tests/glance-default-external-api-1" Dec 01 09:00:54 crc kubenswrapper[4813]: I1201 09:00:54.590009 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/0aa6a534-c6a3-4a53-9e0e-053abec1ad18-lib-modules\") pod \"glance-default-internal-api-1\" (UID: \"0aa6a534-c6a3-4a53-9e0e-053abec1ad18\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Dec 01 09:00:54 crc kubenswrapper[4813]: I1201 09:00:54.590042 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/04117ae2-817e-41a3-9d1e-608290bf79c8-etc-nvme\") pod \"glance-default-external-api-1\" (UID: \"04117ae2-817e-41a3-9d1e-608290bf79c8\") " pod="glance-kuttl-tests/glance-default-external-api-1" Dec 01 09:00:54 crc kubenswrapper[4813]: I1201 09:00:54.590198 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/423d0710-1fe1-4cd3-99a8-3c32c7f08ee3-logs\") pod \"glance-default-external-api-2\" (UID: \"423d0710-1fe1-4cd3-99a8-3c32c7f08ee3\") " pod="glance-kuttl-tests/glance-default-external-api-2" Dec 01 09:00:54 crc kubenswrapper[4813]: I1201 09:00:54.591462 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/04117ae2-817e-41a3-9d1e-608290bf79c8-logs\") pod \"glance-default-external-api-1\" (UID: \"04117ae2-817e-41a3-9d1e-608290bf79c8\") " pod="glance-kuttl-tests/glance-default-external-api-1" Dec 01 09:00:54 crc kubenswrapper[4813]: I1201 09:00:54.610672 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/04117ae2-817e-41a3-9d1e-608290bf79c8-scripts\") pod \"glance-default-external-api-1\" (UID: \"04117ae2-817e-41a3-9d1e-608290bf79c8\") " pod="glance-kuttl-tests/glance-default-external-api-1" Dec 01 09:00:54 crc kubenswrapper[4813]: I1201 09:00:54.611780 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/423d0710-1fe1-4cd3-99a8-3c32c7f08ee3-config-data\") pod \"glance-default-external-api-2\" (UID: \"423d0710-1fe1-4cd3-99a8-3c32c7f08ee3\") " pod="glance-kuttl-tests/glance-default-external-api-2" Dec 01 09:00:54 crc kubenswrapper[4813]: I1201 09:00:54.611843 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/423d0710-1fe1-4cd3-99a8-3c32c7f08ee3-scripts\") pod \"glance-default-external-api-2\" (UID: \"423d0710-1fe1-4cd3-99a8-3c32c7f08ee3\") " pod="glance-kuttl-tests/glance-default-external-api-2" Dec 01 09:00:54 crc kubenswrapper[4813]: I1201 09:00:54.620270 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/04117ae2-817e-41a3-9d1e-608290bf79c8-config-data\") pod \"glance-default-external-api-1\" (UID: \"04117ae2-817e-41a3-9d1e-608290bf79c8\") " pod="glance-kuttl-tests/glance-default-external-api-1" Dec 01 09:00:54 crc kubenswrapper[4813]: I1201 09:00:54.626615 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2d8d9\" (UniqueName: \"kubernetes.io/projected/04117ae2-817e-41a3-9d1e-608290bf79c8-kube-api-access-2d8d9\") pod \"glance-default-external-api-1\" (UID: \"04117ae2-817e-41a3-9d1e-608290bf79c8\") " pod="glance-kuttl-tests/glance-default-external-api-1" Dec 01 09:00:54 crc kubenswrapper[4813]: I1201 09:00:54.638486 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kf7jh\" (UniqueName: \"kubernetes.io/projected/423d0710-1fe1-4cd3-99a8-3c32c7f08ee3-kube-api-access-kf7jh\") pod \"glance-default-external-api-2\" (UID: \"423d0710-1fe1-4cd3-99a8-3c32c7f08ee3\") " pod="glance-kuttl-tests/glance-default-external-api-2" Dec 01 09:00:54 crc kubenswrapper[4813]: I1201 09:00:54.639258 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage14-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage14-crc\") pod \"glance-default-external-api-2\" (UID: \"423d0710-1fe1-4cd3-99a8-3c32c7f08ee3\") " pod="glance-kuttl-tests/glance-default-external-api-2" Dec 01 09:00:54 crc kubenswrapper[4813]: I1201 09:00:54.640062 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-external-api-1\" (UID: \"04117ae2-817e-41a3-9d1e-608290bf79c8\") " pod="glance-kuttl-tests/glance-default-external-api-1" Dec 01 09:00:54 crc kubenswrapper[4813]: I1201 09:00:54.654563 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-external-api-1\" (UID: \"04117ae2-817e-41a3-9d1e-608290bf79c8\") " pod="glance-kuttl-tests/glance-default-external-api-1" Dec 01 09:00:54 crc kubenswrapper[4813]: I1201 09:00:54.659120 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-external-api-2\" (UID: \"423d0710-1fe1-4cd3-99a8-3c32c7f08ee3\") " pod="glance-kuttl-tests/glance-default-external-api-2" Dec 01 09:00:54 crc kubenswrapper[4813]: I1201 09:00:54.666836 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-2" Dec 01 09:00:54 crc kubenswrapper[4813]: I1201 09:00:54.686509 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-1" Dec 01 09:00:54 crc kubenswrapper[4813]: I1201 09:00:54.691839 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/0aa6a534-c6a3-4a53-9e0e-053abec1ad18-etc-nvme\") pod \"glance-default-internal-api-1\" (UID: \"0aa6a534-c6a3-4a53-9e0e-053abec1ad18\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Dec 01 09:00:54 crc kubenswrapper[4813]: I1201 09:00:54.691889 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-internal-api-1\" (UID: \"0aa6a534-c6a3-4a53-9e0e-053abec1ad18\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Dec 01 09:00:54 crc kubenswrapper[4813]: I1201 09:00:54.691909 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0aa6a534-c6a3-4a53-9e0e-053abec1ad18-scripts\") pod \"glance-default-internal-api-1\" (UID: \"0aa6a534-c6a3-4a53-9e0e-053abec1ad18\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Dec 01 09:00:54 crc kubenswrapper[4813]: I1201 09:00:54.691933 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/0aa6a534-c6a3-4a53-9e0e-053abec1ad18-httpd-run\") pod \"glance-default-internal-api-1\" (UID: \"0aa6a534-c6a3-4a53-9e0e-053abec1ad18\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Dec 01 09:00:54 crc kubenswrapper[4813]: I1201 09:00:54.692048 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/a1ad44fa-8926-4242-8d07-1be670bb316a-var-locks-brick\") pod \"glance-default-internal-api-2\" (UID: \"a1ad44fa-8926-4242-8d07-1be670bb316a\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Dec 01 09:00:54 crc kubenswrapper[4813]: I1201 09:00:54.692070 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ng4mg\" (UniqueName: \"kubernetes.io/projected/a1ad44fa-8926-4242-8d07-1be670bb316a-kube-api-access-ng4mg\") pod \"glance-default-internal-api-2\" (UID: \"a1ad44fa-8926-4242-8d07-1be670bb316a\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Dec 01 09:00:54 crc kubenswrapper[4813]: I1201 09:00:54.692076 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/0aa6a534-c6a3-4a53-9e0e-053abec1ad18-etc-nvme\") pod \"glance-default-internal-api-1\" (UID: \"0aa6a534-c6a3-4a53-9e0e-053abec1ad18\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Dec 01 09:00:54 crc kubenswrapper[4813]: I1201 09:00:54.692084 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/a1ad44fa-8926-4242-8d07-1be670bb316a-httpd-run\") pod \"glance-default-internal-api-2\" (UID: \"a1ad44fa-8926-4242-8d07-1be670bb316a\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Dec 01 09:00:54 crc kubenswrapper[4813]: I1201 09:00:54.692146 4813 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-internal-api-1\" (UID: \"0aa6a534-c6a3-4a53-9e0e-053abec1ad18\") device mount path \"/mnt/openstack/pv11\"" pod="glance-kuttl-tests/glance-default-internal-api-1" Dec 01 09:00:54 crc kubenswrapper[4813]: I1201 09:00:54.692163 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/a1ad44fa-8926-4242-8d07-1be670bb316a-var-locks-brick\") pod \"glance-default-internal-api-2\" (UID: \"a1ad44fa-8926-4242-8d07-1be670bb316a\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Dec 01 09:00:54 crc kubenswrapper[4813]: I1201 09:00:54.692299 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a1ad44fa-8926-4242-8d07-1be670bb316a-logs\") pod \"glance-default-internal-api-2\" (UID: \"a1ad44fa-8926-4242-8d07-1be670bb316a\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Dec 01 09:00:54 crc kubenswrapper[4813]: I1201 09:00:54.692339 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage19-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage19-crc\") pod \"glance-default-internal-api-2\" (UID: \"a1ad44fa-8926-4242-8d07-1be670bb316a\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Dec 01 09:00:54 crc kubenswrapper[4813]: I1201 09:00:54.692362 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage18-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage18-crc\") pod \"glance-default-internal-api-1\" (UID: \"0aa6a534-c6a3-4a53-9e0e-053abec1ad18\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Dec 01 09:00:54 crc kubenswrapper[4813]: I1201 09:00:54.692392 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0aa6a534-c6a3-4a53-9e0e-053abec1ad18-config-data\") pod \"glance-default-internal-api-1\" (UID: \"0aa6a534-c6a3-4a53-9e0e-053abec1ad18\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Dec 01 09:00:54 crc kubenswrapper[4813]: I1201 09:00:54.692470 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/a1ad44fa-8926-4242-8d07-1be670bb316a-httpd-run\") pod \"glance-default-internal-api-2\" (UID: \"a1ad44fa-8926-4242-8d07-1be670bb316a\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Dec 01 09:00:54 crc kubenswrapper[4813]: I1201 09:00:54.692514 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/0aa6a534-c6a3-4a53-9e0e-053abec1ad18-httpd-run\") pod \"glance-default-internal-api-1\" (UID: \"0aa6a534-c6a3-4a53-9e0e-053abec1ad18\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Dec 01 09:00:54 crc kubenswrapper[4813]: I1201 09:00:54.692583 4813 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage18-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage18-crc\") pod \"glance-default-internal-api-1\" (UID: \"0aa6a534-c6a3-4a53-9e0e-053abec1ad18\") device mount path \"/mnt/openstack/pv18\"" pod="glance-kuttl-tests/glance-default-internal-api-1" Dec 01 09:00:54 crc kubenswrapper[4813]: I1201 09:00:54.692602 4813 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage19-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage19-crc\") pod \"glance-default-internal-api-2\" (UID: \"a1ad44fa-8926-4242-8d07-1be670bb316a\") device mount path \"/mnt/openstack/pv19\"" pod="glance-kuttl-tests/glance-default-internal-api-2" Dec 01 09:00:54 crc kubenswrapper[4813]: I1201 09:00:54.692673 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/a1ad44fa-8926-4242-8d07-1be670bb316a-sys\") pod \"glance-default-internal-api-2\" (UID: \"a1ad44fa-8926-4242-8d07-1be670bb316a\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Dec 01 09:00:54 crc kubenswrapper[4813]: I1201 09:00:54.692809 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a1ad44fa-8926-4242-8d07-1be670bb316a-logs\") pod \"glance-default-internal-api-2\" (UID: \"a1ad44fa-8926-4242-8d07-1be670bb316a\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Dec 01 09:00:54 crc kubenswrapper[4813]: I1201 09:00:54.692883 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/a1ad44fa-8926-4242-8d07-1be670bb316a-sys\") pod \"glance-default-internal-api-2\" (UID: \"a1ad44fa-8926-4242-8d07-1be670bb316a\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Dec 01 09:00:54 crc kubenswrapper[4813]: I1201 09:00:54.692928 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/0aa6a534-c6a3-4a53-9e0e-053abec1ad18-lib-modules\") pod \"glance-default-internal-api-1\" (UID: \"0aa6a534-c6a3-4a53-9e0e-053abec1ad18\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Dec 01 09:00:54 crc kubenswrapper[4813]: I1201 09:00:54.692978 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lxn6d\" (UniqueName: \"kubernetes.io/projected/0aa6a534-c6a3-4a53-9e0e-053abec1ad18-kube-api-access-lxn6d\") pod \"glance-default-internal-api-1\" (UID: \"0aa6a534-c6a3-4a53-9e0e-053abec1ad18\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Dec 01 09:00:54 crc kubenswrapper[4813]: I1201 09:00:54.693010 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a1ad44fa-8926-4242-8d07-1be670bb316a-scripts\") pod \"glance-default-internal-api-2\" (UID: \"a1ad44fa-8926-4242-8d07-1be670bb316a\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Dec 01 09:00:54 crc kubenswrapper[4813]: I1201 09:00:54.693014 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/0aa6a534-c6a3-4a53-9e0e-053abec1ad18-lib-modules\") pod \"glance-default-internal-api-1\" (UID: \"0aa6a534-c6a3-4a53-9e0e-053abec1ad18\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Dec 01 09:00:54 crc kubenswrapper[4813]: I1201 09:00:54.693031 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/0aa6a534-c6a3-4a53-9e0e-053abec1ad18-var-locks-brick\") pod \"glance-default-internal-api-1\" (UID: \"0aa6a534-c6a3-4a53-9e0e-053abec1ad18\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Dec 01 09:00:54 crc kubenswrapper[4813]: I1201 09:00:54.693054 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0aa6a534-c6a3-4a53-9e0e-053abec1ad18-logs\") pod \"glance-default-internal-api-1\" (UID: \"0aa6a534-c6a3-4a53-9e0e-053abec1ad18\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Dec 01 09:00:54 crc kubenswrapper[4813]: I1201 09:00:54.693095 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-internal-api-2\" (UID: \"a1ad44fa-8926-4242-8d07-1be670bb316a\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Dec 01 09:00:54 crc kubenswrapper[4813]: I1201 09:00:54.693120 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a1ad44fa-8926-4242-8d07-1be670bb316a-config-data\") pod \"glance-default-internal-api-2\" (UID: \"a1ad44fa-8926-4242-8d07-1be670bb316a\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Dec 01 09:00:54 crc kubenswrapper[4813]: I1201 09:00:54.693155 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/0aa6a534-c6a3-4a53-9e0e-053abec1ad18-etc-iscsi\") pod \"glance-default-internal-api-1\" (UID: \"0aa6a534-c6a3-4a53-9e0e-053abec1ad18\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Dec 01 09:00:54 crc kubenswrapper[4813]: I1201 09:00:54.693182 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/a1ad44fa-8926-4242-8d07-1be670bb316a-run\") pod \"glance-default-internal-api-2\" (UID: \"a1ad44fa-8926-4242-8d07-1be670bb316a\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Dec 01 09:00:54 crc kubenswrapper[4813]: I1201 09:00:54.693218 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/0aa6a534-c6a3-4a53-9e0e-053abec1ad18-run\") pod \"glance-default-internal-api-1\" (UID: \"0aa6a534-c6a3-4a53-9e0e-053abec1ad18\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Dec 01 09:00:54 crc kubenswrapper[4813]: I1201 09:00:54.693245 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/a1ad44fa-8926-4242-8d07-1be670bb316a-dev\") pod \"glance-default-internal-api-2\" (UID: \"a1ad44fa-8926-4242-8d07-1be670bb316a\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Dec 01 09:00:54 crc kubenswrapper[4813]: I1201 09:00:54.693274 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/a1ad44fa-8926-4242-8d07-1be670bb316a-lib-modules\") pod \"glance-default-internal-api-2\" (UID: \"a1ad44fa-8926-4242-8d07-1be670bb316a\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Dec 01 09:00:54 crc kubenswrapper[4813]: I1201 09:00:54.693307 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/0aa6a534-c6a3-4a53-9e0e-053abec1ad18-sys\") pod \"glance-default-internal-api-1\" (UID: \"0aa6a534-c6a3-4a53-9e0e-053abec1ad18\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Dec 01 09:00:54 crc kubenswrapper[4813]: I1201 09:00:54.693343 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/a1ad44fa-8926-4242-8d07-1be670bb316a-etc-nvme\") pod \"glance-default-internal-api-2\" (UID: \"a1ad44fa-8926-4242-8d07-1be670bb316a\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Dec 01 09:00:54 crc kubenswrapper[4813]: I1201 09:00:54.693407 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/0aa6a534-c6a3-4a53-9e0e-053abec1ad18-dev\") pod \"glance-default-internal-api-1\" (UID: \"0aa6a534-c6a3-4a53-9e0e-053abec1ad18\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Dec 01 09:00:54 crc kubenswrapper[4813]: I1201 09:00:54.693436 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/a1ad44fa-8926-4242-8d07-1be670bb316a-etc-iscsi\") pod \"glance-default-internal-api-2\" (UID: \"a1ad44fa-8926-4242-8d07-1be670bb316a\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Dec 01 09:00:54 crc kubenswrapper[4813]: I1201 09:00:54.693507 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0aa6a534-c6a3-4a53-9e0e-053abec1ad18-logs\") pod \"glance-default-internal-api-1\" (UID: \"0aa6a534-c6a3-4a53-9e0e-053abec1ad18\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Dec 01 09:00:54 crc kubenswrapper[4813]: I1201 09:00:54.693518 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/a1ad44fa-8926-4242-8d07-1be670bb316a-etc-iscsi\") pod \"glance-default-internal-api-2\" (UID: \"a1ad44fa-8926-4242-8d07-1be670bb316a\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Dec 01 09:00:54 crc kubenswrapper[4813]: I1201 09:00:54.693615 4813 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-internal-api-2\" (UID: \"a1ad44fa-8926-4242-8d07-1be670bb316a\") device mount path \"/mnt/openstack/pv08\"" pod="glance-kuttl-tests/glance-default-internal-api-2" Dec 01 09:00:54 crc kubenswrapper[4813]: I1201 09:00:54.694036 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/a1ad44fa-8926-4242-8d07-1be670bb316a-dev\") pod \"glance-default-internal-api-2\" (UID: \"a1ad44fa-8926-4242-8d07-1be670bb316a\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Dec 01 09:00:54 crc kubenswrapper[4813]: I1201 09:00:54.697089 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/0aa6a534-c6a3-4a53-9e0e-053abec1ad18-sys\") pod \"glance-default-internal-api-1\" (UID: \"0aa6a534-c6a3-4a53-9e0e-053abec1ad18\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Dec 01 09:00:54 crc kubenswrapper[4813]: I1201 09:00:54.697131 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/a1ad44fa-8926-4242-8d07-1be670bb316a-run\") pod \"glance-default-internal-api-2\" (UID: \"a1ad44fa-8926-4242-8d07-1be670bb316a\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Dec 01 09:00:54 crc kubenswrapper[4813]: I1201 09:00:54.697106 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/a1ad44fa-8926-4242-8d07-1be670bb316a-lib-modules\") pod \"glance-default-internal-api-2\" (UID: \"a1ad44fa-8926-4242-8d07-1be670bb316a\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Dec 01 09:00:54 crc kubenswrapper[4813]: I1201 09:00:54.697170 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/0aa6a534-c6a3-4a53-9e0e-053abec1ad18-etc-iscsi\") pod \"glance-default-internal-api-1\" (UID: \"0aa6a534-c6a3-4a53-9e0e-053abec1ad18\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Dec 01 09:00:54 crc kubenswrapper[4813]: I1201 09:00:54.697207 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/a1ad44fa-8926-4242-8d07-1be670bb316a-etc-nvme\") pod \"glance-default-internal-api-2\" (UID: \"a1ad44fa-8926-4242-8d07-1be670bb316a\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Dec 01 09:00:54 crc kubenswrapper[4813]: I1201 09:00:54.697219 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/0aa6a534-c6a3-4a53-9e0e-053abec1ad18-var-locks-brick\") pod \"glance-default-internal-api-1\" (UID: \"0aa6a534-c6a3-4a53-9e0e-053abec1ad18\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Dec 01 09:00:54 crc kubenswrapper[4813]: I1201 09:00:54.697259 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/0aa6a534-c6a3-4a53-9e0e-053abec1ad18-run\") pod \"glance-default-internal-api-1\" (UID: \"0aa6a534-c6a3-4a53-9e0e-053abec1ad18\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Dec 01 09:00:54 crc kubenswrapper[4813]: I1201 09:00:54.697265 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/0aa6a534-c6a3-4a53-9e0e-053abec1ad18-dev\") pod \"glance-default-internal-api-1\" (UID: \"0aa6a534-c6a3-4a53-9e0e-053abec1ad18\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Dec 01 09:00:54 crc kubenswrapper[4813]: I1201 09:00:54.698235 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a1ad44fa-8926-4242-8d07-1be670bb316a-scripts\") pod \"glance-default-internal-api-2\" (UID: \"a1ad44fa-8926-4242-8d07-1be670bb316a\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Dec 01 09:00:54 crc kubenswrapper[4813]: I1201 09:00:54.700565 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a1ad44fa-8926-4242-8d07-1be670bb316a-config-data\") pod \"glance-default-internal-api-2\" (UID: \"a1ad44fa-8926-4242-8d07-1be670bb316a\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Dec 01 09:00:54 crc kubenswrapper[4813]: I1201 09:00:54.714300 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0aa6a534-c6a3-4a53-9e0e-053abec1ad18-scripts\") pod \"glance-default-internal-api-1\" (UID: \"0aa6a534-c6a3-4a53-9e0e-053abec1ad18\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Dec 01 09:00:54 crc kubenswrapper[4813]: I1201 09:00:54.715632 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lxn6d\" (UniqueName: \"kubernetes.io/projected/0aa6a534-c6a3-4a53-9e0e-053abec1ad18-kube-api-access-lxn6d\") pod \"glance-default-internal-api-1\" (UID: \"0aa6a534-c6a3-4a53-9e0e-053abec1ad18\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Dec 01 09:00:54 crc kubenswrapper[4813]: I1201 09:00:54.716978 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ng4mg\" (UniqueName: \"kubernetes.io/projected/a1ad44fa-8926-4242-8d07-1be670bb316a-kube-api-access-ng4mg\") pod \"glance-default-internal-api-2\" (UID: \"a1ad44fa-8926-4242-8d07-1be670bb316a\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Dec 01 09:00:54 crc kubenswrapper[4813]: I1201 09:00:54.718136 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-internal-api-1\" (UID: \"0aa6a534-c6a3-4a53-9e0e-053abec1ad18\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Dec 01 09:00:54 crc kubenswrapper[4813]: I1201 09:00:54.721580 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0aa6a534-c6a3-4a53-9e0e-053abec1ad18-config-data\") pod \"glance-default-internal-api-1\" (UID: \"0aa6a534-c6a3-4a53-9e0e-053abec1ad18\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Dec 01 09:00:54 crc kubenswrapper[4813]: I1201 09:00:54.728086 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage19-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage19-crc\") pod \"glance-default-internal-api-2\" (UID: \"a1ad44fa-8926-4242-8d07-1be670bb316a\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Dec 01 09:00:54 crc kubenswrapper[4813]: I1201 09:00:54.731820 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage18-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage18-crc\") pod \"glance-default-internal-api-1\" (UID: \"0aa6a534-c6a3-4a53-9e0e-053abec1ad18\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Dec 01 09:00:54 crc kubenswrapper[4813]: I1201 09:00:54.733522 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-internal-api-2\" (UID: \"a1ad44fa-8926-4242-8d07-1be670bb316a\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Dec 01 09:00:54 crc kubenswrapper[4813]: I1201 09:00:54.758070 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-1" Dec 01 09:00:54 crc kubenswrapper[4813]: I1201 09:00:54.776451 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-2" Dec 01 09:00:55 crc kubenswrapper[4813]: I1201 09:00:55.162945 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-2"] Dec 01 09:00:55 crc kubenswrapper[4813]: W1201 09:00:55.166686 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod423d0710_1fe1_4cd3_99a8_3c32c7f08ee3.slice/crio-9c6e09813fd442db9893498a25bea07881988ffcb4f35aa1f98e6098821204bf WatchSource:0}: Error finding container 9c6e09813fd442db9893498a25bea07881988ffcb4f35aa1f98e6098821204bf: Status 404 returned error can't find the container with id 9c6e09813fd442db9893498a25bea07881988ffcb4f35aa1f98e6098821204bf Dec 01 09:00:55 crc kubenswrapper[4813]: I1201 09:00:55.218517 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-1"] Dec 01 09:00:55 crc kubenswrapper[4813]: W1201 09:00:55.228268 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod04117ae2_817e_41a3_9d1e_608290bf79c8.slice/crio-bf3f286e8b96b9acd055929a895eb4acdd9ac6e3fe4ba996b2eb0f330fc26cb5 WatchSource:0}: Error finding container bf3f286e8b96b9acd055929a895eb4acdd9ac6e3fe4ba996b2eb0f330fc26cb5: Status 404 returned error can't find the container with id bf3f286e8b96b9acd055929a895eb4acdd9ac6e3fe4ba996b2eb0f330fc26cb5 Dec 01 09:00:55 crc kubenswrapper[4813]: W1201 09:00:55.267791 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0aa6a534_c6a3_4a53_9e0e_053abec1ad18.slice/crio-8f3aed550b27f97a08a10cee80d952c37af41981f48f8d45b75cd4a3c43e581d WatchSource:0}: Error finding container 8f3aed550b27f97a08a10cee80d952c37af41981f48f8d45b75cd4a3c43e581d: Status 404 returned error can't find the container with id 8f3aed550b27f97a08a10cee80d952c37af41981f48f8d45b75cd4a3c43e581d Dec 01 09:00:55 crc kubenswrapper[4813]: I1201 09:00:55.269010 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-1"] Dec 01 09:00:55 crc kubenswrapper[4813]: I1201 09:00:55.304091 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-2"] Dec 01 09:00:55 crc kubenswrapper[4813]: I1201 09:00:55.833663 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-1" event={"ID":"04117ae2-817e-41a3-9d1e-608290bf79c8","Type":"ContainerStarted","Data":"90e758fc6a5909bf47d08f54ce59ec6cc42413a4b45450a12c19628fd6b4b65a"} Dec 01 09:00:55 crc kubenswrapper[4813]: I1201 09:00:55.834251 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-1" event={"ID":"04117ae2-817e-41a3-9d1e-608290bf79c8","Type":"ContainerStarted","Data":"5d358a07cc765b1837f58741f50bb9830d64307b8dd0e5fdf8374abe188632f0"} Dec 01 09:00:55 crc kubenswrapper[4813]: I1201 09:00:55.834268 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-1" event={"ID":"04117ae2-817e-41a3-9d1e-608290bf79c8","Type":"ContainerStarted","Data":"bf3f286e8b96b9acd055929a895eb4acdd9ac6e3fe4ba996b2eb0f330fc26cb5"} Dec 01 09:00:55 crc kubenswrapper[4813]: I1201 09:00:55.839163 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-2" event={"ID":"423d0710-1fe1-4cd3-99a8-3c32c7f08ee3","Type":"ContainerStarted","Data":"90cbe765e7add7cded884a053b1d4011790b191b5bcf8f9ff77962aac2b08465"} Dec 01 09:00:55 crc kubenswrapper[4813]: I1201 09:00:55.842522 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-2" event={"ID":"423d0710-1fe1-4cd3-99a8-3c32c7f08ee3","Type":"ContainerStarted","Data":"fb9b27183e883119a77dd18a3b7a19b007eb008117aaf3be3b5e4b130ec25168"} Dec 01 09:00:55 crc kubenswrapper[4813]: I1201 09:00:55.843086 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-2" event={"ID":"423d0710-1fe1-4cd3-99a8-3c32c7f08ee3","Type":"ContainerStarted","Data":"9c6e09813fd442db9893498a25bea07881988ffcb4f35aa1f98e6098821204bf"} Dec 01 09:00:55 crc kubenswrapper[4813]: I1201 09:00:55.844873 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-2" event={"ID":"a1ad44fa-8926-4242-8d07-1be670bb316a","Type":"ContainerStarted","Data":"2492fb9e1af85b083b689d0c6632ed2df1d7490d5e150942607f61d53cb32927"} Dec 01 09:00:55 crc kubenswrapper[4813]: I1201 09:00:55.845473 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-2" event={"ID":"a1ad44fa-8926-4242-8d07-1be670bb316a","Type":"ContainerStarted","Data":"4b2c981db1be1cb0ad44d85ebd0a03f2804f09cc54bf129113bcdc31fc351ec1"} Dec 01 09:00:55 crc kubenswrapper[4813]: I1201 09:00:55.845627 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-2" event={"ID":"a1ad44fa-8926-4242-8d07-1be670bb316a","Type":"ContainerStarted","Data":"365d7029c577d0bb9289a1315fca0d7da073cf3f875c3459b9c3b30a7c4989e2"} Dec 01 09:00:55 crc kubenswrapper[4813]: I1201 09:00:55.846801 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-1" event={"ID":"0aa6a534-c6a3-4a53-9e0e-053abec1ad18","Type":"ContainerStarted","Data":"4944e053e3f36dfa018553d3360246ed1700a07826ef1257bc626a9403453457"} Dec 01 09:00:55 crc kubenswrapper[4813]: I1201 09:00:55.846852 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-1" event={"ID":"0aa6a534-c6a3-4a53-9e0e-053abec1ad18","Type":"ContainerStarted","Data":"2dc52cea20901792c12a12a1696045834e8c701e8121159246f5779bbccc129e"} Dec 01 09:00:55 crc kubenswrapper[4813]: I1201 09:00:55.846864 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-1" event={"ID":"0aa6a534-c6a3-4a53-9e0e-053abec1ad18","Type":"ContainerStarted","Data":"8f3aed550b27f97a08a10cee80d952c37af41981f48f8d45b75cd4a3c43e581d"} Dec 01 09:00:56 crc kubenswrapper[4813]: I1201 09:00:56.857641 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-1" event={"ID":"0aa6a534-c6a3-4a53-9e0e-053abec1ad18","Type":"ContainerStarted","Data":"e7119831013f4469ee90aac24417f23bf871d60a2f6a6fa406330a8ae5341961"} Dec 01 09:00:56 crc kubenswrapper[4813]: I1201 09:00:56.860268 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-1" event={"ID":"04117ae2-817e-41a3-9d1e-608290bf79c8","Type":"ContainerStarted","Data":"d146c532187e7bef8b1e6a46c551d1170628077d2e1064a0515ab6373ef7c582"} Dec 01 09:00:56 crc kubenswrapper[4813]: I1201 09:00:56.862278 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-2" event={"ID":"423d0710-1fe1-4cd3-99a8-3c32c7f08ee3","Type":"ContainerStarted","Data":"57e7ca31a120bf9b4028782b4fb2f42008ca564c5ed34445f45356f8934be07e"} Dec 01 09:00:56 crc kubenswrapper[4813]: I1201 09:00:56.864926 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-2" event={"ID":"a1ad44fa-8926-4242-8d07-1be670bb316a","Type":"ContainerStarted","Data":"79765df0acf0dc4c25b5cbef928bc5f3a577552a1ebbe77160f8ee8151390716"} Dec 01 09:00:56 crc kubenswrapper[4813]: I1201 09:00:56.918023 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-default-external-api-2" podStartSLOduration=3.917957584 podStartE2EDuration="3.917957584s" podCreationTimestamp="2025-12-01 09:00:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 09:00:56.91427463 +0000 UTC m=+1241.537196216" watchObservedRunningTime="2025-12-01 09:00:56.917957584 +0000 UTC m=+1241.540879170" Dec 01 09:00:56 crc kubenswrapper[4813]: I1201 09:00:56.919412 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-default-internal-api-1" podStartSLOduration=3.919404834 podStartE2EDuration="3.919404834s" podCreationTimestamp="2025-12-01 09:00:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 09:00:56.894355351 +0000 UTC m=+1241.517276947" watchObservedRunningTime="2025-12-01 09:00:56.919404834 +0000 UTC m=+1241.542326420" Dec 01 09:00:56 crc kubenswrapper[4813]: I1201 09:00:56.942302 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-default-external-api-1" podStartSLOduration=3.942262786 podStartE2EDuration="3.942262786s" podCreationTimestamp="2025-12-01 09:00:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 09:00:56.939472658 +0000 UTC m=+1241.562394244" watchObservedRunningTime="2025-12-01 09:00:56.942262786 +0000 UTC m=+1241.565184372" Dec 01 09:00:57 crc kubenswrapper[4813]: I1201 09:00:57.007134 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-default-internal-api-2" podStartSLOduration=4.007108165 podStartE2EDuration="4.007108165s" podCreationTimestamp="2025-12-01 09:00:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 09:00:56.980270633 +0000 UTC m=+1241.603192219" watchObservedRunningTime="2025-12-01 09:00:57.007108165 +0000 UTC m=+1241.630029751" Dec 01 09:01:00 crc kubenswrapper[4813]: I1201 09:01:00.129185 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/keystone-cron-29409661-442fd"] Dec 01 09:01:00 crc kubenswrapper[4813]: I1201 09:01:00.130868 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-cron-29409661-442fd" Dec 01 09:01:00 crc kubenswrapper[4813]: I1201 09:01:00.149531 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/keystone-cron-29409661-442fd"] Dec 01 09:01:00 crc kubenswrapper[4813]: I1201 09:01:00.214828 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b247a555-0805-4884-a0d5-d2799423bf94-config-data\") pod \"keystone-cron-29409661-442fd\" (UID: \"b247a555-0805-4884-a0d5-d2799423bf94\") " pod="glance-kuttl-tests/keystone-cron-29409661-442fd" Dec 01 09:01:00 crc kubenswrapper[4813]: I1201 09:01:00.214953 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gdwrh\" (UniqueName: \"kubernetes.io/projected/b247a555-0805-4884-a0d5-d2799423bf94-kube-api-access-gdwrh\") pod \"keystone-cron-29409661-442fd\" (UID: \"b247a555-0805-4884-a0d5-d2799423bf94\") " pod="glance-kuttl-tests/keystone-cron-29409661-442fd" Dec 01 09:01:00 crc kubenswrapper[4813]: I1201 09:01:00.215097 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/b247a555-0805-4884-a0d5-d2799423bf94-fernet-keys\") pod \"keystone-cron-29409661-442fd\" (UID: \"b247a555-0805-4884-a0d5-d2799423bf94\") " pod="glance-kuttl-tests/keystone-cron-29409661-442fd" Dec 01 09:01:00 crc kubenswrapper[4813]: I1201 09:01:00.316517 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/b247a555-0805-4884-a0d5-d2799423bf94-fernet-keys\") pod \"keystone-cron-29409661-442fd\" (UID: \"b247a555-0805-4884-a0d5-d2799423bf94\") " pod="glance-kuttl-tests/keystone-cron-29409661-442fd" Dec 01 09:01:00 crc kubenswrapper[4813]: I1201 09:01:00.316592 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b247a555-0805-4884-a0d5-d2799423bf94-config-data\") pod \"keystone-cron-29409661-442fd\" (UID: \"b247a555-0805-4884-a0d5-d2799423bf94\") " pod="glance-kuttl-tests/keystone-cron-29409661-442fd" Dec 01 09:01:00 crc kubenswrapper[4813]: I1201 09:01:00.316662 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gdwrh\" (UniqueName: \"kubernetes.io/projected/b247a555-0805-4884-a0d5-d2799423bf94-kube-api-access-gdwrh\") pod \"keystone-cron-29409661-442fd\" (UID: \"b247a555-0805-4884-a0d5-d2799423bf94\") " pod="glance-kuttl-tests/keystone-cron-29409661-442fd" Dec 01 09:01:00 crc kubenswrapper[4813]: I1201 09:01:00.329013 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/b247a555-0805-4884-a0d5-d2799423bf94-fernet-keys\") pod \"keystone-cron-29409661-442fd\" (UID: \"b247a555-0805-4884-a0d5-d2799423bf94\") " pod="glance-kuttl-tests/keystone-cron-29409661-442fd" Dec 01 09:01:00 crc kubenswrapper[4813]: I1201 09:01:00.329593 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b247a555-0805-4884-a0d5-d2799423bf94-config-data\") pod \"keystone-cron-29409661-442fd\" (UID: \"b247a555-0805-4884-a0d5-d2799423bf94\") " pod="glance-kuttl-tests/keystone-cron-29409661-442fd" Dec 01 09:01:00 crc kubenswrapper[4813]: I1201 09:01:00.333403 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gdwrh\" (UniqueName: \"kubernetes.io/projected/b247a555-0805-4884-a0d5-d2799423bf94-kube-api-access-gdwrh\") pod \"keystone-cron-29409661-442fd\" (UID: \"b247a555-0805-4884-a0d5-d2799423bf94\") " pod="glance-kuttl-tests/keystone-cron-29409661-442fd" Dec 01 09:01:00 crc kubenswrapper[4813]: I1201 09:01:00.450878 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-cron-29409661-442fd" Dec 01 09:01:00 crc kubenswrapper[4813]: I1201 09:01:00.869637 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/keystone-cron-29409661-442fd"] Dec 01 09:01:00 crc kubenswrapper[4813]: W1201 09:01:00.873359 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb247a555_0805_4884_a0d5_d2799423bf94.slice/crio-ce8951f87637a04c25ce0efcf5f932fe3f48c3f77748b8304c3ce02697cc3973 WatchSource:0}: Error finding container ce8951f87637a04c25ce0efcf5f932fe3f48c3f77748b8304c3ce02697cc3973: Status 404 returned error can't find the container with id ce8951f87637a04c25ce0efcf5f932fe3f48c3f77748b8304c3ce02697cc3973 Dec 01 09:01:00 crc kubenswrapper[4813]: I1201 09:01:00.928898 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-cron-29409661-442fd" event={"ID":"b247a555-0805-4884-a0d5-d2799423bf94","Type":"ContainerStarted","Data":"ce8951f87637a04c25ce0efcf5f932fe3f48c3f77748b8304c3ce02697cc3973"} Dec 01 09:01:01 crc kubenswrapper[4813]: I1201 09:01:01.941991 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-cron-29409661-442fd" event={"ID":"b247a555-0805-4884-a0d5-d2799423bf94","Type":"ContainerStarted","Data":"34f7e76fca6a7ec2ea3bd96b3cde58af3d3c87d1a190335eb7651ffc30db835e"} Dec 01 09:01:01 crc kubenswrapper[4813]: I1201 09:01:01.974358 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/keystone-cron-29409661-442fd" podStartSLOduration=1.974328853 podStartE2EDuration="1.974328853s" podCreationTimestamp="2025-12-01 09:01:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 09:01:01.95994643 +0000 UTC m=+1246.582868106" watchObservedRunningTime="2025-12-01 09:01:01.974328853 +0000 UTC m=+1246.597250479" Dec 01 09:01:03 crc kubenswrapper[4813]: I1201 09:01:03.959308 4813 generic.go:334] "Generic (PLEG): container finished" podID="b247a555-0805-4884-a0d5-d2799423bf94" containerID="34f7e76fca6a7ec2ea3bd96b3cde58af3d3c87d1a190335eb7651ffc30db835e" exitCode=0 Dec 01 09:01:03 crc kubenswrapper[4813]: I1201 09:01:03.959362 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-cron-29409661-442fd" event={"ID":"b247a555-0805-4884-a0d5-d2799423bf94","Type":"ContainerDied","Data":"34f7e76fca6a7ec2ea3bd96b3cde58af3d3c87d1a190335eb7651ffc30db835e"} Dec 01 09:01:04 crc kubenswrapper[4813]: I1201 09:01:04.667374 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-external-api-2" Dec 01 09:01:04 crc kubenswrapper[4813]: I1201 09:01:04.667737 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-external-api-2" Dec 01 09:01:04 crc kubenswrapper[4813]: I1201 09:01:04.667756 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-external-api-2" Dec 01 09:01:04 crc kubenswrapper[4813]: I1201 09:01:04.687125 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-external-api-1" Dec 01 09:01:04 crc kubenswrapper[4813]: I1201 09:01:04.687260 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-external-api-1" Dec 01 09:01:04 crc kubenswrapper[4813]: I1201 09:01:04.687284 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-external-api-1" Dec 01 09:01:04 crc kubenswrapper[4813]: I1201 09:01:04.702780 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-external-api-2" Dec 01 09:01:04 crc kubenswrapper[4813]: I1201 09:01:04.708404 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-external-api-2" Dec 01 09:01:04 crc kubenswrapper[4813]: I1201 09:01:04.718542 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-external-api-1" Dec 01 09:01:04 crc kubenswrapper[4813]: I1201 09:01:04.724851 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-external-api-2" Dec 01 09:01:04 crc kubenswrapper[4813]: I1201 09:01:04.728063 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-external-api-1" Dec 01 09:01:04 crc kubenswrapper[4813]: I1201 09:01:04.739937 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-external-api-1" Dec 01 09:01:04 crc kubenswrapper[4813]: I1201 09:01:04.759088 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-internal-api-1" Dec 01 09:01:04 crc kubenswrapper[4813]: I1201 09:01:04.759137 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-internal-api-1" Dec 01 09:01:04 crc kubenswrapper[4813]: I1201 09:01:04.759148 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-internal-api-1" Dec 01 09:01:04 crc kubenswrapper[4813]: I1201 09:01:04.778259 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-internal-api-2" Dec 01 09:01:04 crc kubenswrapper[4813]: I1201 09:01:04.778303 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-internal-api-2" Dec 01 09:01:04 crc kubenswrapper[4813]: I1201 09:01:04.778314 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-internal-api-2" Dec 01 09:01:04 crc kubenswrapper[4813]: I1201 09:01:04.780057 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-internal-api-1" Dec 01 09:01:04 crc kubenswrapper[4813]: I1201 09:01:04.787458 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-internal-api-1" Dec 01 09:01:04 crc kubenswrapper[4813]: I1201 09:01:04.799893 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-internal-api-1" Dec 01 09:01:04 crc kubenswrapper[4813]: I1201 09:01:04.814933 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-internal-api-2" Dec 01 09:01:04 crc kubenswrapper[4813]: I1201 09:01:04.818801 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-internal-api-2" Dec 01 09:01:04 crc kubenswrapper[4813]: I1201 09:01:04.833759 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-internal-api-2" Dec 01 09:01:04 crc kubenswrapper[4813]: I1201 09:01:04.967732 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-external-api-2" Dec 01 09:01:04 crc kubenswrapper[4813]: I1201 09:01:04.967766 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-internal-api-2" Dec 01 09:01:04 crc kubenswrapper[4813]: I1201 09:01:04.967777 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-external-api-2" Dec 01 09:01:04 crc kubenswrapper[4813]: I1201 09:01:04.967786 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-internal-api-1" Dec 01 09:01:04 crc kubenswrapper[4813]: I1201 09:01:04.967796 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-internal-api-2" Dec 01 09:01:04 crc kubenswrapper[4813]: I1201 09:01:04.967806 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-internal-api-2" Dec 01 09:01:04 crc kubenswrapper[4813]: I1201 09:01:04.967816 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-external-api-1" Dec 01 09:01:04 crc kubenswrapper[4813]: I1201 09:01:04.967826 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-internal-api-1" Dec 01 09:01:04 crc kubenswrapper[4813]: I1201 09:01:04.967834 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-external-api-1" Dec 01 09:01:04 crc kubenswrapper[4813]: I1201 09:01:04.967842 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-external-api-1" Dec 01 09:01:04 crc kubenswrapper[4813]: I1201 09:01:04.967850 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-internal-api-1" Dec 01 09:01:04 crc kubenswrapper[4813]: I1201 09:01:04.967857 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-external-api-2" Dec 01 09:01:04 crc kubenswrapper[4813]: I1201 09:01:04.981851 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-external-api-2" Dec 01 09:01:04 crc kubenswrapper[4813]: I1201 09:01:04.982298 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-external-api-2" Dec 01 09:01:04 crc kubenswrapper[4813]: I1201 09:01:04.985279 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-internal-api-1" Dec 01 09:01:04 crc kubenswrapper[4813]: I1201 09:01:04.986160 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-external-api-2" Dec 01 09:01:04 crc kubenswrapper[4813]: I1201 09:01:04.986430 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-external-api-1" Dec 01 09:01:04 crc kubenswrapper[4813]: I1201 09:01:04.992427 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-external-api-1" Dec 01 09:01:04 crc kubenswrapper[4813]: I1201 09:01:04.992612 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-internal-api-1" Dec 01 09:01:04 crc kubenswrapper[4813]: I1201 09:01:04.992679 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-internal-api-1" Dec 01 09:01:04 crc kubenswrapper[4813]: I1201 09:01:04.997633 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-external-api-1" Dec 01 09:01:04 crc kubenswrapper[4813]: I1201 09:01:04.999669 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-internal-api-2" Dec 01 09:01:05 crc kubenswrapper[4813]: I1201 09:01:05.001860 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-internal-api-2" Dec 01 09:01:05 crc kubenswrapper[4813]: I1201 09:01:05.013370 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-internal-api-2" Dec 01 09:01:05 crc kubenswrapper[4813]: I1201 09:01:05.292439 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-cron-29409661-442fd" Dec 01 09:01:05 crc kubenswrapper[4813]: I1201 09:01:05.353527 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gdwrh\" (UniqueName: \"kubernetes.io/projected/b247a555-0805-4884-a0d5-d2799423bf94-kube-api-access-gdwrh\") pod \"b247a555-0805-4884-a0d5-d2799423bf94\" (UID: \"b247a555-0805-4884-a0d5-d2799423bf94\") " Dec 01 09:01:05 crc kubenswrapper[4813]: I1201 09:01:05.353574 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b247a555-0805-4884-a0d5-d2799423bf94-config-data\") pod \"b247a555-0805-4884-a0d5-d2799423bf94\" (UID: \"b247a555-0805-4884-a0d5-d2799423bf94\") " Dec 01 09:01:05 crc kubenswrapper[4813]: I1201 09:01:05.353656 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/b247a555-0805-4884-a0d5-d2799423bf94-fernet-keys\") pod \"b247a555-0805-4884-a0d5-d2799423bf94\" (UID: \"b247a555-0805-4884-a0d5-d2799423bf94\") " Dec 01 09:01:05 crc kubenswrapper[4813]: I1201 09:01:05.359762 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b247a555-0805-4884-a0d5-d2799423bf94-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "b247a555-0805-4884-a0d5-d2799423bf94" (UID: "b247a555-0805-4884-a0d5-d2799423bf94"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:01:05 crc kubenswrapper[4813]: I1201 09:01:05.368073 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b247a555-0805-4884-a0d5-d2799423bf94-kube-api-access-gdwrh" (OuterVolumeSpecName: "kube-api-access-gdwrh") pod "b247a555-0805-4884-a0d5-d2799423bf94" (UID: "b247a555-0805-4884-a0d5-d2799423bf94"). InnerVolumeSpecName "kube-api-access-gdwrh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:01:05 crc kubenswrapper[4813]: I1201 09:01:05.396907 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b247a555-0805-4884-a0d5-d2799423bf94-config-data" (OuterVolumeSpecName: "config-data") pod "b247a555-0805-4884-a0d5-d2799423bf94" (UID: "b247a555-0805-4884-a0d5-d2799423bf94"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:01:05 crc kubenswrapper[4813]: I1201 09:01:05.456285 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gdwrh\" (UniqueName: \"kubernetes.io/projected/b247a555-0805-4884-a0d5-d2799423bf94-kube-api-access-gdwrh\") on node \"crc\" DevicePath \"\"" Dec 01 09:01:05 crc kubenswrapper[4813]: I1201 09:01:05.456328 4813 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b247a555-0805-4884-a0d5-d2799423bf94-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 09:01:05 crc kubenswrapper[4813]: I1201 09:01:05.456340 4813 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/b247a555-0805-4884-a0d5-d2799423bf94-fernet-keys\") on node \"crc\" DevicePath \"\"" Dec 01 09:01:05 crc kubenswrapper[4813]: I1201 09:01:05.976834 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-2"] Dec 01 09:01:05 crc kubenswrapper[4813]: I1201 09:01:05.980489 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-cron-29409661-442fd" event={"ID":"b247a555-0805-4884-a0d5-d2799423bf94","Type":"ContainerDied","Data":"ce8951f87637a04c25ce0efcf5f932fe3f48c3f77748b8304c3ce02697cc3973"} Dec 01 09:01:05 crc kubenswrapper[4813]: I1201 09:01:05.980544 4813 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ce8951f87637a04c25ce0efcf5f932fe3f48c3f77748b8304c3ce02697cc3973" Dec 01 09:01:05 crc kubenswrapper[4813]: I1201 09:01:05.980676 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-cron-29409661-442fd" Dec 01 09:01:05 crc kubenswrapper[4813]: I1201 09:01:05.993827 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-1"] Dec 01 09:01:06 crc kubenswrapper[4813]: I1201 09:01:06.278521 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-2"] Dec 01 09:01:06 crc kubenswrapper[4813]: I1201 09:01:06.287315 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-1"] Dec 01 09:01:06 crc kubenswrapper[4813]: I1201 09:01:06.989187 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-external-api-2" podUID="423d0710-1fe1-4cd3-99a8-3c32c7f08ee3" containerName="glance-api" containerID="cri-o://57e7ca31a120bf9b4028782b4fb2f42008ca564c5ed34445f45356f8934be07e" gracePeriod=30 Dec 01 09:01:06 crc kubenswrapper[4813]: I1201 09:01:06.989222 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-external-api-2" podUID="423d0710-1fe1-4cd3-99a8-3c32c7f08ee3" containerName="glance-httpd" containerID="cri-o://90cbe765e7add7cded884a053b1d4011790b191b5bcf8f9ff77962aac2b08465" gracePeriod=30 Dec 01 09:01:06 crc kubenswrapper[4813]: I1201 09:01:06.989187 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-external-api-2" podUID="423d0710-1fe1-4cd3-99a8-3c32c7f08ee3" containerName="glance-log" containerID="cri-o://fb9b27183e883119a77dd18a3b7a19b007eb008117aaf3be3b5e4b130ec25168" gracePeriod=30 Dec 01 09:01:08 crc kubenswrapper[4813]: I1201 09:01:08.011029 4813 generic.go:334] "Generic (PLEG): container finished" podID="423d0710-1fe1-4cd3-99a8-3c32c7f08ee3" containerID="57e7ca31a120bf9b4028782b4fb2f42008ca564c5ed34445f45356f8934be07e" exitCode=0 Dec 01 09:01:08 crc kubenswrapper[4813]: I1201 09:01:08.011367 4813 generic.go:334] "Generic (PLEG): container finished" podID="423d0710-1fe1-4cd3-99a8-3c32c7f08ee3" containerID="90cbe765e7add7cded884a053b1d4011790b191b5bcf8f9ff77962aac2b08465" exitCode=0 Dec 01 09:01:08 crc kubenswrapper[4813]: I1201 09:01:08.011378 4813 generic.go:334] "Generic (PLEG): container finished" podID="423d0710-1fe1-4cd3-99a8-3c32c7f08ee3" containerID="fb9b27183e883119a77dd18a3b7a19b007eb008117aaf3be3b5e4b130ec25168" exitCode=143 Dec 01 09:01:08 crc kubenswrapper[4813]: I1201 09:01:08.011116 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-2" event={"ID":"423d0710-1fe1-4cd3-99a8-3c32c7f08ee3","Type":"ContainerDied","Data":"57e7ca31a120bf9b4028782b4fb2f42008ca564c5ed34445f45356f8934be07e"} Dec 01 09:01:08 crc kubenswrapper[4813]: I1201 09:01:08.011563 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-2" event={"ID":"423d0710-1fe1-4cd3-99a8-3c32c7f08ee3","Type":"ContainerDied","Data":"90cbe765e7add7cded884a053b1d4011790b191b5bcf8f9ff77962aac2b08465"} Dec 01 09:01:08 crc kubenswrapper[4813]: I1201 09:01:08.011575 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-2" event={"ID":"423d0710-1fe1-4cd3-99a8-3c32c7f08ee3","Type":"ContainerDied","Data":"fb9b27183e883119a77dd18a3b7a19b007eb008117aaf3be3b5e4b130ec25168"} Dec 01 09:01:08 crc kubenswrapper[4813]: I1201 09:01:08.011750 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-external-api-1" podUID="04117ae2-817e-41a3-9d1e-608290bf79c8" containerName="glance-log" containerID="cri-o://5d358a07cc765b1837f58741f50bb9830d64307b8dd0e5fdf8374abe188632f0" gracePeriod=30 Dec 01 09:01:08 crc kubenswrapper[4813]: I1201 09:01:08.012017 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-external-api-1" podUID="04117ae2-817e-41a3-9d1e-608290bf79c8" containerName="glance-api" containerID="cri-o://d146c532187e7bef8b1e6a46c551d1170628077d2e1064a0515ab6373ef7c582" gracePeriod=30 Dec 01 09:01:08 crc kubenswrapper[4813]: I1201 09:01:08.012057 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-external-api-1" podUID="04117ae2-817e-41a3-9d1e-608290bf79c8" containerName="glance-httpd" containerID="cri-o://90e758fc6a5909bf47d08f54ce59ec6cc42413a4b45450a12c19628fd6b4b65a" gracePeriod=30 Dec 01 09:01:08 crc kubenswrapper[4813]: I1201 09:01:08.012238 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-internal-api-2" podUID="a1ad44fa-8926-4242-8d07-1be670bb316a" containerName="glance-log" containerID="cri-o://4b2c981db1be1cb0ad44d85ebd0a03f2804f09cc54bf129113bcdc31fc351ec1" gracePeriod=30 Dec 01 09:01:08 crc kubenswrapper[4813]: I1201 09:01:08.012389 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-internal-api-2" podUID="a1ad44fa-8926-4242-8d07-1be670bb316a" containerName="glance-httpd" containerID="cri-o://2492fb9e1af85b083b689d0c6632ed2df1d7490d5e150942607f61d53cb32927" gracePeriod=30 Dec 01 09:01:08 crc kubenswrapper[4813]: I1201 09:01:08.012389 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-internal-api-2" podUID="a1ad44fa-8926-4242-8d07-1be670bb316a" containerName="glance-api" containerID="cri-o://79765df0acf0dc4c25b5cbef928bc5f3a577552a1ebbe77160f8ee8151390716" gracePeriod=30 Dec 01 09:01:08 crc kubenswrapper[4813]: I1201 09:01:08.012461 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-internal-api-1" podUID="0aa6a534-c6a3-4a53-9e0e-053abec1ad18" containerName="glance-log" containerID="cri-o://2dc52cea20901792c12a12a1696045834e8c701e8121159246f5779bbccc129e" gracePeriod=30 Dec 01 09:01:08 crc kubenswrapper[4813]: I1201 09:01:08.012497 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-internal-api-1" podUID="0aa6a534-c6a3-4a53-9e0e-053abec1ad18" containerName="glance-api" containerID="cri-o://e7119831013f4469ee90aac24417f23bf871d60a2f6a6fa406330a8ae5341961" gracePeriod=30 Dec 01 09:01:08 crc kubenswrapper[4813]: I1201 09:01:08.012499 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-internal-api-1" podUID="0aa6a534-c6a3-4a53-9e0e-053abec1ad18" containerName="glance-httpd" containerID="cri-o://4944e053e3f36dfa018553d3360246ed1700a07826ef1257bc626a9403453457" gracePeriod=30 Dec 01 09:01:08 crc kubenswrapper[4813]: I1201 09:01:08.208019 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-2" Dec 01 09:01:08 crc kubenswrapper[4813]: I1201 09:01:08.314661 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/423d0710-1fe1-4cd3-99a8-3c32c7f08ee3-httpd-run\") pod \"423d0710-1fe1-4cd3-99a8-3c32c7f08ee3\" (UID: \"423d0710-1fe1-4cd3-99a8-3c32c7f08ee3\") " Dec 01 09:01:08 crc kubenswrapper[4813]: I1201 09:01:08.314733 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/423d0710-1fe1-4cd3-99a8-3c32c7f08ee3-run\") pod \"423d0710-1fe1-4cd3-99a8-3c32c7f08ee3\" (UID: \"423d0710-1fe1-4cd3-99a8-3c32c7f08ee3\") " Dec 01 09:01:08 crc kubenswrapper[4813]: I1201 09:01:08.314779 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/423d0710-1fe1-4cd3-99a8-3c32c7f08ee3-var-locks-brick\") pod \"423d0710-1fe1-4cd3-99a8-3c32c7f08ee3\" (UID: \"423d0710-1fe1-4cd3-99a8-3c32c7f08ee3\") " Dec 01 09:01:08 crc kubenswrapper[4813]: I1201 09:01:08.314868 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/423d0710-1fe1-4cd3-99a8-3c32c7f08ee3-sys\") pod \"423d0710-1fe1-4cd3-99a8-3c32c7f08ee3\" (UID: \"423d0710-1fe1-4cd3-99a8-3c32c7f08ee3\") " Dec 01 09:01:08 crc kubenswrapper[4813]: I1201 09:01:08.314914 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance-cache\" (UniqueName: \"kubernetes.io/local-volume/local-storage14-crc\") pod \"423d0710-1fe1-4cd3-99a8-3c32c7f08ee3\" (UID: \"423d0710-1fe1-4cd3-99a8-3c32c7f08ee3\") " Dec 01 09:01:08 crc kubenswrapper[4813]: I1201 09:01:08.314931 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/423d0710-1fe1-4cd3-99a8-3c32c7f08ee3-lib-modules\") pod \"423d0710-1fe1-4cd3-99a8-3c32c7f08ee3\" (UID: \"423d0710-1fe1-4cd3-99a8-3c32c7f08ee3\") " Dec 01 09:01:08 crc kubenswrapper[4813]: I1201 09:01:08.314930 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/423d0710-1fe1-4cd3-99a8-3c32c7f08ee3-run" (OuterVolumeSpecName: "run") pod "423d0710-1fe1-4cd3-99a8-3c32c7f08ee3" (UID: "423d0710-1fe1-4cd3-99a8-3c32c7f08ee3"). InnerVolumeSpecName "run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 09:01:08 crc kubenswrapper[4813]: I1201 09:01:08.315007 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/423d0710-1fe1-4cd3-99a8-3c32c7f08ee3-sys" (OuterVolumeSpecName: "sys") pod "423d0710-1fe1-4cd3-99a8-3c32c7f08ee3" (UID: "423d0710-1fe1-4cd3-99a8-3c32c7f08ee3"). InnerVolumeSpecName "sys". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 09:01:08 crc kubenswrapper[4813]: I1201 09:01:08.315015 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/423d0710-1fe1-4cd3-99a8-3c32c7f08ee3-logs\") pod \"423d0710-1fe1-4cd3-99a8-3c32c7f08ee3\" (UID: \"423d0710-1fe1-4cd3-99a8-3c32c7f08ee3\") " Dec 01 09:01:08 crc kubenswrapper[4813]: I1201 09:01:08.315029 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/423d0710-1fe1-4cd3-99a8-3c32c7f08ee3-var-locks-brick" (OuterVolumeSpecName: "var-locks-brick") pod "423d0710-1fe1-4cd3-99a8-3c32c7f08ee3" (UID: "423d0710-1fe1-4cd3-99a8-3c32c7f08ee3"). InnerVolumeSpecName "var-locks-brick". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 09:01:08 crc kubenswrapper[4813]: I1201 09:01:08.315034 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/423d0710-1fe1-4cd3-99a8-3c32c7f08ee3-etc-nvme\") pod \"423d0710-1fe1-4cd3-99a8-3c32c7f08ee3\" (UID: \"423d0710-1fe1-4cd3-99a8-3c32c7f08ee3\") " Dec 01 09:01:08 crc kubenswrapper[4813]: I1201 09:01:08.315048 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/423d0710-1fe1-4cd3-99a8-3c32c7f08ee3-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "423d0710-1fe1-4cd3-99a8-3c32c7f08ee3" (UID: "423d0710-1fe1-4cd3-99a8-3c32c7f08ee3"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 09:01:08 crc kubenswrapper[4813]: I1201 09:01:08.315084 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/423d0710-1fe1-4cd3-99a8-3c32c7f08ee3-etc-nvme" (OuterVolumeSpecName: "etc-nvme") pod "423d0710-1fe1-4cd3-99a8-3c32c7f08ee3" (UID: "423d0710-1fe1-4cd3-99a8-3c32c7f08ee3"). InnerVolumeSpecName "etc-nvme". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 09:01:08 crc kubenswrapper[4813]: I1201 09:01:08.315109 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/423d0710-1fe1-4cd3-99a8-3c32c7f08ee3-dev\") pod \"423d0710-1fe1-4cd3-99a8-3c32c7f08ee3\" (UID: \"423d0710-1fe1-4cd3-99a8-3c32c7f08ee3\") " Dec 01 09:01:08 crc kubenswrapper[4813]: I1201 09:01:08.315115 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/423d0710-1fe1-4cd3-99a8-3c32c7f08ee3-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "423d0710-1fe1-4cd3-99a8-3c32c7f08ee3" (UID: "423d0710-1fe1-4cd3-99a8-3c32c7f08ee3"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 09:01:08 crc kubenswrapper[4813]: I1201 09:01:08.315149 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/423d0710-1fe1-4cd3-99a8-3c32c7f08ee3-etc-iscsi\") pod \"423d0710-1fe1-4cd3-99a8-3c32c7f08ee3\" (UID: \"423d0710-1fe1-4cd3-99a8-3c32c7f08ee3\") " Dec 01 09:01:08 crc kubenswrapper[4813]: I1201 09:01:08.315243 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kf7jh\" (UniqueName: \"kubernetes.io/projected/423d0710-1fe1-4cd3-99a8-3c32c7f08ee3-kube-api-access-kf7jh\") pod \"423d0710-1fe1-4cd3-99a8-3c32c7f08ee3\" (UID: \"423d0710-1fe1-4cd3-99a8-3c32c7f08ee3\") " Dec 01 09:01:08 crc kubenswrapper[4813]: I1201 09:01:08.315274 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/423d0710-1fe1-4cd3-99a8-3c32c7f08ee3-config-data\") pod \"423d0710-1fe1-4cd3-99a8-3c32c7f08ee3\" (UID: \"423d0710-1fe1-4cd3-99a8-3c32c7f08ee3\") " Dec 01 09:01:08 crc kubenswrapper[4813]: I1201 09:01:08.315291 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/423d0710-1fe1-4cd3-99a8-3c32c7f08ee3-scripts\") pod \"423d0710-1fe1-4cd3-99a8-3c32c7f08ee3\" (UID: \"423d0710-1fe1-4cd3-99a8-3c32c7f08ee3\") " Dec 01 09:01:08 crc kubenswrapper[4813]: I1201 09:01:08.315311 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"423d0710-1fe1-4cd3-99a8-3c32c7f08ee3\" (UID: \"423d0710-1fe1-4cd3-99a8-3c32c7f08ee3\") " Dec 01 09:01:08 crc kubenswrapper[4813]: I1201 09:01:08.315378 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/423d0710-1fe1-4cd3-99a8-3c32c7f08ee3-logs" (OuterVolumeSpecName: "logs") pod "423d0710-1fe1-4cd3-99a8-3c32c7f08ee3" (UID: "423d0710-1fe1-4cd3-99a8-3c32c7f08ee3"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 09:01:08 crc kubenswrapper[4813]: I1201 09:01:08.315519 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/423d0710-1fe1-4cd3-99a8-3c32c7f08ee3-dev" (OuterVolumeSpecName: "dev") pod "423d0710-1fe1-4cd3-99a8-3c32c7f08ee3" (UID: "423d0710-1fe1-4cd3-99a8-3c32c7f08ee3"). InnerVolumeSpecName "dev". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 09:01:08 crc kubenswrapper[4813]: I1201 09:01:08.315641 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/423d0710-1fe1-4cd3-99a8-3c32c7f08ee3-etc-iscsi" (OuterVolumeSpecName: "etc-iscsi") pod "423d0710-1fe1-4cd3-99a8-3c32c7f08ee3" (UID: "423d0710-1fe1-4cd3-99a8-3c32c7f08ee3"). InnerVolumeSpecName "etc-iscsi". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 09:01:08 crc kubenswrapper[4813]: I1201 09:01:08.316099 4813 reconciler_common.go:293] "Volume detached for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/423d0710-1fe1-4cd3-99a8-3c32c7f08ee3-sys\") on node \"crc\" DevicePath \"\"" Dec 01 09:01:08 crc kubenswrapper[4813]: I1201 09:01:08.316117 4813 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/423d0710-1fe1-4cd3-99a8-3c32c7f08ee3-lib-modules\") on node \"crc\" DevicePath \"\"" Dec 01 09:01:08 crc kubenswrapper[4813]: I1201 09:01:08.316127 4813 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/423d0710-1fe1-4cd3-99a8-3c32c7f08ee3-logs\") on node \"crc\" DevicePath \"\"" Dec 01 09:01:08 crc kubenswrapper[4813]: I1201 09:01:08.316135 4813 reconciler_common.go:293] "Volume detached for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/423d0710-1fe1-4cd3-99a8-3c32c7f08ee3-etc-nvme\") on node \"crc\" DevicePath \"\"" Dec 01 09:01:08 crc kubenswrapper[4813]: I1201 09:01:08.316142 4813 reconciler_common.go:293] "Volume detached for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/423d0710-1fe1-4cd3-99a8-3c32c7f08ee3-dev\") on node \"crc\" DevicePath \"\"" Dec 01 09:01:08 crc kubenswrapper[4813]: I1201 09:01:08.316150 4813 reconciler_common.go:293] "Volume detached for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/423d0710-1fe1-4cd3-99a8-3c32c7f08ee3-etc-iscsi\") on node \"crc\" DevicePath \"\"" Dec 01 09:01:08 crc kubenswrapper[4813]: I1201 09:01:08.316158 4813 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/423d0710-1fe1-4cd3-99a8-3c32c7f08ee3-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 01 09:01:08 crc kubenswrapper[4813]: I1201 09:01:08.316169 4813 reconciler_common.go:293] "Volume detached for volume \"run\" (UniqueName: \"kubernetes.io/host-path/423d0710-1fe1-4cd3-99a8-3c32c7f08ee3-run\") on node \"crc\" DevicePath \"\"" Dec 01 09:01:08 crc kubenswrapper[4813]: I1201 09:01:08.316177 4813 reconciler_common.go:293] "Volume detached for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/423d0710-1fe1-4cd3-99a8-3c32c7f08ee3-var-locks-brick\") on node \"crc\" DevicePath \"\"" Dec 01 09:01:08 crc kubenswrapper[4813]: I1201 09:01:08.323174 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage07-crc" (OuterVolumeSpecName: "glance") pod "423d0710-1fe1-4cd3-99a8-3c32c7f08ee3" (UID: "423d0710-1fe1-4cd3-99a8-3c32c7f08ee3"). InnerVolumeSpecName "local-storage07-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 01 09:01:08 crc kubenswrapper[4813]: I1201 09:01:08.323175 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/423d0710-1fe1-4cd3-99a8-3c32c7f08ee3-scripts" (OuterVolumeSpecName: "scripts") pod "423d0710-1fe1-4cd3-99a8-3c32c7f08ee3" (UID: "423d0710-1fe1-4cd3-99a8-3c32c7f08ee3"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:01:08 crc kubenswrapper[4813]: I1201 09:01:08.323245 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage14-crc" (OuterVolumeSpecName: "glance-cache") pod "423d0710-1fe1-4cd3-99a8-3c32c7f08ee3" (UID: "423d0710-1fe1-4cd3-99a8-3c32c7f08ee3"). InnerVolumeSpecName "local-storage14-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 01 09:01:08 crc kubenswrapper[4813]: I1201 09:01:08.323252 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/423d0710-1fe1-4cd3-99a8-3c32c7f08ee3-kube-api-access-kf7jh" (OuterVolumeSpecName: "kube-api-access-kf7jh") pod "423d0710-1fe1-4cd3-99a8-3c32c7f08ee3" (UID: "423d0710-1fe1-4cd3-99a8-3c32c7f08ee3"). InnerVolumeSpecName "kube-api-access-kf7jh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:01:08 crc kubenswrapper[4813]: I1201 09:01:08.387547 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/423d0710-1fe1-4cd3-99a8-3c32c7f08ee3-config-data" (OuterVolumeSpecName: "config-data") pod "423d0710-1fe1-4cd3-99a8-3c32c7f08ee3" (UID: "423d0710-1fe1-4cd3-99a8-3c32c7f08ee3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:01:08 crc kubenswrapper[4813]: I1201 09:01:08.417819 4813 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage14-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage14-crc\") on node \"crc\" " Dec 01 09:01:08 crc kubenswrapper[4813]: I1201 09:01:08.417852 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kf7jh\" (UniqueName: \"kubernetes.io/projected/423d0710-1fe1-4cd3-99a8-3c32c7f08ee3-kube-api-access-kf7jh\") on node \"crc\" DevicePath \"\"" Dec 01 09:01:08 crc kubenswrapper[4813]: I1201 09:01:08.417866 4813 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/423d0710-1fe1-4cd3-99a8-3c32c7f08ee3-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 09:01:08 crc kubenswrapper[4813]: I1201 09:01:08.417876 4813 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/423d0710-1fe1-4cd3-99a8-3c32c7f08ee3-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 09:01:08 crc kubenswrapper[4813]: I1201 09:01:08.417920 4813 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" " Dec 01 09:01:08 crc kubenswrapper[4813]: I1201 09:01:08.433259 4813 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage14-crc" (UniqueName: "kubernetes.io/local-volume/local-storage14-crc") on node "crc" Dec 01 09:01:08 crc kubenswrapper[4813]: I1201 09:01:08.435123 4813 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage07-crc" (UniqueName: "kubernetes.io/local-volume/local-storage07-crc") on node "crc" Dec 01 09:01:08 crc kubenswrapper[4813]: I1201 09:01:08.519649 4813 reconciler_common.go:293] "Volume detached for volume \"local-storage14-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage14-crc\") on node \"crc\" DevicePath \"\"" Dec 01 09:01:08 crc kubenswrapper[4813]: I1201 09:01:08.519689 4813 reconciler_common.go:293] "Volume detached for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" DevicePath \"\"" Dec 01 09:01:08 crc kubenswrapper[4813]: I1201 09:01:08.854850 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-1" Dec 01 09:01:08 crc kubenswrapper[4813]: I1201 09:01:08.926570 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d8d9\" (UniqueName: \"kubernetes.io/projected/04117ae2-817e-41a3-9d1e-608290bf79c8-kube-api-access-2d8d9\") pod \"04117ae2-817e-41a3-9d1e-608290bf79c8\" (UID: \"04117ae2-817e-41a3-9d1e-608290bf79c8\") " Dec 01 09:01:08 crc kubenswrapper[4813]: I1201 09:01:08.926649 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/04117ae2-817e-41a3-9d1e-608290bf79c8-var-locks-brick\") pod \"04117ae2-817e-41a3-9d1e-608290bf79c8\" (UID: \"04117ae2-817e-41a3-9d1e-608290bf79c8\") " Dec 01 09:01:08 crc kubenswrapper[4813]: I1201 09:01:08.926726 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance-cache\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"04117ae2-817e-41a3-9d1e-608290bf79c8\" (UID: \"04117ae2-817e-41a3-9d1e-608290bf79c8\") " Dec 01 09:01:08 crc kubenswrapper[4813]: I1201 09:01:08.926753 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"04117ae2-817e-41a3-9d1e-608290bf79c8\" (UID: \"04117ae2-817e-41a3-9d1e-608290bf79c8\") " Dec 01 09:01:08 crc kubenswrapper[4813]: I1201 09:01:08.926821 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/04117ae2-817e-41a3-9d1e-608290bf79c8-etc-iscsi\") pod \"04117ae2-817e-41a3-9d1e-608290bf79c8\" (UID: \"04117ae2-817e-41a3-9d1e-608290bf79c8\") " Dec 01 09:01:08 crc kubenswrapper[4813]: I1201 09:01:08.926851 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/04117ae2-817e-41a3-9d1e-608290bf79c8-scripts\") pod \"04117ae2-817e-41a3-9d1e-608290bf79c8\" (UID: \"04117ae2-817e-41a3-9d1e-608290bf79c8\") " Dec 01 09:01:08 crc kubenswrapper[4813]: I1201 09:01:08.926911 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/04117ae2-817e-41a3-9d1e-608290bf79c8-lib-modules\") pod \"04117ae2-817e-41a3-9d1e-608290bf79c8\" (UID: \"04117ae2-817e-41a3-9d1e-608290bf79c8\") " Dec 01 09:01:08 crc kubenswrapper[4813]: I1201 09:01:08.926996 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/04117ae2-817e-41a3-9d1e-608290bf79c8-sys\") pod \"04117ae2-817e-41a3-9d1e-608290bf79c8\" (UID: \"04117ae2-817e-41a3-9d1e-608290bf79c8\") " Dec 01 09:01:08 crc kubenswrapper[4813]: I1201 09:01:08.927389 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/04117ae2-817e-41a3-9d1e-608290bf79c8-dev\") pod \"04117ae2-817e-41a3-9d1e-608290bf79c8\" (UID: \"04117ae2-817e-41a3-9d1e-608290bf79c8\") " Dec 01 09:01:08 crc kubenswrapper[4813]: I1201 09:01:08.927460 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/04117ae2-817e-41a3-9d1e-608290bf79c8-etc-nvme\") pod \"04117ae2-817e-41a3-9d1e-608290bf79c8\" (UID: \"04117ae2-817e-41a3-9d1e-608290bf79c8\") " Dec 01 09:01:08 crc kubenswrapper[4813]: I1201 09:01:08.927551 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/04117ae2-817e-41a3-9d1e-608290bf79c8-httpd-run\") pod \"04117ae2-817e-41a3-9d1e-608290bf79c8\" (UID: \"04117ae2-817e-41a3-9d1e-608290bf79c8\") " Dec 01 09:01:08 crc kubenswrapper[4813]: I1201 09:01:08.928178 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/04117ae2-817e-41a3-9d1e-608290bf79c8-run\") pod \"04117ae2-817e-41a3-9d1e-608290bf79c8\" (UID: \"04117ae2-817e-41a3-9d1e-608290bf79c8\") " Dec 01 09:01:08 crc kubenswrapper[4813]: I1201 09:01:08.928217 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/04117ae2-817e-41a3-9d1e-608290bf79c8-config-data\") pod \"04117ae2-817e-41a3-9d1e-608290bf79c8\" (UID: \"04117ae2-817e-41a3-9d1e-608290bf79c8\") " Dec 01 09:01:08 crc kubenswrapper[4813]: I1201 09:01:08.928238 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/04117ae2-817e-41a3-9d1e-608290bf79c8-logs\") pod \"04117ae2-817e-41a3-9d1e-608290bf79c8\" (UID: \"04117ae2-817e-41a3-9d1e-608290bf79c8\") " Dec 01 09:01:08 crc kubenswrapper[4813]: I1201 09:01:08.929013 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/04117ae2-817e-41a3-9d1e-608290bf79c8-logs" (OuterVolumeSpecName: "logs") pod "04117ae2-817e-41a3-9d1e-608290bf79c8" (UID: "04117ae2-817e-41a3-9d1e-608290bf79c8"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 09:01:08 crc kubenswrapper[4813]: I1201 09:01:08.929433 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/04117ae2-817e-41a3-9d1e-608290bf79c8-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "04117ae2-817e-41a3-9d1e-608290bf79c8" (UID: "04117ae2-817e-41a3-9d1e-608290bf79c8"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 09:01:08 crc kubenswrapper[4813]: I1201 09:01:08.929473 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/04117ae2-817e-41a3-9d1e-608290bf79c8-run" (OuterVolumeSpecName: "run") pod "04117ae2-817e-41a3-9d1e-608290bf79c8" (UID: "04117ae2-817e-41a3-9d1e-608290bf79c8"). InnerVolumeSpecName "run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 09:01:08 crc kubenswrapper[4813]: I1201 09:01:08.929507 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/04117ae2-817e-41a3-9d1e-608290bf79c8-etc-iscsi" (OuterVolumeSpecName: "etc-iscsi") pod "04117ae2-817e-41a3-9d1e-608290bf79c8" (UID: "04117ae2-817e-41a3-9d1e-608290bf79c8"). InnerVolumeSpecName "etc-iscsi". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 09:01:08 crc kubenswrapper[4813]: I1201 09:01:08.929561 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/04117ae2-817e-41a3-9d1e-608290bf79c8-sys" (OuterVolumeSpecName: "sys") pod "04117ae2-817e-41a3-9d1e-608290bf79c8" (UID: "04117ae2-817e-41a3-9d1e-608290bf79c8"). InnerVolumeSpecName "sys". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 09:01:08 crc kubenswrapper[4813]: I1201 09:01:08.929596 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/04117ae2-817e-41a3-9d1e-608290bf79c8-dev" (OuterVolumeSpecName: "dev") pod "04117ae2-817e-41a3-9d1e-608290bf79c8" (UID: "04117ae2-817e-41a3-9d1e-608290bf79c8"). InnerVolumeSpecName "dev". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 09:01:08 crc kubenswrapper[4813]: I1201 09:01:08.929675 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/04117ae2-817e-41a3-9d1e-608290bf79c8-var-locks-brick" (OuterVolumeSpecName: "var-locks-brick") pod "04117ae2-817e-41a3-9d1e-608290bf79c8" (UID: "04117ae2-817e-41a3-9d1e-608290bf79c8"). InnerVolumeSpecName "var-locks-brick". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 09:01:08 crc kubenswrapper[4813]: I1201 09:01:08.929747 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/04117ae2-817e-41a3-9d1e-608290bf79c8-etc-nvme" (OuterVolumeSpecName: "etc-nvme") pod "04117ae2-817e-41a3-9d1e-608290bf79c8" (UID: "04117ae2-817e-41a3-9d1e-608290bf79c8"). InnerVolumeSpecName "etc-nvme". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 09:01:08 crc kubenswrapper[4813]: I1201 09:01:08.929752 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/04117ae2-817e-41a3-9d1e-608290bf79c8-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "04117ae2-817e-41a3-9d1e-608290bf79c8" (UID: "04117ae2-817e-41a3-9d1e-608290bf79c8"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 09:01:08 crc kubenswrapper[4813]: I1201 09:01:08.933619 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/04117ae2-817e-41a3-9d1e-608290bf79c8-scripts" (OuterVolumeSpecName: "scripts") pod "04117ae2-817e-41a3-9d1e-608290bf79c8" (UID: "04117ae2-817e-41a3-9d1e-608290bf79c8"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:01:08 crc kubenswrapper[4813]: I1201 09:01:08.934074 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage04-crc" (OuterVolumeSpecName: "glance") pod "04117ae2-817e-41a3-9d1e-608290bf79c8" (UID: "04117ae2-817e-41a3-9d1e-608290bf79c8"). InnerVolumeSpecName "local-storage04-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 01 09:01:08 crc kubenswrapper[4813]: I1201 09:01:08.934362 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/04117ae2-817e-41a3-9d1e-608290bf79c8-kube-api-access-2d8d9" (OuterVolumeSpecName: "kube-api-access-2d8d9") pod "04117ae2-817e-41a3-9d1e-608290bf79c8" (UID: "04117ae2-817e-41a3-9d1e-608290bf79c8"). InnerVolumeSpecName "kube-api-access-2d8d9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:01:08 crc kubenswrapper[4813]: I1201 09:01:08.935071 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage03-crc" (OuterVolumeSpecName: "glance-cache") pod "04117ae2-817e-41a3-9d1e-608290bf79c8" (UID: "04117ae2-817e-41a3-9d1e-608290bf79c8"). InnerVolumeSpecName "local-storage03-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 01 09:01:08 crc kubenswrapper[4813]: I1201 09:01:08.997769 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-1" Dec 01 09:01:09 crc kubenswrapper[4813]: I1201 09:01:09.023812 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/04117ae2-817e-41a3-9d1e-608290bf79c8-config-data" (OuterVolumeSpecName: "config-data") pod "04117ae2-817e-41a3-9d1e-608290bf79c8" (UID: "04117ae2-817e-41a3-9d1e-608290bf79c8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:01:09 crc kubenswrapper[4813]: I1201 09:01:09.032255 4813 reconciler_common.go:293] "Volume detached for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/04117ae2-817e-41a3-9d1e-608290bf79c8-dev\") on node \"crc\" DevicePath \"\"" Dec 01 09:01:09 crc kubenswrapper[4813]: I1201 09:01:09.032295 4813 reconciler_common.go:293] "Volume detached for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/04117ae2-817e-41a3-9d1e-608290bf79c8-etc-nvme\") on node \"crc\" DevicePath \"\"" Dec 01 09:01:09 crc kubenswrapper[4813]: I1201 09:01:09.032308 4813 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/04117ae2-817e-41a3-9d1e-608290bf79c8-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 01 09:01:09 crc kubenswrapper[4813]: I1201 09:01:09.032319 4813 reconciler_common.go:293] "Volume detached for volume \"run\" (UniqueName: \"kubernetes.io/host-path/04117ae2-817e-41a3-9d1e-608290bf79c8-run\") on node \"crc\" DevicePath \"\"" Dec 01 09:01:09 crc kubenswrapper[4813]: I1201 09:01:09.032330 4813 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/04117ae2-817e-41a3-9d1e-608290bf79c8-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 09:01:09 crc kubenswrapper[4813]: I1201 09:01:09.032341 4813 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/04117ae2-817e-41a3-9d1e-608290bf79c8-logs\") on node \"crc\" DevicePath \"\"" Dec 01 09:01:09 crc kubenswrapper[4813]: I1201 09:01:09.032353 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d8d9\" (UniqueName: \"kubernetes.io/projected/04117ae2-817e-41a3-9d1e-608290bf79c8-kube-api-access-2d8d9\") on node \"crc\" DevicePath \"\"" Dec 01 09:01:09 crc kubenswrapper[4813]: I1201 09:01:09.032366 4813 reconciler_common.go:293] "Volume detached for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/04117ae2-817e-41a3-9d1e-608290bf79c8-var-locks-brick\") on node \"crc\" DevicePath \"\"" Dec 01 09:01:09 crc kubenswrapper[4813]: I1201 09:01:09.032399 4813 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" " Dec 01 09:01:09 crc kubenswrapper[4813]: I1201 09:01:09.032415 4813 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" " Dec 01 09:01:09 crc kubenswrapper[4813]: I1201 09:01:09.032426 4813 reconciler_common.go:293] "Volume detached for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/04117ae2-817e-41a3-9d1e-608290bf79c8-etc-iscsi\") on node \"crc\" DevicePath \"\"" Dec 01 09:01:09 crc kubenswrapper[4813]: I1201 09:01:09.032436 4813 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/04117ae2-817e-41a3-9d1e-608290bf79c8-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 09:01:09 crc kubenswrapper[4813]: I1201 09:01:09.032448 4813 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/04117ae2-817e-41a3-9d1e-608290bf79c8-lib-modules\") on node \"crc\" DevicePath \"\"" Dec 01 09:01:09 crc kubenswrapper[4813]: I1201 09:01:09.032462 4813 reconciler_common.go:293] "Volume detached for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/04117ae2-817e-41a3-9d1e-608290bf79c8-sys\") on node \"crc\" DevicePath \"\"" Dec 01 09:01:09 crc kubenswrapper[4813]: I1201 09:01:09.041802 4813 generic.go:334] "Generic (PLEG): container finished" podID="0aa6a534-c6a3-4a53-9e0e-053abec1ad18" containerID="e7119831013f4469ee90aac24417f23bf871d60a2f6a6fa406330a8ae5341961" exitCode=0 Dec 01 09:01:09 crc kubenswrapper[4813]: I1201 09:01:09.041844 4813 generic.go:334] "Generic (PLEG): container finished" podID="0aa6a534-c6a3-4a53-9e0e-053abec1ad18" containerID="4944e053e3f36dfa018553d3360246ed1700a07826ef1257bc626a9403453457" exitCode=0 Dec 01 09:01:09 crc kubenswrapper[4813]: I1201 09:01:09.041891 4813 generic.go:334] "Generic (PLEG): container finished" podID="0aa6a534-c6a3-4a53-9e0e-053abec1ad18" containerID="2dc52cea20901792c12a12a1696045834e8c701e8121159246f5779bbccc129e" exitCode=143 Dec 01 09:01:09 crc kubenswrapper[4813]: I1201 09:01:09.042320 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-1" event={"ID":"0aa6a534-c6a3-4a53-9e0e-053abec1ad18","Type":"ContainerDied","Data":"e7119831013f4469ee90aac24417f23bf871d60a2f6a6fa406330a8ae5341961"} Dec 01 09:01:09 crc kubenswrapper[4813]: I1201 09:01:09.042388 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-1" event={"ID":"0aa6a534-c6a3-4a53-9e0e-053abec1ad18","Type":"ContainerDied","Data":"4944e053e3f36dfa018553d3360246ed1700a07826ef1257bc626a9403453457"} Dec 01 09:01:09 crc kubenswrapper[4813]: I1201 09:01:09.042401 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-1" event={"ID":"0aa6a534-c6a3-4a53-9e0e-053abec1ad18","Type":"ContainerDied","Data":"2dc52cea20901792c12a12a1696045834e8c701e8121159246f5779bbccc129e"} Dec 01 09:01:09 crc kubenswrapper[4813]: I1201 09:01:09.042409 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-1" event={"ID":"0aa6a534-c6a3-4a53-9e0e-053abec1ad18","Type":"ContainerDied","Data":"8f3aed550b27f97a08a10cee80d952c37af41981f48f8d45b75cd4a3c43e581d"} Dec 01 09:01:09 crc kubenswrapper[4813]: I1201 09:01:09.042447 4813 scope.go:117] "RemoveContainer" containerID="e7119831013f4469ee90aac24417f23bf871d60a2f6a6fa406330a8ae5341961" Dec 01 09:01:09 crc kubenswrapper[4813]: I1201 09:01:09.042615 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-1" Dec 01 09:01:09 crc kubenswrapper[4813]: I1201 09:01:09.054431 4813 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage04-crc" (UniqueName: "kubernetes.io/local-volume/local-storage04-crc") on node "crc" Dec 01 09:01:09 crc kubenswrapper[4813]: I1201 09:01:09.062331 4813 generic.go:334] "Generic (PLEG): container finished" podID="04117ae2-817e-41a3-9d1e-608290bf79c8" containerID="d146c532187e7bef8b1e6a46c551d1170628077d2e1064a0515ab6373ef7c582" exitCode=0 Dec 01 09:01:09 crc kubenswrapper[4813]: I1201 09:01:09.062370 4813 generic.go:334] "Generic (PLEG): container finished" podID="04117ae2-817e-41a3-9d1e-608290bf79c8" containerID="90e758fc6a5909bf47d08f54ce59ec6cc42413a4b45450a12c19628fd6b4b65a" exitCode=0 Dec 01 09:01:09 crc kubenswrapper[4813]: I1201 09:01:09.062382 4813 generic.go:334] "Generic (PLEG): container finished" podID="04117ae2-817e-41a3-9d1e-608290bf79c8" containerID="5d358a07cc765b1837f58741f50bb9830d64307b8dd0e5fdf8374abe188632f0" exitCode=143 Dec 01 09:01:09 crc kubenswrapper[4813]: I1201 09:01:09.062429 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-1" event={"ID":"04117ae2-817e-41a3-9d1e-608290bf79c8","Type":"ContainerDied","Data":"d146c532187e7bef8b1e6a46c551d1170628077d2e1064a0515ab6373ef7c582"} Dec 01 09:01:09 crc kubenswrapper[4813]: I1201 09:01:09.062461 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-1" event={"ID":"04117ae2-817e-41a3-9d1e-608290bf79c8","Type":"ContainerDied","Data":"90e758fc6a5909bf47d08f54ce59ec6cc42413a4b45450a12c19628fd6b4b65a"} Dec 01 09:01:09 crc kubenswrapper[4813]: I1201 09:01:09.062476 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-1" event={"ID":"04117ae2-817e-41a3-9d1e-608290bf79c8","Type":"ContainerDied","Data":"5d358a07cc765b1837f58741f50bb9830d64307b8dd0e5fdf8374abe188632f0"} Dec 01 09:01:09 crc kubenswrapper[4813]: I1201 09:01:09.062487 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-1" event={"ID":"04117ae2-817e-41a3-9d1e-608290bf79c8","Type":"ContainerDied","Data":"bf3f286e8b96b9acd055929a895eb4acdd9ac6e3fe4ba996b2eb0f330fc26cb5"} Dec 01 09:01:09 crc kubenswrapper[4813]: I1201 09:01:09.062568 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-1" Dec 01 09:01:09 crc kubenswrapper[4813]: I1201 09:01:09.065118 4813 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage03-crc" (UniqueName: "kubernetes.io/local-volume/local-storage03-crc") on node "crc" Dec 01 09:01:09 crc kubenswrapper[4813]: I1201 09:01:09.073344 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-2" event={"ID":"423d0710-1fe1-4cd3-99a8-3c32c7f08ee3","Type":"ContainerDied","Data":"9c6e09813fd442db9893498a25bea07881988ffcb4f35aa1f98e6098821204bf"} Dec 01 09:01:09 crc kubenswrapper[4813]: I1201 09:01:09.073361 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-2" Dec 01 09:01:09 crc kubenswrapper[4813]: I1201 09:01:09.080778 4813 generic.go:334] "Generic (PLEG): container finished" podID="a1ad44fa-8926-4242-8d07-1be670bb316a" containerID="79765df0acf0dc4c25b5cbef928bc5f3a577552a1ebbe77160f8ee8151390716" exitCode=0 Dec 01 09:01:09 crc kubenswrapper[4813]: I1201 09:01:09.080812 4813 generic.go:334] "Generic (PLEG): container finished" podID="a1ad44fa-8926-4242-8d07-1be670bb316a" containerID="2492fb9e1af85b083b689d0c6632ed2df1d7490d5e150942607f61d53cb32927" exitCode=0 Dec 01 09:01:09 crc kubenswrapper[4813]: I1201 09:01:09.080821 4813 generic.go:334] "Generic (PLEG): container finished" podID="a1ad44fa-8926-4242-8d07-1be670bb316a" containerID="4b2c981db1be1cb0ad44d85ebd0a03f2804f09cc54bf129113bcdc31fc351ec1" exitCode=143 Dec 01 09:01:09 crc kubenswrapper[4813]: I1201 09:01:09.080849 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-2" event={"ID":"a1ad44fa-8926-4242-8d07-1be670bb316a","Type":"ContainerDied","Data":"79765df0acf0dc4c25b5cbef928bc5f3a577552a1ebbe77160f8ee8151390716"} Dec 01 09:01:09 crc kubenswrapper[4813]: I1201 09:01:09.080885 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-2" event={"ID":"a1ad44fa-8926-4242-8d07-1be670bb316a","Type":"ContainerDied","Data":"2492fb9e1af85b083b689d0c6632ed2df1d7490d5e150942607f61d53cb32927"} Dec 01 09:01:09 crc kubenswrapper[4813]: I1201 09:01:09.080902 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-2" event={"ID":"a1ad44fa-8926-4242-8d07-1be670bb316a","Type":"ContainerDied","Data":"4b2c981db1be1cb0ad44d85ebd0a03f2804f09cc54bf129113bcdc31fc351ec1"} Dec 01 09:01:09 crc kubenswrapper[4813]: I1201 09:01:09.113075 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-2"] Dec 01 09:01:09 crc kubenswrapper[4813]: I1201 09:01:09.135745 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-2"] Dec 01 09:01:09 crc kubenswrapper[4813]: I1201 09:01:09.136291 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0aa6a534-c6a3-4a53-9e0e-053abec1ad18-config-data\") pod \"0aa6a534-c6a3-4a53-9e0e-053abec1ad18\" (UID: \"0aa6a534-c6a3-4a53-9e0e-053abec1ad18\") " Dec 01 09:01:09 crc kubenswrapper[4813]: I1201 09:01:09.136347 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/0aa6a534-c6a3-4a53-9e0e-053abec1ad18-httpd-run\") pod \"0aa6a534-c6a3-4a53-9e0e-053abec1ad18\" (UID: \"0aa6a534-c6a3-4a53-9e0e-053abec1ad18\") " Dec 01 09:01:09 crc kubenswrapper[4813]: I1201 09:01:09.136424 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/0aa6a534-c6a3-4a53-9e0e-053abec1ad18-etc-nvme\") pod \"0aa6a534-c6a3-4a53-9e0e-053abec1ad18\" (UID: \"0aa6a534-c6a3-4a53-9e0e-053abec1ad18\") " Dec 01 09:01:09 crc kubenswrapper[4813]: I1201 09:01:09.136474 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/0aa6a534-c6a3-4a53-9e0e-053abec1ad18-sys\") pod \"0aa6a534-c6a3-4a53-9e0e-053abec1ad18\" (UID: \"0aa6a534-c6a3-4a53-9e0e-053abec1ad18\") " Dec 01 09:01:09 crc kubenswrapper[4813]: I1201 09:01:09.136511 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/0aa6a534-c6a3-4a53-9e0e-053abec1ad18-etc-iscsi\") pod \"0aa6a534-c6a3-4a53-9e0e-053abec1ad18\" (UID: \"0aa6a534-c6a3-4a53-9e0e-053abec1ad18\") " Dec 01 09:01:09 crc kubenswrapper[4813]: I1201 09:01:09.136561 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0aa6a534-c6a3-4a53-9e0e-053abec1ad18-scripts\") pod \"0aa6a534-c6a3-4a53-9e0e-053abec1ad18\" (UID: \"0aa6a534-c6a3-4a53-9e0e-053abec1ad18\") " Dec 01 09:01:09 crc kubenswrapper[4813]: I1201 09:01:09.136615 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lxn6d\" (UniqueName: \"kubernetes.io/projected/0aa6a534-c6a3-4a53-9e0e-053abec1ad18-kube-api-access-lxn6d\") pod \"0aa6a534-c6a3-4a53-9e0e-053abec1ad18\" (UID: \"0aa6a534-c6a3-4a53-9e0e-053abec1ad18\") " Dec 01 09:01:09 crc kubenswrapper[4813]: I1201 09:01:09.136640 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/0aa6a534-c6a3-4a53-9e0e-053abec1ad18-var-locks-brick\") pod \"0aa6a534-c6a3-4a53-9e0e-053abec1ad18\" (UID: \"0aa6a534-c6a3-4a53-9e0e-053abec1ad18\") " Dec 01 09:01:09 crc kubenswrapper[4813]: I1201 09:01:09.136688 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage18-crc\") pod \"0aa6a534-c6a3-4a53-9e0e-053abec1ad18\" (UID: \"0aa6a534-c6a3-4a53-9e0e-053abec1ad18\") " Dec 01 09:01:09 crc kubenswrapper[4813]: I1201 09:01:09.136736 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/0aa6a534-c6a3-4a53-9e0e-053abec1ad18-dev\") pod \"0aa6a534-c6a3-4a53-9e0e-053abec1ad18\" (UID: \"0aa6a534-c6a3-4a53-9e0e-053abec1ad18\") " Dec 01 09:01:09 crc kubenswrapper[4813]: I1201 09:01:09.136767 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/0aa6a534-c6a3-4a53-9e0e-053abec1ad18-lib-modules\") pod \"0aa6a534-c6a3-4a53-9e0e-053abec1ad18\" (UID: \"0aa6a534-c6a3-4a53-9e0e-053abec1ad18\") " Dec 01 09:01:09 crc kubenswrapper[4813]: I1201 09:01:09.136805 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0aa6a534-c6a3-4a53-9e0e-053abec1ad18-logs\") pod \"0aa6a534-c6a3-4a53-9e0e-053abec1ad18\" (UID: \"0aa6a534-c6a3-4a53-9e0e-053abec1ad18\") " Dec 01 09:01:09 crc kubenswrapper[4813]: I1201 09:01:09.136833 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/0aa6a534-c6a3-4a53-9e0e-053abec1ad18-run\") pod \"0aa6a534-c6a3-4a53-9e0e-053abec1ad18\" (UID: \"0aa6a534-c6a3-4a53-9e0e-053abec1ad18\") " Dec 01 09:01:09 crc kubenswrapper[4813]: I1201 09:01:09.136853 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance-cache\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"0aa6a534-c6a3-4a53-9e0e-053abec1ad18\" (UID: \"0aa6a534-c6a3-4a53-9e0e-053abec1ad18\") " Dec 01 09:01:09 crc kubenswrapper[4813]: I1201 09:01:09.137229 4813 reconciler_common.go:293] "Volume detached for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" DevicePath \"\"" Dec 01 09:01:09 crc kubenswrapper[4813]: I1201 09:01:09.137253 4813 reconciler_common.go:293] "Volume detached for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" DevicePath \"\"" Dec 01 09:01:09 crc kubenswrapper[4813]: I1201 09:01:09.137639 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0aa6a534-c6a3-4a53-9e0e-053abec1ad18-dev" (OuterVolumeSpecName: "dev") pod "0aa6a534-c6a3-4a53-9e0e-053abec1ad18" (UID: "0aa6a534-c6a3-4a53-9e0e-053abec1ad18"). InnerVolumeSpecName "dev". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 09:01:09 crc kubenswrapper[4813]: I1201 09:01:09.137708 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0aa6a534-c6a3-4a53-9e0e-053abec1ad18-var-locks-brick" (OuterVolumeSpecName: "var-locks-brick") pod "0aa6a534-c6a3-4a53-9e0e-053abec1ad18" (UID: "0aa6a534-c6a3-4a53-9e0e-053abec1ad18"). InnerVolumeSpecName "var-locks-brick". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 09:01:09 crc kubenswrapper[4813]: I1201 09:01:09.137793 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0aa6a534-c6a3-4a53-9e0e-053abec1ad18-etc-nvme" (OuterVolumeSpecName: "etc-nvme") pod "0aa6a534-c6a3-4a53-9e0e-053abec1ad18" (UID: "0aa6a534-c6a3-4a53-9e0e-053abec1ad18"). InnerVolumeSpecName "etc-nvme". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 09:01:09 crc kubenswrapper[4813]: I1201 09:01:09.141272 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0aa6a534-c6a3-4a53-9e0e-053abec1ad18-sys" (OuterVolumeSpecName: "sys") pod "0aa6a534-c6a3-4a53-9e0e-053abec1ad18" (UID: "0aa6a534-c6a3-4a53-9e0e-053abec1ad18"). InnerVolumeSpecName "sys". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 09:01:09 crc kubenswrapper[4813]: I1201 09:01:09.141342 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0aa6a534-c6a3-4a53-9e0e-053abec1ad18-etc-iscsi" (OuterVolumeSpecName: "etc-iscsi") pod "0aa6a534-c6a3-4a53-9e0e-053abec1ad18" (UID: "0aa6a534-c6a3-4a53-9e0e-053abec1ad18"). InnerVolumeSpecName "etc-iscsi". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 09:01:09 crc kubenswrapper[4813]: I1201 09:01:09.141656 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0aa6a534-c6a3-4a53-9e0e-053abec1ad18-logs" (OuterVolumeSpecName: "logs") pod "0aa6a534-c6a3-4a53-9e0e-053abec1ad18" (UID: "0aa6a534-c6a3-4a53-9e0e-053abec1ad18"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 09:01:09 crc kubenswrapper[4813]: I1201 09:01:09.141716 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0aa6a534-c6a3-4a53-9e0e-053abec1ad18-run" (OuterVolumeSpecName: "run") pod "0aa6a534-c6a3-4a53-9e0e-053abec1ad18" (UID: "0aa6a534-c6a3-4a53-9e0e-053abec1ad18"). InnerVolumeSpecName "run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 09:01:09 crc kubenswrapper[4813]: I1201 09:01:09.141656 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0aa6a534-c6a3-4a53-9e0e-053abec1ad18-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "0aa6a534-c6a3-4a53-9e0e-053abec1ad18" (UID: "0aa6a534-c6a3-4a53-9e0e-053abec1ad18"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 09:01:09 crc kubenswrapper[4813]: I1201 09:01:09.141766 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0aa6a534-c6a3-4a53-9e0e-053abec1ad18-kube-api-access-lxn6d" (OuterVolumeSpecName: "kube-api-access-lxn6d") pod "0aa6a534-c6a3-4a53-9e0e-053abec1ad18" (UID: "0aa6a534-c6a3-4a53-9e0e-053abec1ad18"). InnerVolumeSpecName "kube-api-access-lxn6d". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:01:09 crc kubenswrapper[4813]: I1201 09:01:09.142082 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0aa6a534-c6a3-4a53-9e0e-053abec1ad18-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "0aa6a534-c6a3-4a53-9e0e-053abec1ad18" (UID: "0aa6a534-c6a3-4a53-9e0e-053abec1ad18"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 09:01:09 crc kubenswrapper[4813]: I1201 09:01:09.142240 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage11-crc" (OuterVolumeSpecName: "glance-cache") pod "0aa6a534-c6a3-4a53-9e0e-053abec1ad18" (UID: "0aa6a534-c6a3-4a53-9e0e-053abec1ad18"). InnerVolumeSpecName "local-storage11-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 01 09:01:09 crc kubenswrapper[4813]: I1201 09:01:09.142398 4813 scope.go:117] "RemoveContainer" containerID="4944e053e3f36dfa018553d3360246ed1700a07826ef1257bc626a9403453457" Dec 01 09:01:09 crc kubenswrapper[4813]: I1201 09:01:09.143759 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0aa6a534-c6a3-4a53-9e0e-053abec1ad18-scripts" (OuterVolumeSpecName: "scripts") pod "0aa6a534-c6a3-4a53-9e0e-053abec1ad18" (UID: "0aa6a534-c6a3-4a53-9e0e-053abec1ad18"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:01:09 crc kubenswrapper[4813]: I1201 09:01:09.163879 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-1"] Dec 01 09:01:09 crc kubenswrapper[4813]: I1201 09:01:09.177623 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage18-crc" (OuterVolumeSpecName: "glance") pod "0aa6a534-c6a3-4a53-9e0e-053abec1ad18" (UID: "0aa6a534-c6a3-4a53-9e0e-053abec1ad18"). InnerVolumeSpecName "local-storage18-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 01 09:01:09 crc kubenswrapper[4813]: I1201 09:01:09.181421 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-1"] Dec 01 09:01:09 crc kubenswrapper[4813]: I1201 09:01:09.239866 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lxn6d\" (UniqueName: \"kubernetes.io/projected/0aa6a534-c6a3-4a53-9e0e-053abec1ad18-kube-api-access-lxn6d\") on node \"crc\" DevicePath \"\"" Dec 01 09:01:09 crc kubenswrapper[4813]: I1201 09:01:09.239901 4813 reconciler_common.go:293] "Volume detached for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/0aa6a534-c6a3-4a53-9e0e-053abec1ad18-var-locks-brick\") on node \"crc\" DevicePath \"\"" Dec 01 09:01:09 crc kubenswrapper[4813]: I1201 09:01:09.239934 4813 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage18-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage18-crc\") on node \"crc\" " Dec 01 09:01:09 crc kubenswrapper[4813]: I1201 09:01:09.239945 4813 reconciler_common.go:293] "Volume detached for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/0aa6a534-c6a3-4a53-9e0e-053abec1ad18-dev\") on node \"crc\" DevicePath \"\"" Dec 01 09:01:09 crc kubenswrapper[4813]: I1201 09:01:09.239953 4813 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/0aa6a534-c6a3-4a53-9e0e-053abec1ad18-lib-modules\") on node \"crc\" DevicePath \"\"" Dec 01 09:01:09 crc kubenswrapper[4813]: I1201 09:01:09.239960 4813 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0aa6a534-c6a3-4a53-9e0e-053abec1ad18-logs\") on node \"crc\" DevicePath \"\"" Dec 01 09:01:09 crc kubenswrapper[4813]: I1201 09:01:09.239982 4813 reconciler_common.go:293] "Volume detached for volume \"run\" (UniqueName: \"kubernetes.io/host-path/0aa6a534-c6a3-4a53-9e0e-053abec1ad18-run\") on node \"crc\" DevicePath \"\"" Dec 01 09:01:09 crc kubenswrapper[4813]: I1201 09:01:09.239993 4813 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" " Dec 01 09:01:09 crc kubenswrapper[4813]: I1201 09:01:09.240003 4813 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/0aa6a534-c6a3-4a53-9e0e-053abec1ad18-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 01 09:01:09 crc kubenswrapper[4813]: I1201 09:01:09.240010 4813 reconciler_common.go:293] "Volume detached for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/0aa6a534-c6a3-4a53-9e0e-053abec1ad18-etc-nvme\") on node \"crc\" DevicePath \"\"" Dec 01 09:01:09 crc kubenswrapper[4813]: I1201 09:01:09.240018 4813 reconciler_common.go:293] "Volume detached for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/0aa6a534-c6a3-4a53-9e0e-053abec1ad18-sys\") on node \"crc\" DevicePath \"\"" Dec 01 09:01:09 crc kubenswrapper[4813]: I1201 09:01:09.240027 4813 reconciler_common.go:293] "Volume detached for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/0aa6a534-c6a3-4a53-9e0e-053abec1ad18-etc-iscsi\") on node \"crc\" DevicePath \"\"" Dec 01 09:01:09 crc kubenswrapper[4813]: I1201 09:01:09.240035 4813 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0aa6a534-c6a3-4a53-9e0e-053abec1ad18-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 09:01:09 crc kubenswrapper[4813]: I1201 09:01:09.252940 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0aa6a534-c6a3-4a53-9e0e-053abec1ad18-config-data" (OuterVolumeSpecName: "config-data") pod "0aa6a534-c6a3-4a53-9e0e-053abec1ad18" (UID: "0aa6a534-c6a3-4a53-9e0e-053abec1ad18"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:01:09 crc kubenswrapper[4813]: I1201 09:01:09.255956 4813 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage18-crc" (UniqueName: "kubernetes.io/local-volume/local-storage18-crc") on node "crc" Dec 01 09:01:09 crc kubenswrapper[4813]: I1201 09:01:09.266748 4813 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage11-crc" (UniqueName: "kubernetes.io/local-volume/local-storage11-crc") on node "crc" Dec 01 09:01:09 crc kubenswrapper[4813]: I1201 09:01:09.313028 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-2" Dec 01 09:01:09 crc kubenswrapper[4813]: I1201 09:01:09.319143 4813 scope.go:117] "RemoveContainer" containerID="2dc52cea20901792c12a12a1696045834e8c701e8121159246f5779bbccc129e" Dec 01 09:01:09 crc kubenswrapper[4813]: I1201 09:01:09.337763 4813 scope.go:117] "RemoveContainer" containerID="e7119831013f4469ee90aac24417f23bf871d60a2f6a6fa406330a8ae5341961" Dec 01 09:01:09 crc kubenswrapper[4813]: E1201 09:01:09.338338 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e7119831013f4469ee90aac24417f23bf871d60a2f6a6fa406330a8ae5341961\": container with ID starting with e7119831013f4469ee90aac24417f23bf871d60a2f6a6fa406330a8ae5341961 not found: ID does not exist" containerID="e7119831013f4469ee90aac24417f23bf871d60a2f6a6fa406330a8ae5341961" Dec 01 09:01:09 crc kubenswrapper[4813]: I1201 09:01:09.338370 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e7119831013f4469ee90aac24417f23bf871d60a2f6a6fa406330a8ae5341961"} err="failed to get container status \"e7119831013f4469ee90aac24417f23bf871d60a2f6a6fa406330a8ae5341961\": rpc error: code = NotFound desc = could not find container \"e7119831013f4469ee90aac24417f23bf871d60a2f6a6fa406330a8ae5341961\": container with ID starting with e7119831013f4469ee90aac24417f23bf871d60a2f6a6fa406330a8ae5341961 not found: ID does not exist" Dec 01 09:01:09 crc kubenswrapper[4813]: I1201 09:01:09.338392 4813 scope.go:117] "RemoveContainer" containerID="4944e053e3f36dfa018553d3360246ed1700a07826ef1257bc626a9403453457" Dec 01 09:01:09 crc kubenswrapper[4813]: E1201 09:01:09.338805 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4944e053e3f36dfa018553d3360246ed1700a07826ef1257bc626a9403453457\": container with ID starting with 4944e053e3f36dfa018553d3360246ed1700a07826ef1257bc626a9403453457 not found: ID does not exist" containerID="4944e053e3f36dfa018553d3360246ed1700a07826ef1257bc626a9403453457" Dec 01 09:01:09 crc kubenswrapper[4813]: I1201 09:01:09.338828 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4944e053e3f36dfa018553d3360246ed1700a07826ef1257bc626a9403453457"} err="failed to get container status \"4944e053e3f36dfa018553d3360246ed1700a07826ef1257bc626a9403453457\": rpc error: code = NotFound desc = could not find container \"4944e053e3f36dfa018553d3360246ed1700a07826ef1257bc626a9403453457\": container with ID starting with 4944e053e3f36dfa018553d3360246ed1700a07826ef1257bc626a9403453457 not found: ID does not exist" Dec 01 09:01:09 crc kubenswrapper[4813]: I1201 09:01:09.338841 4813 scope.go:117] "RemoveContainer" containerID="2dc52cea20901792c12a12a1696045834e8c701e8121159246f5779bbccc129e" Dec 01 09:01:09 crc kubenswrapper[4813]: E1201 09:01:09.339194 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2dc52cea20901792c12a12a1696045834e8c701e8121159246f5779bbccc129e\": container with ID starting with 2dc52cea20901792c12a12a1696045834e8c701e8121159246f5779bbccc129e not found: ID does not exist" containerID="2dc52cea20901792c12a12a1696045834e8c701e8121159246f5779bbccc129e" Dec 01 09:01:09 crc kubenswrapper[4813]: I1201 09:01:09.339215 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2dc52cea20901792c12a12a1696045834e8c701e8121159246f5779bbccc129e"} err="failed to get container status \"2dc52cea20901792c12a12a1696045834e8c701e8121159246f5779bbccc129e\": rpc error: code = NotFound desc = could not find container \"2dc52cea20901792c12a12a1696045834e8c701e8121159246f5779bbccc129e\": container with ID starting with 2dc52cea20901792c12a12a1696045834e8c701e8121159246f5779bbccc129e not found: ID does not exist" Dec 01 09:01:09 crc kubenswrapper[4813]: I1201 09:01:09.339227 4813 scope.go:117] "RemoveContainer" containerID="e7119831013f4469ee90aac24417f23bf871d60a2f6a6fa406330a8ae5341961" Dec 01 09:01:09 crc kubenswrapper[4813]: I1201 09:01:09.339543 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e7119831013f4469ee90aac24417f23bf871d60a2f6a6fa406330a8ae5341961"} err="failed to get container status \"e7119831013f4469ee90aac24417f23bf871d60a2f6a6fa406330a8ae5341961\": rpc error: code = NotFound desc = could not find container \"e7119831013f4469ee90aac24417f23bf871d60a2f6a6fa406330a8ae5341961\": container with ID starting with e7119831013f4469ee90aac24417f23bf871d60a2f6a6fa406330a8ae5341961 not found: ID does not exist" Dec 01 09:01:09 crc kubenswrapper[4813]: I1201 09:01:09.339607 4813 scope.go:117] "RemoveContainer" containerID="4944e053e3f36dfa018553d3360246ed1700a07826ef1257bc626a9403453457" Dec 01 09:01:09 crc kubenswrapper[4813]: I1201 09:01:09.339986 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4944e053e3f36dfa018553d3360246ed1700a07826ef1257bc626a9403453457"} err="failed to get container status \"4944e053e3f36dfa018553d3360246ed1700a07826ef1257bc626a9403453457\": rpc error: code = NotFound desc = could not find container \"4944e053e3f36dfa018553d3360246ed1700a07826ef1257bc626a9403453457\": container with ID starting with 4944e053e3f36dfa018553d3360246ed1700a07826ef1257bc626a9403453457 not found: ID does not exist" Dec 01 09:01:09 crc kubenswrapper[4813]: I1201 09:01:09.340007 4813 scope.go:117] "RemoveContainer" containerID="2dc52cea20901792c12a12a1696045834e8c701e8121159246f5779bbccc129e" Dec 01 09:01:09 crc kubenswrapper[4813]: I1201 09:01:09.340761 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2dc52cea20901792c12a12a1696045834e8c701e8121159246f5779bbccc129e"} err="failed to get container status \"2dc52cea20901792c12a12a1696045834e8c701e8121159246f5779bbccc129e\": rpc error: code = NotFound desc = could not find container \"2dc52cea20901792c12a12a1696045834e8c701e8121159246f5779bbccc129e\": container with ID starting with 2dc52cea20901792c12a12a1696045834e8c701e8121159246f5779bbccc129e not found: ID does not exist" Dec 01 09:01:09 crc kubenswrapper[4813]: I1201 09:01:09.340817 4813 scope.go:117] "RemoveContainer" containerID="e7119831013f4469ee90aac24417f23bf871d60a2f6a6fa406330a8ae5341961" Dec 01 09:01:09 crc kubenswrapper[4813]: I1201 09:01:09.340891 4813 reconciler_common.go:293] "Volume detached for volume \"local-storage18-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage18-crc\") on node \"crc\" DevicePath \"\"" Dec 01 09:01:09 crc kubenswrapper[4813]: I1201 09:01:09.340912 4813 reconciler_common.go:293] "Volume detached for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" DevicePath \"\"" Dec 01 09:01:09 crc kubenswrapper[4813]: I1201 09:01:09.340921 4813 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0aa6a534-c6a3-4a53-9e0e-053abec1ad18-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 09:01:09 crc kubenswrapper[4813]: I1201 09:01:09.341422 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e7119831013f4469ee90aac24417f23bf871d60a2f6a6fa406330a8ae5341961"} err="failed to get container status \"e7119831013f4469ee90aac24417f23bf871d60a2f6a6fa406330a8ae5341961\": rpc error: code = NotFound desc = could not find container \"e7119831013f4469ee90aac24417f23bf871d60a2f6a6fa406330a8ae5341961\": container with ID starting with e7119831013f4469ee90aac24417f23bf871d60a2f6a6fa406330a8ae5341961 not found: ID does not exist" Dec 01 09:01:09 crc kubenswrapper[4813]: I1201 09:01:09.341446 4813 scope.go:117] "RemoveContainer" containerID="4944e053e3f36dfa018553d3360246ed1700a07826ef1257bc626a9403453457" Dec 01 09:01:09 crc kubenswrapper[4813]: I1201 09:01:09.341814 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4944e053e3f36dfa018553d3360246ed1700a07826ef1257bc626a9403453457"} err="failed to get container status \"4944e053e3f36dfa018553d3360246ed1700a07826ef1257bc626a9403453457\": rpc error: code = NotFound desc = could not find container \"4944e053e3f36dfa018553d3360246ed1700a07826ef1257bc626a9403453457\": container with ID starting with 4944e053e3f36dfa018553d3360246ed1700a07826ef1257bc626a9403453457 not found: ID does not exist" Dec 01 09:01:09 crc kubenswrapper[4813]: I1201 09:01:09.341831 4813 scope.go:117] "RemoveContainer" containerID="2dc52cea20901792c12a12a1696045834e8c701e8121159246f5779bbccc129e" Dec 01 09:01:09 crc kubenswrapper[4813]: I1201 09:01:09.342116 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2dc52cea20901792c12a12a1696045834e8c701e8121159246f5779bbccc129e"} err="failed to get container status \"2dc52cea20901792c12a12a1696045834e8c701e8121159246f5779bbccc129e\": rpc error: code = NotFound desc = could not find container \"2dc52cea20901792c12a12a1696045834e8c701e8121159246f5779bbccc129e\": container with ID starting with 2dc52cea20901792c12a12a1696045834e8c701e8121159246f5779bbccc129e not found: ID does not exist" Dec 01 09:01:09 crc kubenswrapper[4813]: I1201 09:01:09.342144 4813 scope.go:117] "RemoveContainer" containerID="d146c532187e7bef8b1e6a46c551d1170628077d2e1064a0515ab6373ef7c582" Dec 01 09:01:09 crc kubenswrapper[4813]: I1201 09:01:09.381811 4813 scope.go:117] "RemoveContainer" containerID="90e758fc6a5909bf47d08f54ce59ec6cc42413a4b45450a12c19628fd6b4b65a" Dec 01 09:01:09 crc kubenswrapper[4813]: I1201 09:01:09.384467 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-1"] Dec 01 09:01:09 crc kubenswrapper[4813]: I1201 09:01:09.398613 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-1"] Dec 01 09:01:09 crc kubenswrapper[4813]: I1201 09:01:09.404741 4813 scope.go:117] "RemoveContainer" containerID="5d358a07cc765b1837f58741f50bb9830d64307b8dd0e5fdf8374abe188632f0" Dec 01 09:01:09 crc kubenswrapper[4813]: I1201 09:01:09.422149 4813 scope.go:117] "RemoveContainer" containerID="d146c532187e7bef8b1e6a46c551d1170628077d2e1064a0515ab6373ef7c582" Dec 01 09:01:09 crc kubenswrapper[4813]: E1201 09:01:09.422717 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d146c532187e7bef8b1e6a46c551d1170628077d2e1064a0515ab6373ef7c582\": container with ID starting with d146c532187e7bef8b1e6a46c551d1170628077d2e1064a0515ab6373ef7c582 not found: ID does not exist" containerID="d146c532187e7bef8b1e6a46c551d1170628077d2e1064a0515ab6373ef7c582" Dec 01 09:01:09 crc kubenswrapper[4813]: I1201 09:01:09.422772 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d146c532187e7bef8b1e6a46c551d1170628077d2e1064a0515ab6373ef7c582"} err="failed to get container status \"d146c532187e7bef8b1e6a46c551d1170628077d2e1064a0515ab6373ef7c582\": rpc error: code = NotFound desc = could not find container \"d146c532187e7bef8b1e6a46c551d1170628077d2e1064a0515ab6373ef7c582\": container with ID starting with d146c532187e7bef8b1e6a46c551d1170628077d2e1064a0515ab6373ef7c582 not found: ID does not exist" Dec 01 09:01:09 crc kubenswrapper[4813]: I1201 09:01:09.422810 4813 scope.go:117] "RemoveContainer" containerID="90e758fc6a5909bf47d08f54ce59ec6cc42413a4b45450a12c19628fd6b4b65a" Dec 01 09:01:09 crc kubenswrapper[4813]: E1201 09:01:09.423206 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"90e758fc6a5909bf47d08f54ce59ec6cc42413a4b45450a12c19628fd6b4b65a\": container with ID starting with 90e758fc6a5909bf47d08f54ce59ec6cc42413a4b45450a12c19628fd6b4b65a not found: ID does not exist" containerID="90e758fc6a5909bf47d08f54ce59ec6cc42413a4b45450a12c19628fd6b4b65a" Dec 01 09:01:09 crc kubenswrapper[4813]: I1201 09:01:09.423255 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"90e758fc6a5909bf47d08f54ce59ec6cc42413a4b45450a12c19628fd6b4b65a"} err="failed to get container status \"90e758fc6a5909bf47d08f54ce59ec6cc42413a4b45450a12c19628fd6b4b65a\": rpc error: code = NotFound desc = could not find container \"90e758fc6a5909bf47d08f54ce59ec6cc42413a4b45450a12c19628fd6b4b65a\": container with ID starting with 90e758fc6a5909bf47d08f54ce59ec6cc42413a4b45450a12c19628fd6b4b65a not found: ID does not exist" Dec 01 09:01:09 crc kubenswrapper[4813]: I1201 09:01:09.423297 4813 scope.go:117] "RemoveContainer" containerID="5d358a07cc765b1837f58741f50bb9830d64307b8dd0e5fdf8374abe188632f0" Dec 01 09:01:09 crc kubenswrapper[4813]: E1201 09:01:09.423604 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5d358a07cc765b1837f58741f50bb9830d64307b8dd0e5fdf8374abe188632f0\": container with ID starting with 5d358a07cc765b1837f58741f50bb9830d64307b8dd0e5fdf8374abe188632f0 not found: ID does not exist" containerID="5d358a07cc765b1837f58741f50bb9830d64307b8dd0e5fdf8374abe188632f0" Dec 01 09:01:09 crc kubenswrapper[4813]: I1201 09:01:09.423632 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5d358a07cc765b1837f58741f50bb9830d64307b8dd0e5fdf8374abe188632f0"} err="failed to get container status \"5d358a07cc765b1837f58741f50bb9830d64307b8dd0e5fdf8374abe188632f0\": rpc error: code = NotFound desc = could not find container \"5d358a07cc765b1837f58741f50bb9830d64307b8dd0e5fdf8374abe188632f0\": container with ID starting with 5d358a07cc765b1837f58741f50bb9830d64307b8dd0e5fdf8374abe188632f0 not found: ID does not exist" Dec 01 09:01:09 crc kubenswrapper[4813]: I1201 09:01:09.423660 4813 scope.go:117] "RemoveContainer" containerID="d146c532187e7bef8b1e6a46c551d1170628077d2e1064a0515ab6373ef7c582" Dec 01 09:01:09 crc kubenswrapper[4813]: I1201 09:01:09.424011 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d146c532187e7bef8b1e6a46c551d1170628077d2e1064a0515ab6373ef7c582"} err="failed to get container status \"d146c532187e7bef8b1e6a46c551d1170628077d2e1064a0515ab6373ef7c582\": rpc error: code = NotFound desc = could not find container \"d146c532187e7bef8b1e6a46c551d1170628077d2e1064a0515ab6373ef7c582\": container with ID starting with d146c532187e7bef8b1e6a46c551d1170628077d2e1064a0515ab6373ef7c582 not found: ID does not exist" Dec 01 09:01:09 crc kubenswrapper[4813]: I1201 09:01:09.424038 4813 scope.go:117] "RemoveContainer" containerID="90e758fc6a5909bf47d08f54ce59ec6cc42413a4b45450a12c19628fd6b4b65a" Dec 01 09:01:09 crc kubenswrapper[4813]: I1201 09:01:09.424319 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"90e758fc6a5909bf47d08f54ce59ec6cc42413a4b45450a12c19628fd6b4b65a"} err="failed to get container status \"90e758fc6a5909bf47d08f54ce59ec6cc42413a4b45450a12c19628fd6b4b65a\": rpc error: code = NotFound desc = could not find container \"90e758fc6a5909bf47d08f54ce59ec6cc42413a4b45450a12c19628fd6b4b65a\": container with ID starting with 90e758fc6a5909bf47d08f54ce59ec6cc42413a4b45450a12c19628fd6b4b65a not found: ID does not exist" Dec 01 09:01:09 crc kubenswrapper[4813]: I1201 09:01:09.424339 4813 scope.go:117] "RemoveContainer" containerID="5d358a07cc765b1837f58741f50bb9830d64307b8dd0e5fdf8374abe188632f0" Dec 01 09:01:09 crc kubenswrapper[4813]: I1201 09:01:09.424694 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5d358a07cc765b1837f58741f50bb9830d64307b8dd0e5fdf8374abe188632f0"} err="failed to get container status \"5d358a07cc765b1837f58741f50bb9830d64307b8dd0e5fdf8374abe188632f0\": rpc error: code = NotFound desc = could not find container \"5d358a07cc765b1837f58741f50bb9830d64307b8dd0e5fdf8374abe188632f0\": container with ID starting with 5d358a07cc765b1837f58741f50bb9830d64307b8dd0e5fdf8374abe188632f0 not found: ID does not exist" Dec 01 09:01:09 crc kubenswrapper[4813]: I1201 09:01:09.424712 4813 scope.go:117] "RemoveContainer" containerID="d146c532187e7bef8b1e6a46c551d1170628077d2e1064a0515ab6373ef7c582" Dec 01 09:01:09 crc kubenswrapper[4813]: I1201 09:01:09.424992 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d146c532187e7bef8b1e6a46c551d1170628077d2e1064a0515ab6373ef7c582"} err="failed to get container status \"d146c532187e7bef8b1e6a46c551d1170628077d2e1064a0515ab6373ef7c582\": rpc error: code = NotFound desc = could not find container \"d146c532187e7bef8b1e6a46c551d1170628077d2e1064a0515ab6373ef7c582\": container with ID starting with d146c532187e7bef8b1e6a46c551d1170628077d2e1064a0515ab6373ef7c582 not found: ID does not exist" Dec 01 09:01:09 crc kubenswrapper[4813]: I1201 09:01:09.425018 4813 scope.go:117] "RemoveContainer" containerID="90e758fc6a5909bf47d08f54ce59ec6cc42413a4b45450a12c19628fd6b4b65a" Dec 01 09:01:09 crc kubenswrapper[4813]: I1201 09:01:09.425607 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"90e758fc6a5909bf47d08f54ce59ec6cc42413a4b45450a12c19628fd6b4b65a"} err="failed to get container status \"90e758fc6a5909bf47d08f54ce59ec6cc42413a4b45450a12c19628fd6b4b65a\": rpc error: code = NotFound desc = could not find container \"90e758fc6a5909bf47d08f54ce59ec6cc42413a4b45450a12c19628fd6b4b65a\": container with ID starting with 90e758fc6a5909bf47d08f54ce59ec6cc42413a4b45450a12c19628fd6b4b65a not found: ID does not exist" Dec 01 09:01:09 crc kubenswrapper[4813]: I1201 09:01:09.425627 4813 scope.go:117] "RemoveContainer" containerID="5d358a07cc765b1837f58741f50bb9830d64307b8dd0e5fdf8374abe188632f0" Dec 01 09:01:09 crc kubenswrapper[4813]: I1201 09:01:09.426008 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5d358a07cc765b1837f58741f50bb9830d64307b8dd0e5fdf8374abe188632f0"} err="failed to get container status \"5d358a07cc765b1837f58741f50bb9830d64307b8dd0e5fdf8374abe188632f0\": rpc error: code = NotFound desc = could not find container \"5d358a07cc765b1837f58741f50bb9830d64307b8dd0e5fdf8374abe188632f0\": container with ID starting with 5d358a07cc765b1837f58741f50bb9830d64307b8dd0e5fdf8374abe188632f0 not found: ID does not exist" Dec 01 09:01:09 crc kubenswrapper[4813]: I1201 09:01:09.426061 4813 scope.go:117] "RemoveContainer" containerID="57e7ca31a120bf9b4028782b4fb2f42008ca564c5ed34445f45356f8934be07e" Dec 01 09:01:09 crc kubenswrapper[4813]: I1201 09:01:09.442288 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/a1ad44fa-8926-4242-8d07-1be670bb316a-lib-modules\") pod \"a1ad44fa-8926-4242-8d07-1be670bb316a\" (UID: \"a1ad44fa-8926-4242-8d07-1be670bb316a\") " Dec 01 09:01:09 crc kubenswrapper[4813]: I1201 09:01:09.442341 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/a1ad44fa-8926-4242-8d07-1be670bb316a-run\") pod \"a1ad44fa-8926-4242-8d07-1be670bb316a\" (UID: \"a1ad44fa-8926-4242-8d07-1be670bb316a\") " Dec 01 09:01:09 crc kubenswrapper[4813]: I1201 09:01:09.442379 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a1ad44fa-8926-4242-8d07-1be670bb316a-scripts\") pod \"a1ad44fa-8926-4242-8d07-1be670bb316a\" (UID: \"a1ad44fa-8926-4242-8d07-1be670bb316a\") " Dec 01 09:01:09 crc kubenswrapper[4813]: I1201 09:01:09.442380 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a1ad44fa-8926-4242-8d07-1be670bb316a-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "a1ad44fa-8926-4242-8d07-1be670bb316a" (UID: "a1ad44fa-8926-4242-8d07-1be670bb316a"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 09:01:09 crc kubenswrapper[4813]: I1201 09:01:09.442416 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/a1ad44fa-8926-4242-8d07-1be670bb316a-httpd-run\") pod \"a1ad44fa-8926-4242-8d07-1be670bb316a\" (UID: \"a1ad44fa-8926-4242-8d07-1be670bb316a\") " Dec 01 09:01:09 crc kubenswrapper[4813]: I1201 09:01:09.442466 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/a1ad44fa-8926-4242-8d07-1be670bb316a-dev\") pod \"a1ad44fa-8926-4242-8d07-1be670bb316a\" (UID: \"a1ad44fa-8926-4242-8d07-1be670bb316a\") " Dec 01 09:01:09 crc kubenswrapper[4813]: I1201 09:01:09.442500 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a1ad44fa-8926-4242-8d07-1be670bb316a-run" (OuterVolumeSpecName: "run") pod "a1ad44fa-8926-4242-8d07-1be670bb316a" (UID: "a1ad44fa-8926-4242-8d07-1be670bb316a"). InnerVolumeSpecName "run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 09:01:09 crc kubenswrapper[4813]: I1201 09:01:09.442549 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a1ad44fa-8926-4242-8d07-1be670bb316a-dev" (OuterVolumeSpecName: "dev") pod "a1ad44fa-8926-4242-8d07-1be670bb316a" (UID: "a1ad44fa-8926-4242-8d07-1be670bb316a"). InnerVolumeSpecName "dev". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 09:01:09 crc kubenswrapper[4813]: I1201 09:01:09.442604 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ng4mg\" (UniqueName: \"kubernetes.io/projected/a1ad44fa-8926-4242-8d07-1be670bb316a-kube-api-access-ng4mg\") pod \"a1ad44fa-8926-4242-8d07-1be670bb316a\" (UID: \"a1ad44fa-8926-4242-8d07-1be670bb316a\") " Dec 01 09:01:09 crc kubenswrapper[4813]: I1201 09:01:09.442703 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a1ad44fa-8926-4242-8d07-1be670bb316a-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "a1ad44fa-8926-4242-8d07-1be670bb316a" (UID: "a1ad44fa-8926-4242-8d07-1be670bb316a"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 09:01:09 crc kubenswrapper[4813]: I1201 09:01:09.442978 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a1ad44fa-8926-4242-8d07-1be670bb316a-config-data\") pod \"a1ad44fa-8926-4242-8d07-1be670bb316a\" (UID: \"a1ad44fa-8926-4242-8d07-1be670bb316a\") " Dec 01 09:01:09 crc kubenswrapper[4813]: I1201 09:01:09.443260 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/a1ad44fa-8926-4242-8d07-1be670bb316a-etc-iscsi\") pod \"a1ad44fa-8926-4242-8d07-1be670bb316a\" (UID: \"a1ad44fa-8926-4242-8d07-1be670bb316a\") " Dec 01 09:01:09 crc kubenswrapper[4813]: I1201 09:01:09.443330 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a1ad44fa-8926-4242-8d07-1be670bb316a-etc-iscsi" (OuterVolumeSpecName: "etc-iscsi") pod "a1ad44fa-8926-4242-8d07-1be670bb316a" (UID: "a1ad44fa-8926-4242-8d07-1be670bb316a"). InnerVolumeSpecName "etc-iscsi". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 09:01:09 crc kubenswrapper[4813]: I1201 09:01:09.443295 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a1ad44fa-8926-4242-8d07-1be670bb316a-logs\") pod \"a1ad44fa-8926-4242-8d07-1be670bb316a\" (UID: \"a1ad44fa-8926-4242-8d07-1be670bb316a\") " Dec 01 09:01:09 crc kubenswrapper[4813]: I1201 09:01:09.443459 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/a1ad44fa-8926-4242-8d07-1be670bb316a-sys\") pod \"a1ad44fa-8926-4242-8d07-1be670bb316a\" (UID: \"a1ad44fa-8926-4242-8d07-1be670bb316a\") " Dec 01 09:01:09 crc kubenswrapper[4813]: I1201 09:01:09.443526 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a1ad44fa-8926-4242-8d07-1be670bb316a-sys" (OuterVolumeSpecName: "sys") pod "a1ad44fa-8926-4242-8d07-1be670bb316a" (UID: "a1ad44fa-8926-4242-8d07-1be670bb316a"). InnerVolumeSpecName "sys". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 09:01:09 crc kubenswrapper[4813]: I1201 09:01:09.443580 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance-cache\" (UniqueName: \"kubernetes.io/local-volume/local-storage19-crc\") pod \"a1ad44fa-8926-4242-8d07-1be670bb316a\" (UID: \"a1ad44fa-8926-4242-8d07-1be670bb316a\") " Dec 01 09:01:09 crc kubenswrapper[4813]: I1201 09:01:09.443599 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/a1ad44fa-8926-4242-8d07-1be670bb316a-etc-nvme\") pod \"a1ad44fa-8926-4242-8d07-1be670bb316a\" (UID: \"a1ad44fa-8926-4242-8d07-1be670bb316a\") " Dec 01 09:01:09 crc kubenswrapper[4813]: I1201 09:01:09.443610 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a1ad44fa-8926-4242-8d07-1be670bb316a-logs" (OuterVolumeSpecName: "logs") pod "a1ad44fa-8926-4242-8d07-1be670bb316a" (UID: "a1ad44fa-8926-4242-8d07-1be670bb316a"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 09:01:09 crc kubenswrapper[4813]: I1201 09:01:09.443614 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/a1ad44fa-8926-4242-8d07-1be670bb316a-var-locks-brick\") pod \"a1ad44fa-8926-4242-8d07-1be670bb316a\" (UID: \"a1ad44fa-8926-4242-8d07-1be670bb316a\") " Dec 01 09:01:09 crc kubenswrapper[4813]: I1201 09:01:09.443632 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a1ad44fa-8926-4242-8d07-1be670bb316a-var-locks-brick" (OuterVolumeSpecName: "var-locks-brick") pod "a1ad44fa-8926-4242-8d07-1be670bb316a" (UID: "a1ad44fa-8926-4242-8d07-1be670bb316a"). InnerVolumeSpecName "var-locks-brick". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 09:01:09 crc kubenswrapper[4813]: I1201 09:01:09.443673 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"a1ad44fa-8926-4242-8d07-1be670bb316a\" (UID: \"a1ad44fa-8926-4242-8d07-1be670bb316a\") " Dec 01 09:01:09 crc kubenswrapper[4813]: I1201 09:01:09.443831 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a1ad44fa-8926-4242-8d07-1be670bb316a-etc-nvme" (OuterVolumeSpecName: "etc-nvme") pod "a1ad44fa-8926-4242-8d07-1be670bb316a" (UID: "a1ad44fa-8926-4242-8d07-1be670bb316a"). InnerVolumeSpecName "etc-nvme". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 09:01:09 crc kubenswrapper[4813]: I1201 09:01:09.444264 4813 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a1ad44fa-8926-4242-8d07-1be670bb316a-logs\") on node \"crc\" DevicePath \"\"" Dec 01 09:01:09 crc kubenswrapper[4813]: I1201 09:01:09.444276 4813 reconciler_common.go:293] "Volume detached for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/a1ad44fa-8926-4242-8d07-1be670bb316a-sys\") on node \"crc\" DevicePath \"\"" Dec 01 09:01:09 crc kubenswrapper[4813]: I1201 09:01:09.444286 4813 reconciler_common.go:293] "Volume detached for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/a1ad44fa-8926-4242-8d07-1be670bb316a-var-locks-brick\") on node \"crc\" DevicePath \"\"" Dec 01 09:01:09 crc kubenswrapper[4813]: I1201 09:01:09.444297 4813 reconciler_common.go:293] "Volume detached for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/a1ad44fa-8926-4242-8d07-1be670bb316a-etc-nvme\") on node \"crc\" DevicePath \"\"" Dec 01 09:01:09 crc kubenswrapper[4813]: I1201 09:01:09.444306 4813 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/a1ad44fa-8926-4242-8d07-1be670bb316a-lib-modules\") on node \"crc\" DevicePath \"\"" Dec 01 09:01:09 crc kubenswrapper[4813]: I1201 09:01:09.444314 4813 reconciler_common.go:293] "Volume detached for volume \"run\" (UniqueName: \"kubernetes.io/host-path/a1ad44fa-8926-4242-8d07-1be670bb316a-run\") on node \"crc\" DevicePath \"\"" Dec 01 09:01:09 crc kubenswrapper[4813]: I1201 09:01:09.444322 4813 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/a1ad44fa-8926-4242-8d07-1be670bb316a-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 01 09:01:09 crc kubenswrapper[4813]: I1201 09:01:09.444331 4813 reconciler_common.go:293] "Volume detached for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/a1ad44fa-8926-4242-8d07-1be670bb316a-dev\") on node \"crc\" DevicePath \"\"" Dec 01 09:01:09 crc kubenswrapper[4813]: I1201 09:01:09.444339 4813 reconciler_common.go:293] "Volume detached for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/a1ad44fa-8926-4242-8d07-1be670bb316a-etc-iscsi\") on node \"crc\" DevicePath \"\"" Dec 01 09:01:09 crc kubenswrapper[4813]: I1201 09:01:09.445785 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a1ad44fa-8926-4242-8d07-1be670bb316a-scripts" (OuterVolumeSpecName: "scripts") pod "a1ad44fa-8926-4242-8d07-1be670bb316a" (UID: "a1ad44fa-8926-4242-8d07-1be670bb316a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:01:09 crc kubenswrapper[4813]: I1201 09:01:09.446378 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage08-crc" (OuterVolumeSpecName: "glance") pod "a1ad44fa-8926-4242-8d07-1be670bb316a" (UID: "a1ad44fa-8926-4242-8d07-1be670bb316a"). InnerVolumeSpecName "local-storage08-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 01 09:01:09 crc kubenswrapper[4813]: I1201 09:01:09.447092 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a1ad44fa-8926-4242-8d07-1be670bb316a-kube-api-access-ng4mg" (OuterVolumeSpecName: "kube-api-access-ng4mg") pod "a1ad44fa-8926-4242-8d07-1be670bb316a" (UID: "a1ad44fa-8926-4242-8d07-1be670bb316a"). InnerVolumeSpecName "kube-api-access-ng4mg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:01:09 crc kubenswrapper[4813]: I1201 09:01:09.449046 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage19-crc" (OuterVolumeSpecName: "glance-cache") pod "a1ad44fa-8926-4242-8d07-1be670bb316a" (UID: "a1ad44fa-8926-4242-8d07-1be670bb316a"). InnerVolumeSpecName "local-storage19-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 01 09:01:09 crc kubenswrapper[4813]: I1201 09:01:09.450806 4813 scope.go:117] "RemoveContainer" containerID="90cbe765e7add7cded884a053b1d4011790b191b5bcf8f9ff77962aac2b08465" Dec 01 09:01:09 crc kubenswrapper[4813]: I1201 09:01:09.470437 4813 scope.go:117] "RemoveContainer" containerID="fb9b27183e883119a77dd18a3b7a19b007eb008117aaf3be3b5e4b130ec25168" Dec 01 09:01:09 crc kubenswrapper[4813]: I1201 09:01:09.515239 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a1ad44fa-8926-4242-8d07-1be670bb316a-config-data" (OuterVolumeSpecName: "config-data") pod "a1ad44fa-8926-4242-8d07-1be670bb316a" (UID: "a1ad44fa-8926-4242-8d07-1be670bb316a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:01:09 crc kubenswrapper[4813]: I1201 09:01:09.545517 4813 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a1ad44fa-8926-4242-8d07-1be670bb316a-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 09:01:09 crc kubenswrapper[4813]: I1201 09:01:09.545552 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ng4mg\" (UniqueName: \"kubernetes.io/projected/a1ad44fa-8926-4242-8d07-1be670bb316a-kube-api-access-ng4mg\") on node \"crc\" DevicePath \"\"" Dec 01 09:01:09 crc kubenswrapper[4813]: I1201 09:01:09.545568 4813 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a1ad44fa-8926-4242-8d07-1be670bb316a-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 09:01:09 crc kubenswrapper[4813]: I1201 09:01:09.545603 4813 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage19-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage19-crc\") on node \"crc\" " Dec 01 09:01:09 crc kubenswrapper[4813]: I1201 09:01:09.545618 4813 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" " Dec 01 09:01:09 crc kubenswrapper[4813]: I1201 09:01:09.558986 4813 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage19-crc" (UniqueName: "kubernetes.io/local-volume/local-storage19-crc") on node "crc" Dec 01 09:01:09 crc kubenswrapper[4813]: I1201 09:01:09.560138 4813 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage08-crc" (UniqueName: "kubernetes.io/local-volume/local-storage08-crc") on node "crc" Dec 01 09:01:09 crc kubenswrapper[4813]: I1201 09:01:09.647348 4813 reconciler_common.go:293] "Volume detached for volume \"local-storage19-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage19-crc\") on node \"crc\" DevicePath \"\"" Dec 01 09:01:09 crc kubenswrapper[4813]: I1201 09:01:09.647380 4813 reconciler_common.go:293] "Volume detached for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" DevicePath \"\"" Dec 01 09:01:10 crc kubenswrapper[4813]: I1201 09:01:10.101421 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-2" event={"ID":"a1ad44fa-8926-4242-8d07-1be670bb316a","Type":"ContainerDied","Data":"365d7029c577d0bb9289a1315fca0d7da073cf3f875c3459b9c3b30a7c4989e2"} Dec 01 09:01:10 crc kubenswrapper[4813]: I1201 09:01:10.101481 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-2" Dec 01 09:01:10 crc kubenswrapper[4813]: I1201 09:01:10.101506 4813 scope.go:117] "RemoveContainer" containerID="79765df0acf0dc4c25b5cbef928bc5f3a577552a1ebbe77160f8ee8151390716" Dec 01 09:01:10 crc kubenswrapper[4813]: I1201 09:01:10.131151 4813 scope.go:117] "RemoveContainer" containerID="2492fb9e1af85b083b689d0c6632ed2df1d7490d5e150942607f61d53cb32927" Dec 01 09:01:10 crc kubenswrapper[4813]: I1201 09:01:10.145078 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-2"] Dec 01 09:01:10 crc kubenswrapper[4813]: I1201 09:01:10.152047 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-2"] Dec 01 09:01:10 crc kubenswrapper[4813]: I1201 09:01:10.205940 4813 scope.go:117] "RemoveContainer" containerID="4b2c981db1be1cb0ad44d85ebd0a03f2804f09cc54bf129113bcdc31fc351ec1" Dec 01 09:01:10 crc kubenswrapper[4813]: I1201 09:01:10.405205 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="04117ae2-817e-41a3-9d1e-608290bf79c8" path="/var/lib/kubelet/pods/04117ae2-817e-41a3-9d1e-608290bf79c8/volumes" Dec 01 09:01:10 crc kubenswrapper[4813]: I1201 09:01:10.407562 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0aa6a534-c6a3-4a53-9e0e-053abec1ad18" path="/var/lib/kubelet/pods/0aa6a534-c6a3-4a53-9e0e-053abec1ad18/volumes" Dec 01 09:01:10 crc kubenswrapper[4813]: I1201 09:01:10.409327 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="423d0710-1fe1-4cd3-99a8-3c32c7f08ee3" path="/var/lib/kubelet/pods/423d0710-1fe1-4cd3-99a8-3c32c7f08ee3/volumes" Dec 01 09:01:10 crc kubenswrapper[4813]: I1201 09:01:10.412304 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a1ad44fa-8926-4242-8d07-1be670bb316a" path="/var/lib/kubelet/pods/a1ad44fa-8926-4242-8d07-1be670bb316a/volumes" Dec 01 09:01:11 crc kubenswrapper[4813]: I1201 09:01:11.359174 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-0"] Dec 01 09:01:11 crc kubenswrapper[4813]: I1201 09:01:11.359542 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-external-api-0" podUID="e738e940-82c7-481a-8a7d-069215ae99c7" containerName="glance-log" containerID="cri-o://c7f8843d64fb50ae7e49ac4a05f42b1eac5e18fac3aeb36b3a1e4e47e9457f20" gracePeriod=30 Dec 01 09:01:11 crc kubenswrapper[4813]: I1201 09:01:11.359599 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-external-api-0" podUID="e738e940-82c7-481a-8a7d-069215ae99c7" containerName="glance-api" containerID="cri-o://c6d074d29e895601fa04da4eca3bc4fca8102be4e3f3ac5b7fa6de4021e4f833" gracePeriod=30 Dec 01 09:01:11 crc kubenswrapper[4813]: I1201 09:01:11.359613 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-external-api-0" podUID="e738e940-82c7-481a-8a7d-069215ae99c7" containerName="glance-httpd" containerID="cri-o://e3c4d4cd844c6982eeee8cf67db9ce5ebdfe25059f786d79a084320b82fe71d6" gracePeriod=30 Dec 01 09:01:11 crc kubenswrapper[4813]: I1201 09:01:11.788856 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Dec 01 09:01:11 crc kubenswrapper[4813]: I1201 09:01:11.789420 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-internal-api-0" podUID="5e438749-e7d5-4ca0-98e1-6d3319286720" containerName="glance-log" containerID="cri-o://dcb779b0095ccf2057b4fa7f52ea855e49228c7d0c78ae4beaafc5c8e1f860ba" gracePeriod=30 Dec 01 09:01:11 crc kubenswrapper[4813]: I1201 09:01:11.789671 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-internal-api-0" podUID="5e438749-e7d5-4ca0-98e1-6d3319286720" containerName="glance-api" containerID="cri-o://6604eac2164f12eccf3a2895e8fbaf218ce287738434acbba0a5cadfe7fcc82f" gracePeriod=30 Dec 01 09:01:11 crc kubenswrapper[4813]: I1201 09:01:11.789674 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-internal-api-0" podUID="5e438749-e7d5-4ca0-98e1-6d3319286720" containerName="glance-httpd" containerID="cri-o://0fba4493026f7e0bb5c77eac41f7dd9e9be04fb93081af5767525ccad569ecc3" gracePeriod=30 Dec 01 09:01:12 crc kubenswrapper[4813]: I1201 09:01:12.144114 4813 generic.go:334] "Generic (PLEG): container finished" podID="e738e940-82c7-481a-8a7d-069215ae99c7" containerID="c6d074d29e895601fa04da4eca3bc4fca8102be4e3f3ac5b7fa6de4021e4f833" exitCode=0 Dec 01 09:01:12 crc kubenswrapper[4813]: I1201 09:01:12.144152 4813 generic.go:334] "Generic (PLEG): container finished" podID="e738e940-82c7-481a-8a7d-069215ae99c7" containerID="e3c4d4cd844c6982eeee8cf67db9ce5ebdfe25059f786d79a084320b82fe71d6" exitCode=0 Dec 01 09:01:12 crc kubenswrapper[4813]: I1201 09:01:12.144166 4813 generic.go:334] "Generic (PLEG): container finished" podID="e738e940-82c7-481a-8a7d-069215ae99c7" containerID="c7f8843d64fb50ae7e49ac4a05f42b1eac5e18fac3aeb36b3a1e4e47e9457f20" exitCode=143 Dec 01 09:01:12 crc kubenswrapper[4813]: I1201 09:01:12.144199 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-0" event={"ID":"e738e940-82c7-481a-8a7d-069215ae99c7","Type":"ContainerDied","Data":"c6d074d29e895601fa04da4eca3bc4fca8102be4e3f3ac5b7fa6de4021e4f833"} Dec 01 09:01:12 crc kubenswrapper[4813]: I1201 09:01:12.144273 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-0" event={"ID":"e738e940-82c7-481a-8a7d-069215ae99c7","Type":"ContainerDied","Data":"e3c4d4cd844c6982eeee8cf67db9ce5ebdfe25059f786d79a084320b82fe71d6"} Dec 01 09:01:12 crc kubenswrapper[4813]: I1201 09:01:12.144328 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-0" event={"ID":"e738e940-82c7-481a-8a7d-069215ae99c7","Type":"ContainerDied","Data":"c7f8843d64fb50ae7e49ac4a05f42b1eac5e18fac3aeb36b3a1e4e47e9457f20"} Dec 01 09:01:12 crc kubenswrapper[4813]: I1201 09:01:12.147878 4813 generic.go:334] "Generic (PLEG): container finished" podID="5e438749-e7d5-4ca0-98e1-6d3319286720" containerID="0fba4493026f7e0bb5c77eac41f7dd9e9be04fb93081af5767525ccad569ecc3" exitCode=0 Dec 01 09:01:12 crc kubenswrapper[4813]: I1201 09:01:12.147912 4813 generic.go:334] "Generic (PLEG): container finished" podID="5e438749-e7d5-4ca0-98e1-6d3319286720" containerID="dcb779b0095ccf2057b4fa7f52ea855e49228c7d0c78ae4beaafc5c8e1f860ba" exitCode=143 Dec 01 09:01:12 crc kubenswrapper[4813]: I1201 09:01:12.147934 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"5e438749-e7d5-4ca0-98e1-6d3319286720","Type":"ContainerDied","Data":"0fba4493026f7e0bb5c77eac41f7dd9e9be04fb93081af5767525ccad569ecc3"} Dec 01 09:01:12 crc kubenswrapper[4813]: I1201 09:01:12.147973 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"5e438749-e7d5-4ca0-98e1-6d3319286720","Type":"ContainerDied","Data":"dcb779b0095ccf2057b4fa7f52ea855e49228c7d0c78ae4beaafc5c8e1f860ba"} Dec 01 09:01:12 crc kubenswrapper[4813]: I1201 09:01:12.448613 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-0" Dec 01 09:01:12 crc kubenswrapper[4813]: I1201 09:01:12.603487 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/e738e940-82c7-481a-8a7d-069215ae99c7-etc-nvme\") pod \"e738e940-82c7-481a-8a7d-069215ae99c7\" (UID: \"e738e940-82c7-481a-8a7d-069215ae99c7\") " Dec 01 09:01:12 crc kubenswrapper[4813]: I1201 09:01:12.603589 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance-cache\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"e738e940-82c7-481a-8a7d-069215ae99c7\" (UID: \"e738e940-82c7-481a-8a7d-069215ae99c7\") " Dec 01 09:01:12 crc kubenswrapper[4813]: I1201 09:01:12.603637 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/e738e940-82c7-481a-8a7d-069215ae99c7-run\") pod \"e738e940-82c7-481a-8a7d-069215ae99c7\" (UID: \"e738e940-82c7-481a-8a7d-069215ae99c7\") " Dec 01 09:01:12 crc kubenswrapper[4813]: I1201 09:01:12.603711 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/e738e940-82c7-481a-8a7d-069215ae99c7-lib-modules\") pod \"e738e940-82c7-481a-8a7d-069215ae99c7\" (UID: \"e738e940-82c7-481a-8a7d-069215ae99c7\") " Dec 01 09:01:12 crc kubenswrapper[4813]: I1201 09:01:12.603746 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/e738e940-82c7-481a-8a7d-069215ae99c7-etc-iscsi\") pod \"e738e940-82c7-481a-8a7d-069215ae99c7\" (UID: \"e738e940-82c7-481a-8a7d-069215ae99c7\") " Dec 01 09:01:12 crc kubenswrapper[4813]: I1201 09:01:12.603782 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/e738e940-82c7-481a-8a7d-069215ae99c7-sys\") pod \"e738e940-82c7-481a-8a7d-069215ae99c7\" (UID: \"e738e940-82c7-481a-8a7d-069215ae99c7\") " Dec 01 09:01:12 crc kubenswrapper[4813]: I1201 09:01:12.603831 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e738e940-82c7-481a-8a7d-069215ae99c7-logs\") pod \"e738e940-82c7-481a-8a7d-069215ae99c7\" (UID: \"e738e940-82c7-481a-8a7d-069215ae99c7\") " Dec 01 09:01:12 crc kubenswrapper[4813]: I1201 09:01:12.604135 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e738e940-82c7-481a-8a7d-069215ae99c7-run" (OuterVolumeSpecName: "run") pod "e738e940-82c7-481a-8a7d-069215ae99c7" (UID: "e738e940-82c7-481a-8a7d-069215ae99c7"). InnerVolumeSpecName "run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 09:01:12 crc kubenswrapper[4813]: I1201 09:01:12.604206 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e738e940-82c7-481a-8a7d-069215ae99c7-etc-iscsi" (OuterVolumeSpecName: "etc-iscsi") pod "e738e940-82c7-481a-8a7d-069215ae99c7" (UID: "e738e940-82c7-481a-8a7d-069215ae99c7"). InnerVolumeSpecName "etc-iscsi". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 09:01:12 crc kubenswrapper[4813]: I1201 09:01:12.604236 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e738e940-82c7-481a-8a7d-069215ae99c7-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "e738e940-82c7-481a-8a7d-069215ae99c7" (UID: "e738e940-82c7-481a-8a7d-069215ae99c7"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 09:01:12 crc kubenswrapper[4813]: I1201 09:01:12.604261 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e738e940-82c7-481a-8a7d-069215ae99c7-sys" (OuterVolumeSpecName: "sys") pod "e738e940-82c7-481a-8a7d-069215ae99c7" (UID: "e738e940-82c7-481a-8a7d-069215ae99c7"). InnerVolumeSpecName "sys". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 09:01:12 crc kubenswrapper[4813]: I1201 09:01:12.604172 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e738e940-82c7-481a-8a7d-069215ae99c7-etc-nvme" (OuterVolumeSpecName: "etc-nvme") pod "e738e940-82c7-481a-8a7d-069215ae99c7" (UID: "e738e940-82c7-481a-8a7d-069215ae99c7"). InnerVolumeSpecName "etc-nvme". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 09:01:12 crc kubenswrapper[4813]: I1201 09:01:12.604696 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e738e940-82c7-481a-8a7d-069215ae99c7-logs" (OuterVolumeSpecName: "logs") pod "e738e940-82c7-481a-8a7d-069215ae99c7" (UID: "e738e940-82c7-481a-8a7d-069215ae99c7"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 09:01:12 crc kubenswrapper[4813]: I1201 09:01:12.605122 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/e738e940-82c7-481a-8a7d-069215ae99c7-dev\") pod \"e738e940-82c7-481a-8a7d-069215ae99c7\" (UID: \"e738e940-82c7-481a-8a7d-069215ae99c7\") " Dec 01 09:01:12 crc kubenswrapper[4813]: I1201 09:01:12.605195 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/e738e940-82c7-481a-8a7d-069215ae99c7-var-locks-brick\") pod \"e738e940-82c7-481a-8a7d-069215ae99c7\" (UID: \"e738e940-82c7-481a-8a7d-069215ae99c7\") " Dec 01 09:01:12 crc kubenswrapper[4813]: I1201 09:01:12.605224 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e738e940-82c7-481a-8a7d-069215ae99c7-scripts\") pod \"e738e940-82c7-481a-8a7d-069215ae99c7\" (UID: \"e738e940-82c7-481a-8a7d-069215ae99c7\") " Dec 01 09:01:12 crc kubenswrapper[4813]: I1201 09:01:12.605274 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e738e940-82c7-481a-8a7d-069215ae99c7-dev" (OuterVolumeSpecName: "dev") pod "e738e940-82c7-481a-8a7d-069215ae99c7" (UID: "e738e940-82c7-481a-8a7d-069215ae99c7"). InnerVolumeSpecName "dev". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 09:01:12 crc kubenswrapper[4813]: I1201 09:01:12.605296 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9787f\" (UniqueName: \"kubernetes.io/projected/e738e940-82c7-481a-8a7d-069215ae99c7-kube-api-access-9787f\") pod \"e738e940-82c7-481a-8a7d-069215ae99c7\" (UID: \"e738e940-82c7-481a-8a7d-069215ae99c7\") " Dec 01 09:01:12 crc kubenswrapper[4813]: I1201 09:01:12.605313 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e738e940-82c7-481a-8a7d-069215ae99c7-var-locks-brick" (OuterVolumeSpecName: "var-locks-brick") pod "e738e940-82c7-481a-8a7d-069215ae99c7" (UID: "e738e940-82c7-481a-8a7d-069215ae99c7"). InnerVolumeSpecName "var-locks-brick". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 09:01:12 crc kubenswrapper[4813]: I1201 09:01:12.605342 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage13-crc\") pod \"e738e940-82c7-481a-8a7d-069215ae99c7\" (UID: \"e738e940-82c7-481a-8a7d-069215ae99c7\") " Dec 01 09:01:12 crc kubenswrapper[4813]: I1201 09:01:12.605371 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e738e940-82c7-481a-8a7d-069215ae99c7-config-data\") pod \"e738e940-82c7-481a-8a7d-069215ae99c7\" (UID: \"e738e940-82c7-481a-8a7d-069215ae99c7\") " Dec 01 09:01:12 crc kubenswrapper[4813]: I1201 09:01:12.605447 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e738e940-82c7-481a-8a7d-069215ae99c7-httpd-run\") pod \"e738e940-82c7-481a-8a7d-069215ae99c7\" (UID: \"e738e940-82c7-481a-8a7d-069215ae99c7\") " Dec 01 09:01:12 crc kubenswrapper[4813]: I1201 09:01:12.606154 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e738e940-82c7-481a-8a7d-069215ae99c7-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "e738e940-82c7-481a-8a7d-069215ae99c7" (UID: "e738e940-82c7-481a-8a7d-069215ae99c7"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 09:01:12 crc kubenswrapper[4813]: I1201 09:01:12.606366 4813 reconciler_common.go:293] "Volume detached for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/e738e940-82c7-481a-8a7d-069215ae99c7-etc-nvme\") on node \"crc\" DevicePath \"\"" Dec 01 09:01:12 crc kubenswrapper[4813]: I1201 09:01:12.606438 4813 reconciler_common.go:293] "Volume detached for volume \"run\" (UniqueName: \"kubernetes.io/host-path/e738e940-82c7-481a-8a7d-069215ae99c7-run\") on node \"crc\" DevicePath \"\"" Dec 01 09:01:12 crc kubenswrapper[4813]: I1201 09:01:12.606459 4813 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/e738e940-82c7-481a-8a7d-069215ae99c7-lib-modules\") on node \"crc\" DevicePath \"\"" Dec 01 09:01:12 crc kubenswrapper[4813]: I1201 09:01:12.606476 4813 reconciler_common.go:293] "Volume detached for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/e738e940-82c7-481a-8a7d-069215ae99c7-etc-iscsi\") on node \"crc\" DevicePath \"\"" Dec 01 09:01:12 crc kubenswrapper[4813]: I1201 09:01:12.606530 4813 reconciler_common.go:293] "Volume detached for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/e738e940-82c7-481a-8a7d-069215ae99c7-sys\") on node \"crc\" DevicePath \"\"" Dec 01 09:01:12 crc kubenswrapper[4813]: I1201 09:01:12.606546 4813 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e738e940-82c7-481a-8a7d-069215ae99c7-logs\") on node \"crc\" DevicePath \"\"" Dec 01 09:01:12 crc kubenswrapper[4813]: I1201 09:01:12.606565 4813 reconciler_common.go:293] "Volume detached for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/e738e940-82c7-481a-8a7d-069215ae99c7-dev\") on node \"crc\" DevicePath \"\"" Dec 01 09:01:12 crc kubenswrapper[4813]: I1201 09:01:12.606617 4813 reconciler_common.go:293] "Volume detached for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/e738e940-82c7-481a-8a7d-069215ae99c7-var-locks-brick\") on node \"crc\" DevicePath \"\"" Dec 01 09:01:12 crc kubenswrapper[4813]: I1201 09:01:12.606634 4813 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e738e940-82c7-481a-8a7d-069215ae99c7-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 01 09:01:12 crc kubenswrapper[4813]: I1201 09:01:12.609303 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage13-crc" (OuterVolumeSpecName: "glance") pod "e738e940-82c7-481a-8a7d-069215ae99c7" (UID: "e738e940-82c7-481a-8a7d-069215ae99c7"). InnerVolumeSpecName "local-storage13-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 01 09:01:12 crc kubenswrapper[4813]: I1201 09:01:12.609674 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage12-crc" (OuterVolumeSpecName: "glance-cache") pod "e738e940-82c7-481a-8a7d-069215ae99c7" (UID: "e738e940-82c7-481a-8a7d-069215ae99c7"). InnerVolumeSpecName "local-storage12-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 01 09:01:12 crc kubenswrapper[4813]: I1201 09:01:12.609906 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e738e940-82c7-481a-8a7d-069215ae99c7-kube-api-access-9787f" (OuterVolumeSpecName: "kube-api-access-9787f") pod "e738e940-82c7-481a-8a7d-069215ae99c7" (UID: "e738e940-82c7-481a-8a7d-069215ae99c7"). InnerVolumeSpecName "kube-api-access-9787f". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:01:12 crc kubenswrapper[4813]: I1201 09:01:12.611539 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e738e940-82c7-481a-8a7d-069215ae99c7-scripts" (OuterVolumeSpecName: "scripts") pod "e738e940-82c7-481a-8a7d-069215ae99c7" (UID: "e738e940-82c7-481a-8a7d-069215ae99c7"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:01:12 crc kubenswrapper[4813]: I1201 09:01:12.682804 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e738e940-82c7-481a-8a7d-069215ae99c7-config-data" (OuterVolumeSpecName: "config-data") pod "e738e940-82c7-481a-8a7d-069215ae99c7" (UID: "e738e940-82c7-481a-8a7d-069215ae99c7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:01:12 crc kubenswrapper[4813]: I1201 09:01:12.706361 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:01:12 crc kubenswrapper[4813]: I1201 09:01:12.707623 4813 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e738e940-82c7-481a-8a7d-069215ae99c7-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 09:01:12 crc kubenswrapper[4813]: I1201 09:01:12.707683 4813 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage13-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage13-crc\") on node \"crc\" " Dec 01 09:01:12 crc kubenswrapper[4813]: I1201 09:01:12.707698 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9787f\" (UniqueName: \"kubernetes.io/projected/e738e940-82c7-481a-8a7d-069215ae99c7-kube-api-access-9787f\") on node \"crc\" DevicePath \"\"" Dec 01 09:01:12 crc kubenswrapper[4813]: I1201 09:01:12.707712 4813 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e738e940-82c7-481a-8a7d-069215ae99c7-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 09:01:12 crc kubenswrapper[4813]: I1201 09:01:12.707729 4813 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" " Dec 01 09:01:12 crc kubenswrapper[4813]: I1201 09:01:12.721771 4813 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage12-crc" (UniqueName: "kubernetes.io/local-volume/local-storage12-crc") on node "crc" Dec 01 09:01:12 crc kubenswrapper[4813]: I1201 09:01:12.725763 4813 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage13-crc" (UniqueName: "kubernetes.io/local-volume/local-storage13-crc") on node "crc" Dec 01 09:01:12 crc kubenswrapper[4813]: I1201 09:01:12.809162 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance-cache\" (UniqueName: \"kubernetes.io/local-volume/local-storage17-crc\") pod \"5e438749-e7d5-4ca0-98e1-6d3319286720\" (UID: \"5e438749-e7d5-4ca0-98e1-6d3319286720\") " Dec 01 09:01:12 crc kubenswrapper[4813]: I1201 09:01:12.809257 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5e438749-e7d5-4ca0-98e1-6d3319286720-scripts\") pod \"5e438749-e7d5-4ca0-98e1-6d3319286720\" (UID: \"5e438749-e7d5-4ca0-98e1-6d3319286720\") " Dec 01 09:01:12 crc kubenswrapper[4813]: I1201 09:01:12.809293 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s2gqr\" (UniqueName: \"kubernetes.io/projected/5e438749-e7d5-4ca0-98e1-6d3319286720-kube-api-access-s2gqr\") pod \"5e438749-e7d5-4ca0-98e1-6d3319286720\" (UID: \"5e438749-e7d5-4ca0-98e1-6d3319286720\") " Dec 01 09:01:12 crc kubenswrapper[4813]: I1201 09:01:12.809336 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/5e438749-e7d5-4ca0-98e1-6d3319286720-etc-iscsi\") pod \"5e438749-e7d5-4ca0-98e1-6d3319286720\" (UID: \"5e438749-e7d5-4ca0-98e1-6d3319286720\") " Dec 01 09:01:12 crc kubenswrapper[4813]: I1201 09:01:12.809367 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5e438749-e7d5-4ca0-98e1-6d3319286720-config-data\") pod \"5e438749-e7d5-4ca0-98e1-6d3319286720\" (UID: \"5e438749-e7d5-4ca0-98e1-6d3319286720\") " Dec 01 09:01:12 crc kubenswrapper[4813]: I1201 09:01:12.809392 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"5e438749-e7d5-4ca0-98e1-6d3319286720\" (UID: \"5e438749-e7d5-4ca0-98e1-6d3319286720\") " Dec 01 09:01:12 crc kubenswrapper[4813]: I1201 09:01:12.809413 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/5e438749-e7d5-4ca0-98e1-6d3319286720-lib-modules\") pod \"5e438749-e7d5-4ca0-98e1-6d3319286720\" (UID: \"5e438749-e7d5-4ca0-98e1-6d3319286720\") " Dec 01 09:01:12 crc kubenswrapper[4813]: I1201 09:01:12.809439 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5e438749-e7d5-4ca0-98e1-6d3319286720-httpd-run\") pod \"5e438749-e7d5-4ca0-98e1-6d3319286720\" (UID: \"5e438749-e7d5-4ca0-98e1-6d3319286720\") " Dec 01 09:01:12 crc kubenswrapper[4813]: I1201 09:01:12.809462 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/5e438749-e7d5-4ca0-98e1-6d3319286720-var-locks-brick\") pod \"5e438749-e7d5-4ca0-98e1-6d3319286720\" (UID: \"5e438749-e7d5-4ca0-98e1-6d3319286720\") " Dec 01 09:01:12 crc kubenswrapper[4813]: I1201 09:01:12.809479 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/5e438749-e7d5-4ca0-98e1-6d3319286720-run\") pod \"5e438749-e7d5-4ca0-98e1-6d3319286720\" (UID: \"5e438749-e7d5-4ca0-98e1-6d3319286720\") " Dec 01 09:01:12 crc kubenswrapper[4813]: I1201 09:01:12.809543 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5e438749-e7d5-4ca0-98e1-6d3319286720-logs\") pod \"5e438749-e7d5-4ca0-98e1-6d3319286720\" (UID: \"5e438749-e7d5-4ca0-98e1-6d3319286720\") " Dec 01 09:01:12 crc kubenswrapper[4813]: I1201 09:01:12.809604 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/5e438749-e7d5-4ca0-98e1-6d3319286720-dev\") pod \"5e438749-e7d5-4ca0-98e1-6d3319286720\" (UID: \"5e438749-e7d5-4ca0-98e1-6d3319286720\") " Dec 01 09:01:12 crc kubenswrapper[4813]: I1201 09:01:12.809638 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/5e438749-e7d5-4ca0-98e1-6d3319286720-etc-nvme\") pod \"5e438749-e7d5-4ca0-98e1-6d3319286720\" (UID: \"5e438749-e7d5-4ca0-98e1-6d3319286720\") " Dec 01 09:01:12 crc kubenswrapper[4813]: I1201 09:01:12.809677 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/5e438749-e7d5-4ca0-98e1-6d3319286720-sys\") pod \"5e438749-e7d5-4ca0-98e1-6d3319286720\" (UID: \"5e438749-e7d5-4ca0-98e1-6d3319286720\") " Dec 01 09:01:12 crc kubenswrapper[4813]: I1201 09:01:12.810040 4813 reconciler_common.go:293] "Volume detached for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" DevicePath \"\"" Dec 01 09:01:12 crc kubenswrapper[4813]: I1201 09:01:12.810064 4813 reconciler_common.go:293] "Volume detached for volume \"local-storage13-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage13-crc\") on node \"crc\" DevicePath \"\"" Dec 01 09:01:12 crc kubenswrapper[4813]: I1201 09:01:12.810119 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5e438749-e7d5-4ca0-98e1-6d3319286720-sys" (OuterVolumeSpecName: "sys") pod "5e438749-e7d5-4ca0-98e1-6d3319286720" (UID: "5e438749-e7d5-4ca0-98e1-6d3319286720"). InnerVolumeSpecName "sys". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 09:01:12 crc kubenswrapper[4813]: I1201 09:01:12.810435 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5e438749-e7d5-4ca0-98e1-6d3319286720-run" (OuterVolumeSpecName: "run") pod "5e438749-e7d5-4ca0-98e1-6d3319286720" (UID: "5e438749-e7d5-4ca0-98e1-6d3319286720"). InnerVolumeSpecName "run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 09:01:12 crc kubenswrapper[4813]: I1201 09:01:12.810470 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5e438749-e7d5-4ca0-98e1-6d3319286720-var-locks-brick" (OuterVolumeSpecName: "var-locks-brick") pod "5e438749-e7d5-4ca0-98e1-6d3319286720" (UID: "5e438749-e7d5-4ca0-98e1-6d3319286720"). InnerVolumeSpecName "var-locks-brick". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 09:01:12 crc kubenswrapper[4813]: I1201 09:01:12.810573 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5e438749-e7d5-4ca0-98e1-6d3319286720-dev" (OuterVolumeSpecName: "dev") pod "5e438749-e7d5-4ca0-98e1-6d3319286720" (UID: "5e438749-e7d5-4ca0-98e1-6d3319286720"). InnerVolumeSpecName "dev". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 09:01:12 crc kubenswrapper[4813]: I1201 09:01:12.810603 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5e438749-e7d5-4ca0-98e1-6d3319286720-etc-nvme" (OuterVolumeSpecName: "etc-nvme") pod "5e438749-e7d5-4ca0-98e1-6d3319286720" (UID: "5e438749-e7d5-4ca0-98e1-6d3319286720"). InnerVolumeSpecName "etc-nvme". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 09:01:12 crc kubenswrapper[4813]: I1201 09:01:12.810625 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5e438749-e7d5-4ca0-98e1-6d3319286720-etc-iscsi" (OuterVolumeSpecName: "etc-iscsi") pod "5e438749-e7d5-4ca0-98e1-6d3319286720" (UID: "5e438749-e7d5-4ca0-98e1-6d3319286720"). InnerVolumeSpecName "etc-iscsi". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 09:01:12 crc kubenswrapper[4813]: I1201 09:01:12.810627 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5e438749-e7d5-4ca0-98e1-6d3319286720-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "5e438749-e7d5-4ca0-98e1-6d3319286720" (UID: "5e438749-e7d5-4ca0-98e1-6d3319286720"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 09:01:12 crc kubenswrapper[4813]: I1201 09:01:12.810753 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5e438749-e7d5-4ca0-98e1-6d3319286720-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "5e438749-e7d5-4ca0-98e1-6d3319286720" (UID: "5e438749-e7d5-4ca0-98e1-6d3319286720"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 09:01:12 crc kubenswrapper[4813]: I1201 09:01:12.810934 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5e438749-e7d5-4ca0-98e1-6d3319286720-logs" (OuterVolumeSpecName: "logs") pod "5e438749-e7d5-4ca0-98e1-6d3319286720" (UID: "5e438749-e7d5-4ca0-98e1-6d3319286720"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 09:01:12 crc kubenswrapper[4813]: I1201 09:01:12.813432 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage01-crc" (OuterVolumeSpecName: "glance") pod "5e438749-e7d5-4ca0-98e1-6d3319286720" (UID: "5e438749-e7d5-4ca0-98e1-6d3319286720"). InnerVolumeSpecName "local-storage01-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 01 09:01:12 crc kubenswrapper[4813]: I1201 09:01:12.813509 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5e438749-e7d5-4ca0-98e1-6d3319286720-scripts" (OuterVolumeSpecName: "scripts") pod "5e438749-e7d5-4ca0-98e1-6d3319286720" (UID: "5e438749-e7d5-4ca0-98e1-6d3319286720"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:01:12 crc kubenswrapper[4813]: I1201 09:01:12.813547 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage17-crc" (OuterVolumeSpecName: "glance-cache") pod "5e438749-e7d5-4ca0-98e1-6d3319286720" (UID: "5e438749-e7d5-4ca0-98e1-6d3319286720"). InnerVolumeSpecName "local-storage17-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 01 09:01:12 crc kubenswrapper[4813]: I1201 09:01:12.814051 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5e438749-e7d5-4ca0-98e1-6d3319286720-kube-api-access-s2gqr" (OuterVolumeSpecName: "kube-api-access-s2gqr") pod "5e438749-e7d5-4ca0-98e1-6d3319286720" (UID: "5e438749-e7d5-4ca0-98e1-6d3319286720"). InnerVolumeSpecName "kube-api-access-s2gqr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:01:12 crc kubenswrapper[4813]: I1201 09:01:12.869220 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5e438749-e7d5-4ca0-98e1-6d3319286720-config-data" (OuterVolumeSpecName: "config-data") pod "5e438749-e7d5-4ca0-98e1-6d3319286720" (UID: "5e438749-e7d5-4ca0-98e1-6d3319286720"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:01:12 crc kubenswrapper[4813]: I1201 09:01:12.911330 4813 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5e438749-e7d5-4ca0-98e1-6d3319286720-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 01 09:01:12 crc kubenswrapper[4813]: I1201 09:01:12.911369 4813 reconciler_common.go:293] "Volume detached for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/5e438749-e7d5-4ca0-98e1-6d3319286720-var-locks-brick\") on node \"crc\" DevicePath \"\"" Dec 01 09:01:12 crc kubenswrapper[4813]: I1201 09:01:12.911381 4813 reconciler_common.go:293] "Volume detached for volume \"run\" (UniqueName: \"kubernetes.io/host-path/5e438749-e7d5-4ca0-98e1-6d3319286720-run\") on node \"crc\" DevicePath \"\"" Dec 01 09:01:12 crc kubenswrapper[4813]: I1201 09:01:12.911389 4813 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5e438749-e7d5-4ca0-98e1-6d3319286720-logs\") on node \"crc\" DevicePath \"\"" Dec 01 09:01:12 crc kubenswrapper[4813]: I1201 09:01:12.911396 4813 reconciler_common.go:293] "Volume detached for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/5e438749-e7d5-4ca0-98e1-6d3319286720-dev\") on node \"crc\" DevicePath \"\"" Dec 01 09:01:12 crc kubenswrapper[4813]: I1201 09:01:12.911404 4813 reconciler_common.go:293] "Volume detached for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/5e438749-e7d5-4ca0-98e1-6d3319286720-etc-nvme\") on node \"crc\" DevicePath \"\"" Dec 01 09:01:12 crc kubenswrapper[4813]: I1201 09:01:12.911411 4813 reconciler_common.go:293] "Volume detached for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/5e438749-e7d5-4ca0-98e1-6d3319286720-sys\") on node \"crc\" DevicePath \"\"" Dec 01 09:01:12 crc kubenswrapper[4813]: I1201 09:01:12.911445 4813 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage17-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage17-crc\") on node \"crc\" " Dec 01 09:01:12 crc kubenswrapper[4813]: I1201 09:01:12.911453 4813 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5e438749-e7d5-4ca0-98e1-6d3319286720-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 09:01:12 crc kubenswrapper[4813]: I1201 09:01:12.911463 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s2gqr\" (UniqueName: \"kubernetes.io/projected/5e438749-e7d5-4ca0-98e1-6d3319286720-kube-api-access-s2gqr\") on node \"crc\" DevicePath \"\"" Dec 01 09:01:12 crc kubenswrapper[4813]: I1201 09:01:12.911475 4813 reconciler_common.go:293] "Volume detached for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/5e438749-e7d5-4ca0-98e1-6d3319286720-etc-iscsi\") on node \"crc\" DevicePath \"\"" Dec 01 09:01:12 crc kubenswrapper[4813]: I1201 09:01:12.911485 4813 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5e438749-e7d5-4ca0-98e1-6d3319286720-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 09:01:12 crc kubenswrapper[4813]: I1201 09:01:12.911501 4813 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" " Dec 01 09:01:12 crc kubenswrapper[4813]: I1201 09:01:12.920161 4813 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/5e438749-e7d5-4ca0-98e1-6d3319286720-lib-modules\") on node \"crc\" DevicePath \"\"" Dec 01 09:01:12 crc kubenswrapper[4813]: I1201 09:01:12.924299 4813 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage01-crc" (UniqueName: "kubernetes.io/local-volume/local-storage01-crc") on node "crc" Dec 01 09:01:12 crc kubenswrapper[4813]: I1201 09:01:12.924569 4813 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage17-crc" (UniqueName: "kubernetes.io/local-volume/local-storage17-crc") on node "crc" Dec 01 09:01:13 crc kubenswrapper[4813]: I1201 09:01:13.021947 4813 reconciler_common.go:293] "Volume detached for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" DevicePath \"\"" Dec 01 09:01:13 crc kubenswrapper[4813]: I1201 09:01:13.021997 4813 reconciler_common.go:293] "Volume detached for volume \"local-storage17-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage17-crc\") on node \"crc\" DevicePath \"\"" Dec 01 09:01:13 crc kubenswrapper[4813]: I1201 09:01:13.160581 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-0" event={"ID":"e738e940-82c7-481a-8a7d-069215ae99c7","Type":"ContainerDied","Data":"299e57f1bdbe70875c7a62d154f86a2a573d6cf0d878700c05def95d4a925c71"} Dec 01 09:01:13 crc kubenswrapper[4813]: I1201 09:01:13.160653 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-0" Dec 01 09:01:13 crc kubenswrapper[4813]: I1201 09:01:13.160943 4813 scope.go:117] "RemoveContainer" containerID="c6d074d29e895601fa04da4eca3bc4fca8102be4e3f3ac5b7fa6de4021e4f833" Dec 01 09:01:13 crc kubenswrapper[4813]: I1201 09:01:13.166216 4813 generic.go:334] "Generic (PLEG): container finished" podID="5e438749-e7d5-4ca0-98e1-6d3319286720" containerID="6604eac2164f12eccf3a2895e8fbaf218ce287738434acbba0a5cadfe7fcc82f" exitCode=0 Dec 01 09:01:13 crc kubenswrapper[4813]: I1201 09:01:13.166354 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:01:13 crc kubenswrapper[4813]: I1201 09:01:13.166356 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"5e438749-e7d5-4ca0-98e1-6d3319286720","Type":"ContainerDied","Data":"6604eac2164f12eccf3a2895e8fbaf218ce287738434acbba0a5cadfe7fcc82f"} Dec 01 09:01:13 crc kubenswrapper[4813]: I1201 09:01:13.166642 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"5e438749-e7d5-4ca0-98e1-6d3319286720","Type":"ContainerDied","Data":"2307171a9b2640da2b9a255b325670d959a506be54a9a7da454bee9bc5d3e61d"} Dec 01 09:01:13 crc kubenswrapper[4813]: I1201 09:01:13.204429 4813 scope.go:117] "RemoveContainer" containerID="e3c4d4cd844c6982eeee8cf67db9ce5ebdfe25059f786d79a084320b82fe71d6" Dec 01 09:01:13 crc kubenswrapper[4813]: I1201 09:01:13.205716 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-0"] Dec 01 09:01:13 crc kubenswrapper[4813]: I1201 09:01:13.217500 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-0"] Dec 01 09:01:13 crc kubenswrapper[4813]: I1201 09:01:13.231406 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Dec 01 09:01:13 crc kubenswrapper[4813]: I1201 09:01:13.239092 4813 scope.go:117] "RemoveContainer" containerID="c7f8843d64fb50ae7e49ac4a05f42b1eac5e18fac3aeb36b3a1e4e47e9457f20" Dec 01 09:01:13 crc kubenswrapper[4813]: I1201 09:01:13.241810 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Dec 01 09:01:13 crc kubenswrapper[4813]: I1201 09:01:13.259715 4813 scope.go:117] "RemoveContainer" containerID="6604eac2164f12eccf3a2895e8fbaf218ce287738434acbba0a5cadfe7fcc82f" Dec 01 09:01:13 crc kubenswrapper[4813]: I1201 09:01:13.280611 4813 scope.go:117] "RemoveContainer" containerID="0fba4493026f7e0bb5c77eac41f7dd9e9be04fb93081af5767525ccad569ecc3" Dec 01 09:01:13 crc kubenswrapper[4813]: I1201 09:01:13.306314 4813 scope.go:117] "RemoveContainer" containerID="dcb779b0095ccf2057b4fa7f52ea855e49228c7d0c78ae4beaafc5c8e1f860ba" Dec 01 09:01:13 crc kubenswrapper[4813]: I1201 09:01:13.331777 4813 scope.go:117] "RemoveContainer" containerID="6604eac2164f12eccf3a2895e8fbaf218ce287738434acbba0a5cadfe7fcc82f" Dec 01 09:01:13 crc kubenswrapper[4813]: E1201 09:01:13.332359 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6604eac2164f12eccf3a2895e8fbaf218ce287738434acbba0a5cadfe7fcc82f\": container with ID starting with 6604eac2164f12eccf3a2895e8fbaf218ce287738434acbba0a5cadfe7fcc82f not found: ID does not exist" containerID="6604eac2164f12eccf3a2895e8fbaf218ce287738434acbba0a5cadfe7fcc82f" Dec 01 09:01:13 crc kubenswrapper[4813]: I1201 09:01:13.332419 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6604eac2164f12eccf3a2895e8fbaf218ce287738434acbba0a5cadfe7fcc82f"} err="failed to get container status \"6604eac2164f12eccf3a2895e8fbaf218ce287738434acbba0a5cadfe7fcc82f\": rpc error: code = NotFound desc = could not find container \"6604eac2164f12eccf3a2895e8fbaf218ce287738434acbba0a5cadfe7fcc82f\": container with ID starting with 6604eac2164f12eccf3a2895e8fbaf218ce287738434acbba0a5cadfe7fcc82f not found: ID does not exist" Dec 01 09:01:13 crc kubenswrapper[4813]: I1201 09:01:13.332458 4813 scope.go:117] "RemoveContainer" containerID="0fba4493026f7e0bb5c77eac41f7dd9e9be04fb93081af5767525ccad569ecc3" Dec 01 09:01:13 crc kubenswrapper[4813]: E1201 09:01:13.332880 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0fba4493026f7e0bb5c77eac41f7dd9e9be04fb93081af5767525ccad569ecc3\": container with ID starting with 0fba4493026f7e0bb5c77eac41f7dd9e9be04fb93081af5767525ccad569ecc3 not found: ID does not exist" containerID="0fba4493026f7e0bb5c77eac41f7dd9e9be04fb93081af5767525ccad569ecc3" Dec 01 09:01:13 crc kubenswrapper[4813]: I1201 09:01:13.332955 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0fba4493026f7e0bb5c77eac41f7dd9e9be04fb93081af5767525ccad569ecc3"} err="failed to get container status \"0fba4493026f7e0bb5c77eac41f7dd9e9be04fb93081af5767525ccad569ecc3\": rpc error: code = NotFound desc = could not find container \"0fba4493026f7e0bb5c77eac41f7dd9e9be04fb93081af5767525ccad569ecc3\": container with ID starting with 0fba4493026f7e0bb5c77eac41f7dd9e9be04fb93081af5767525ccad569ecc3 not found: ID does not exist" Dec 01 09:01:13 crc kubenswrapper[4813]: I1201 09:01:13.333038 4813 scope.go:117] "RemoveContainer" containerID="dcb779b0095ccf2057b4fa7f52ea855e49228c7d0c78ae4beaafc5c8e1f860ba" Dec 01 09:01:13 crc kubenswrapper[4813]: E1201 09:01:13.333688 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dcb779b0095ccf2057b4fa7f52ea855e49228c7d0c78ae4beaafc5c8e1f860ba\": container with ID starting with dcb779b0095ccf2057b4fa7f52ea855e49228c7d0c78ae4beaafc5c8e1f860ba not found: ID does not exist" containerID="dcb779b0095ccf2057b4fa7f52ea855e49228c7d0c78ae4beaafc5c8e1f860ba" Dec 01 09:01:13 crc kubenswrapper[4813]: I1201 09:01:13.333725 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dcb779b0095ccf2057b4fa7f52ea855e49228c7d0c78ae4beaafc5c8e1f860ba"} err="failed to get container status \"dcb779b0095ccf2057b4fa7f52ea855e49228c7d0c78ae4beaafc5c8e1f860ba\": rpc error: code = NotFound desc = could not find container \"dcb779b0095ccf2057b4fa7f52ea855e49228c7d0c78ae4beaafc5c8e1f860ba\": container with ID starting with dcb779b0095ccf2057b4fa7f52ea855e49228c7d0c78ae4beaafc5c8e1f860ba not found: ID does not exist" Dec 01 09:01:13 crc kubenswrapper[4813]: I1201 09:01:13.878866 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-db-sync-75cp2"] Dec 01 09:01:13 crc kubenswrapper[4813]: I1201 09:01:13.885767 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-db-sync-75cp2"] Dec 01 09:01:13 crc kubenswrapper[4813]: I1201 09:01:13.919150 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance532e-account-delete-95fc9"] Dec 01 09:01:13 crc kubenswrapper[4813]: E1201 09:01:13.919555 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="423d0710-1fe1-4cd3-99a8-3c32c7f08ee3" containerName="glance-api" Dec 01 09:01:13 crc kubenswrapper[4813]: I1201 09:01:13.919594 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="423d0710-1fe1-4cd3-99a8-3c32c7f08ee3" containerName="glance-api" Dec 01 09:01:13 crc kubenswrapper[4813]: E1201 09:01:13.919632 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0aa6a534-c6a3-4a53-9e0e-053abec1ad18" containerName="glance-httpd" Dec 01 09:01:13 crc kubenswrapper[4813]: I1201 09:01:13.919642 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="0aa6a534-c6a3-4a53-9e0e-053abec1ad18" containerName="glance-httpd" Dec 01 09:01:13 crc kubenswrapper[4813]: E1201 09:01:13.919656 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="423d0710-1fe1-4cd3-99a8-3c32c7f08ee3" containerName="glance-httpd" Dec 01 09:01:13 crc kubenswrapper[4813]: I1201 09:01:13.919666 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="423d0710-1fe1-4cd3-99a8-3c32c7f08ee3" containerName="glance-httpd" Dec 01 09:01:13 crc kubenswrapper[4813]: E1201 09:01:13.919676 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="04117ae2-817e-41a3-9d1e-608290bf79c8" containerName="glance-api" Dec 01 09:01:13 crc kubenswrapper[4813]: I1201 09:01:13.919683 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="04117ae2-817e-41a3-9d1e-608290bf79c8" containerName="glance-api" Dec 01 09:01:13 crc kubenswrapper[4813]: E1201 09:01:13.919704 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5e438749-e7d5-4ca0-98e1-6d3319286720" containerName="glance-httpd" Dec 01 09:01:13 crc kubenswrapper[4813]: I1201 09:01:13.919712 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="5e438749-e7d5-4ca0-98e1-6d3319286720" containerName="glance-httpd" Dec 01 09:01:13 crc kubenswrapper[4813]: E1201 09:01:13.919727 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e738e940-82c7-481a-8a7d-069215ae99c7" containerName="glance-log" Dec 01 09:01:13 crc kubenswrapper[4813]: I1201 09:01:13.919735 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="e738e940-82c7-481a-8a7d-069215ae99c7" containerName="glance-log" Dec 01 09:01:13 crc kubenswrapper[4813]: E1201 09:01:13.919745 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5e438749-e7d5-4ca0-98e1-6d3319286720" containerName="glance-log" Dec 01 09:01:13 crc kubenswrapper[4813]: I1201 09:01:13.919752 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="5e438749-e7d5-4ca0-98e1-6d3319286720" containerName="glance-log" Dec 01 09:01:13 crc kubenswrapper[4813]: E1201 09:01:13.919787 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a1ad44fa-8926-4242-8d07-1be670bb316a" containerName="glance-log" Dec 01 09:01:13 crc kubenswrapper[4813]: I1201 09:01:13.919795 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="a1ad44fa-8926-4242-8d07-1be670bb316a" containerName="glance-log" Dec 01 09:01:13 crc kubenswrapper[4813]: E1201 09:01:13.919811 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e738e940-82c7-481a-8a7d-069215ae99c7" containerName="glance-api" Dec 01 09:01:13 crc kubenswrapper[4813]: I1201 09:01:13.919818 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="e738e940-82c7-481a-8a7d-069215ae99c7" containerName="glance-api" Dec 01 09:01:13 crc kubenswrapper[4813]: E1201 09:01:13.919831 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="04117ae2-817e-41a3-9d1e-608290bf79c8" containerName="glance-log" Dec 01 09:01:13 crc kubenswrapper[4813]: I1201 09:01:13.919839 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="04117ae2-817e-41a3-9d1e-608290bf79c8" containerName="glance-log" Dec 01 09:01:13 crc kubenswrapper[4813]: E1201 09:01:13.919855 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5e438749-e7d5-4ca0-98e1-6d3319286720" containerName="glance-api" Dec 01 09:01:13 crc kubenswrapper[4813]: I1201 09:01:13.919862 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="5e438749-e7d5-4ca0-98e1-6d3319286720" containerName="glance-api" Dec 01 09:01:13 crc kubenswrapper[4813]: E1201 09:01:13.919871 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0aa6a534-c6a3-4a53-9e0e-053abec1ad18" containerName="glance-log" Dec 01 09:01:13 crc kubenswrapper[4813]: I1201 09:01:13.919878 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="0aa6a534-c6a3-4a53-9e0e-053abec1ad18" containerName="glance-log" Dec 01 09:01:13 crc kubenswrapper[4813]: E1201 09:01:13.919887 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0aa6a534-c6a3-4a53-9e0e-053abec1ad18" containerName="glance-api" Dec 01 09:01:13 crc kubenswrapper[4813]: I1201 09:01:13.919895 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="0aa6a534-c6a3-4a53-9e0e-053abec1ad18" containerName="glance-api" Dec 01 09:01:13 crc kubenswrapper[4813]: E1201 09:01:13.919910 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a1ad44fa-8926-4242-8d07-1be670bb316a" containerName="glance-api" Dec 01 09:01:13 crc kubenswrapper[4813]: I1201 09:01:13.919918 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="a1ad44fa-8926-4242-8d07-1be670bb316a" containerName="glance-api" Dec 01 09:01:13 crc kubenswrapper[4813]: E1201 09:01:13.919930 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="04117ae2-817e-41a3-9d1e-608290bf79c8" containerName="glance-httpd" Dec 01 09:01:13 crc kubenswrapper[4813]: I1201 09:01:13.919940 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="04117ae2-817e-41a3-9d1e-608290bf79c8" containerName="glance-httpd" Dec 01 09:01:13 crc kubenswrapper[4813]: E1201 09:01:13.919952 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b247a555-0805-4884-a0d5-d2799423bf94" containerName="keystone-cron" Dec 01 09:01:13 crc kubenswrapper[4813]: I1201 09:01:13.919961 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="b247a555-0805-4884-a0d5-d2799423bf94" containerName="keystone-cron" Dec 01 09:01:13 crc kubenswrapper[4813]: E1201 09:01:13.919993 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e738e940-82c7-481a-8a7d-069215ae99c7" containerName="glance-httpd" Dec 01 09:01:13 crc kubenswrapper[4813]: I1201 09:01:13.920001 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="e738e940-82c7-481a-8a7d-069215ae99c7" containerName="glance-httpd" Dec 01 09:01:13 crc kubenswrapper[4813]: E1201 09:01:13.920011 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="423d0710-1fe1-4cd3-99a8-3c32c7f08ee3" containerName="glance-log" Dec 01 09:01:13 crc kubenswrapper[4813]: I1201 09:01:13.920017 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="423d0710-1fe1-4cd3-99a8-3c32c7f08ee3" containerName="glance-log" Dec 01 09:01:13 crc kubenswrapper[4813]: E1201 09:01:13.920027 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a1ad44fa-8926-4242-8d07-1be670bb316a" containerName="glance-httpd" Dec 01 09:01:13 crc kubenswrapper[4813]: I1201 09:01:13.920032 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="a1ad44fa-8926-4242-8d07-1be670bb316a" containerName="glance-httpd" Dec 01 09:01:13 crc kubenswrapper[4813]: I1201 09:01:13.920210 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="e738e940-82c7-481a-8a7d-069215ae99c7" containerName="glance-log" Dec 01 09:01:13 crc kubenswrapper[4813]: I1201 09:01:13.920233 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="e738e940-82c7-481a-8a7d-069215ae99c7" containerName="glance-httpd" Dec 01 09:01:13 crc kubenswrapper[4813]: I1201 09:01:13.920245 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="a1ad44fa-8926-4242-8d07-1be670bb316a" containerName="glance-httpd" Dec 01 09:01:13 crc kubenswrapper[4813]: I1201 09:01:13.920259 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="a1ad44fa-8926-4242-8d07-1be670bb316a" containerName="glance-log" Dec 01 09:01:13 crc kubenswrapper[4813]: I1201 09:01:13.920273 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="04117ae2-817e-41a3-9d1e-608290bf79c8" containerName="glance-httpd" Dec 01 09:01:13 crc kubenswrapper[4813]: I1201 09:01:13.920284 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="a1ad44fa-8926-4242-8d07-1be670bb316a" containerName="glance-api" Dec 01 09:01:13 crc kubenswrapper[4813]: I1201 09:01:13.920296 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="5e438749-e7d5-4ca0-98e1-6d3319286720" containerName="glance-api" Dec 01 09:01:13 crc kubenswrapper[4813]: I1201 09:01:13.920305 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="423d0710-1fe1-4cd3-99a8-3c32c7f08ee3" containerName="glance-log" Dec 01 09:01:13 crc kubenswrapper[4813]: I1201 09:01:13.920315 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="5e438749-e7d5-4ca0-98e1-6d3319286720" containerName="glance-log" Dec 01 09:01:13 crc kubenswrapper[4813]: I1201 09:01:13.920325 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="5e438749-e7d5-4ca0-98e1-6d3319286720" containerName="glance-httpd" Dec 01 09:01:13 crc kubenswrapper[4813]: I1201 09:01:13.920335 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="04117ae2-817e-41a3-9d1e-608290bf79c8" containerName="glance-api" Dec 01 09:01:13 crc kubenswrapper[4813]: I1201 09:01:13.920348 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="423d0710-1fe1-4cd3-99a8-3c32c7f08ee3" containerName="glance-api" Dec 01 09:01:13 crc kubenswrapper[4813]: I1201 09:01:13.920356 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="423d0710-1fe1-4cd3-99a8-3c32c7f08ee3" containerName="glance-httpd" Dec 01 09:01:13 crc kubenswrapper[4813]: I1201 09:01:13.920365 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="e738e940-82c7-481a-8a7d-069215ae99c7" containerName="glance-api" Dec 01 09:01:13 crc kubenswrapper[4813]: I1201 09:01:13.920374 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="04117ae2-817e-41a3-9d1e-608290bf79c8" containerName="glance-log" Dec 01 09:01:13 crc kubenswrapper[4813]: I1201 09:01:13.920384 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="0aa6a534-c6a3-4a53-9e0e-053abec1ad18" containerName="glance-log" Dec 01 09:01:13 crc kubenswrapper[4813]: I1201 09:01:13.920394 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="0aa6a534-c6a3-4a53-9e0e-053abec1ad18" containerName="glance-api" Dec 01 09:01:13 crc kubenswrapper[4813]: I1201 09:01:13.920402 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="b247a555-0805-4884-a0d5-d2799423bf94" containerName="keystone-cron" Dec 01 09:01:13 crc kubenswrapper[4813]: I1201 09:01:13.920413 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="0aa6a534-c6a3-4a53-9e0e-053abec1ad18" containerName="glance-httpd" Dec 01 09:01:13 crc kubenswrapper[4813]: I1201 09:01:13.921027 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance532e-account-delete-95fc9" Dec 01 09:01:13 crc kubenswrapper[4813]: I1201 09:01:13.927057 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance532e-account-delete-95fc9"] Dec 01 09:01:14 crc kubenswrapper[4813]: I1201 09:01:14.094734 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/450e72b9-61aa-4251-b01b-e9c79833d45b-operator-scripts\") pod \"glance532e-account-delete-95fc9\" (UID: \"450e72b9-61aa-4251-b01b-e9c79833d45b\") " pod="glance-kuttl-tests/glance532e-account-delete-95fc9" Dec 01 09:01:14 crc kubenswrapper[4813]: I1201 09:01:14.094806 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d8gp2\" (UniqueName: \"kubernetes.io/projected/450e72b9-61aa-4251-b01b-e9c79833d45b-kube-api-access-d8gp2\") pod \"glance532e-account-delete-95fc9\" (UID: \"450e72b9-61aa-4251-b01b-e9c79833d45b\") " pod="glance-kuttl-tests/glance532e-account-delete-95fc9" Dec 01 09:01:14 crc kubenswrapper[4813]: I1201 09:01:14.196577 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/450e72b9-61aa-4251-b01b-e9c79833d45b-operator-scripts\") pod \"glance532e-account-delete-95fc9\" (UID: \"450e72b9-61aa-4251-b01b-e9c79833d45b\") " pod="glance-kuttl-tests/glance532e-account-delete-95fc9" Dec 01 09:01:14 crc kubenswrapper[4813]: I1201 09:01:14.196653 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d8gp2\" (UniqueName: \"kubernetes.io/projected/450e72b9-61aa-4251-b01b-e9c79833d45b-kube-api-access-d8gp2\") pod \"glance532e-account-delete-95fc9\" (UID: \"450e72b9-61aa-4251-b01b-e9c79833d45b\") " pod="glance-kuttl-tests/glance532e-account-delete-95fc9" Dec 01 09:01:14 crc kubenswrapper[4813]: I1201 09:01:14.197488 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/450e72b9-61aa-4251-b01b-e9c79833d45b-operator-scripts\") pod \"glance532e-account-delete-95fc9\" (UID: \"450e72b9-61aa-4251-b01b-e9c79833d45b\") " pod="glance-kuttl-tests/glance532e-account-delete-95fc9" Dec 01 09:01:14 crc kubenswrapper[4813]: I1201 09:01:14.224808 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d8gp2\" (UniqueName: \"kubernetes.io/projected/450e72b9-61aa-4251-b01b-e9c79833d45b-kube-api-access-d8gp2\") pod \"glance532e-account-delete-95fc9\" (UID: \"450e72b9-61aa-4251-b01b-e9c79833d45b\") " pod="glance-kuttl-tests/glance532e-account-delete-95fc9" Dec 01 09:01:14 crc kubenswrapper[4813]: I1201 09:01:14.237955 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance532e-account-delete-95fc9" Dec 01 09:01:14 crc kubenswrapper[4813]: I1201 09:01:14.413564 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20498728-2d3d-4736-b3f7-a2f32d3a1c0c" path="/var/lib/kubelet/pods/20498728-2d3d-4736-b3f7-a2f32d3a1c0c/volumes" Dec 01 09:01:14 crc kubenswrapper[4813]: I1201 09:01:14.414714 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5e438749-e7d5-4ca0-98e1-6d3319286720" path="/var/lib/kubelet/pods/5e438749-e7d5-4ca0-98e1-6d3319286720/volumes" Dec 01 09:01:14 crc kubenswrapper[4813]: I1201 09:01:14.415467 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e738e940-82c7-481a-8a7d-069215ae99c7" path="/var/lib/kubelet/pods/e738e940-82c7-481a-8a7d-069215ae99c7/volumes" Dec 01 09:01:14 crc kubenswrapper[4813]: I1201 09:01:14.503105 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance532e-account-delete-95fc9"] Dec 01 09:01:15 crc kubenswrapper[4813]: I1201 09:01:15.186298 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance532e-account-delete-95fc9" event={"ID":"450e72b9-61aa-4251-b01b-e9c79833d45b","Type":"ContainerStarted","Data":"209e365f6e86b127b33bad81a1c8713bf80558613377028d1df9285b7a9ee8e4"} Dec 01 09:01:15 crc kubenswrapper[4813]: I1201 09:01:15.186566 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance532e-account-delete-95fc9" event={"ID":"450e72b9-61aa-4251-b01b-e9c79833d45b","Type":"ContainerStarted","Data":"5f1750d417e5f0f0996655f596d53f9e647deb10a4d81e6c3dabfe6a8cd5b72e"} Dec 01 09:01:15 crc kubenswrapper[4813]: I1201 09:01:15.203844 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance532e-account-delete-95fc9" podStartSLOduration=2.20381804 podStartE2EDuration="2.20381804s" podCreationTimestamp="2025-12-01 09:01:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 09:01:15.199246601 +0000 UTC m=+1259.822168187" watchObservedRunningTime="2025-12-01 09:01:15.20381804 +0000 UTC m=+1259.826739626" Dec 01 09:01:16 crc kubenswrapper[4813]: I1201 09:01:16.199624 4813 generic.go:334] "Generic (PLEG): container finished" podID="450e72b9-61aa-4251-b01b-e9c79833d45b" containerID="209e365f6e86b127b33bad81a1c8713bf80558613377028d1df9285b7a9ee8e4" exitCode=0 Dec 01 09:01:16 crc kubenswrapper[4813]: I1201 09:01:16.199675 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance532e-account-delete-95fc9" event={"ID":"450e72b9-61aa-4251-b01b-e9c79833d45b","Type":"ContainerDied","Data":"209e365f6e86b127b33bad81a1c8713bf80558613377028d1df9285b7a9ee8e4"} Dec 01 09:01:17 crc kubenswrapper[4813]: I1201 09:01:17.221456 4813 patch_prober.go:28] interesting pod/machine-config-daemon-9wn4q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 09:01:17 crc kubenswrapper[4813]: I1201 09:01:17.222066 4813 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9wn4q" podUID="3d423402-e08f-42e9-b60c-6eb9bf2f6d48" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 09:01:17 crc kubenswrapper[4813]: I1201 09:01:17.545888 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance532e-account-delete-95fc9" Dec 01 09:01:17 crc kubenswrapper[4813]: I1201 09:01:17.654914 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/450e72b9-61aa-4251-b01b-e9c79833d45b-operator-scripts\") pod \"450e72b9-61aa-4251-b01b-e9c79833d45b\" (UID: \"450e72b9-61aa-4251-b01b-e9c79833d45b\") " Dec 01 09:01:17 crc kubenswrapper[4813]: I1201 09:01:17.655189 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d8gp2\" (UniqueName: \"kubernetes.io/projected/450e72b9-61aa-4251-b01b-e9c79833d45b-kube-api-access-d8gp2\") pod \"450e72b9-61aa-4251-b01b-e9c79833d45b\" (UID: \"450e72b9-61aa-4251-b01b-e9c79833d45b\") " Dec 01 09:01:17 crc kubenswrapper[4813]: I1201 09:01:17.655720 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/450e72b9-61aa-4251-b01b-e9c79833d45b-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "450e72b9-61aa-4251-b01b-e9c79833d45b" (UID: "450e72b9-61aa-4251-b01b-e9c79833d45b"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 09:01:17 crc kubenswrapper[4813]: I1201 09:01:17.673874 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/450e72b9-61aa-4251-b01b-e9c79833d45b-kube-api-access-d8gp2" (OuterVolumeSpecName: "kube-api-access-d8gp2") pod "450e72b9-61aa-4251-b01b-e9c79833d45b" (UID: "450e72b9-61aa-4251-b01b-e9c79833d45b"). InnerVolumeSpecName "kube-api-access-d8gp2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:01:17 crc kubenswrapper[4813]: I1201 09:01:17.756538 4813 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/450e72b9-61aa-4251-b01b-e9c79833d45b-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 09:01:17 crc kubenswrapper[4813]: I1201 09:01:17.756569 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d8gp2\" (UniqueName: \"kubernetes.io/projected/450e72b9-61aa-4251-b01b-e9c79833d45b-kube-api-access-d8gp2\") on node \"crc\" DevicePath \"\"" Dec 01 09:01:18 crc kubenswrapper[4813]: I1201 09:01:18.223174 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance532e-account-delete-95fc9" event={"ID":"450e72b9-61aa-4251-b01b-e9c79833d45b","Type":"ContainerDied","Data":"5f1750d417e5f0f0996655f596d53f9e647deb10a4d81e6c3dabfe6a8cd5b72e"} Dec 01 09:01:18 crc kubenswrapper[4813]: I1201 09:01:18.223224 4813 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5f1750d417e5f0f0996655f596d53f9e647deb10a4d81e6c3dabfe6a8cd5b72e" Dec 01 09:01:18 crc kubenswrapper[4813]: I1201 09:01:18.223323 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance532e-account-delete-95fc9" Dec 01 09:01:18 crc kubenswrapper[4813]: I1201 09:01:18.955544 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-db-create-nrhnq"] Dec 01 09:01:18 crc kubenswrapper[4813]: I1201 09:01:18.975499 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-db-create-nrhnq"] Dec 01 09:01:18 crc kubenswrapper[4813]: I1201 09:01:18.987299 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance532e-account-delete-95fc9"] Dec 01 09:01:18 crc kubenswrapper[4813]: I1201 09:01:18.993999 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-532e-account-create-update-knrg7"] Dec 01 09:01:19 crc kubenswrapper[4813]: I1201 09:01:19.000709 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance532e-account-delete-95fc9"] Dec 01 09:01:19 crc kubenswrapper[4813]: I1201 09:01:19.017028 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-532e-account-create-update-knrg7"] Dec 01 09:01:19 crc kubenswrapper[4813]: I1201 09:01:19.809063 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-db-create-p4lj2"] Dec 01 09:01:19 crc kubenswrapper[4813]: E1201 09:01:19.809551 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="450e72b9-61aa-4251-b01b-e9c79833d45b" containerName="mariadb-account-delete" Dec 01 09:01:19 crc kubenswrapper[4813]: I1201 09:01:19.809566 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="450e72b9-61aa-4251-b01b-e9c79833d45b" containerName="mariadb-account-delete" Dec 01 09:01:19 crc kubenswrapper[4813]: I1201 09:01:19.809692 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="450e72b9-61aa-4251-b01b-e9c79833d45b" containerName="mariadb-account-delete" Dec 01 09:01:19 crc kubenswrapper[4813]: I1201 09:01:19.810303 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-p4lj2" Dec 01 09:01:19 crc kubenswrapper[4813]: I1201 09:01:19.815984 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-create-p4lj2"] Dec 01 09:01:19 crc kubenswrapper[4813]: I1201 09:01:19.837804 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-edab-account-create-update-gcfkg"] Dec 01 09:01:19 crc kubenswrapper[4813]: I1201 09:01:19.838923 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-edab-account-create-update-gcfkg" Dec 01 09:01:19 crc kubenswrapper[4813]: I1201 09:01:19.842751 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-edab-account-create-update-gcfkg"] Dec 01 09:01:19 crc kubenswrapper[4813]: I1201 09:01:19.843338 4813 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-db-secret" Dec 01 09:01:19 crc kubenswrapper[4813]: I1201 09:01:19.906643 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dr55r\" (UniqueName: \"kubernetes.io/projected/5dc86ea3-5e32-4c91-9c14-da28209fd375-kube-api-access-dr55r\") pod \"glance-db-create-p4lj2\" (UID: \"5dc86ea3-5e32-4c91-9c14-da28209fd375\") " pod="glance-kuttl-tests/glance-db-create-p4lj2" Dec 01 09:01:19 crc kubenswrapper[4813]: I1201 09:01:19.906717 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5dc86ea3-5e32-4c91-9c14-da28209fd375-operator-scripts\") pod \"glance-db-create-p4lj2\" (UID: \"5dc86ea3-5e32-4c91-9c14-da28209fd375\") " pod="glance-kuttl-tests/glance-db-create-p4lj2" Dec 01 09:01:20 crc kubenswrapper[4813]: I1201 09:01:20.008375 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dr55r\" (UniqueName: \"kubernetes.io/projected/5dc86ea3-5e32-4c91-9c14-da28209fd375-kube-api-access-dr55r\") pod \"glance-db-create-p4lj2\" (UID: \"5dc86ea3-5e32-4c91-9c14-da28209fd375\") " pod="glance-kuttl-tests/glance-db-create-p4lj2" Dec 01 09:01:20 crc kubenswrapper[4813]: I1201 09:01:20.008437 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5dc86ea3-5e32-4c91-9c14-da28209fd375-operator-scripts\") pod \"glance-db-create-p4lj2\" (UID: \"5dc86ea3-5e32-4c91-9c14-da28209fd375\") " pod="glance-kuttl-tests/glance-db-create-p4lj2" Dec 01 09:01:20 crc kubenswrapper[4813]: I1201 09:01:20.008468 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/85bca524-2f65-4994-8471-8c45ac8cdd22-operator-scripts\") pod \"glance-edab-account-create-update-gcfkg\" (UID: \"85bca524-2f65-4994-8471-8c45ac8cdd22\") " pod="glance-kuttl-tests/glance-edab-account-create-update-gcfkg" Dec 01 09:01:20 crc kubenswrapper[4813]: I1201 09:01:20.008508 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xr86d\" (UniqueName: \"kubernetes.io/projected/85bca524-2f65-4994-8471-8c45ac8cdd22-kube-api-access-xr86d\") pod \"glance-edab-account-create-update-gcfkg\" (UID: \"85bca524-2f65-4994-8471-8c45ac8cdd22\") " pod="glance-kuttl-tests/glance-edab-account-create-update-gcfkg" Dec 01 09:01:20 crc kubenswrapper[4813]: I1201 09:01:20.009238 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5dc86ea3-5e32-4c91-9c14-da28209fd375-operator-scripts\") pod \"glance-db-create-p4lj2\" (UID: \"5dc86ea3-5e32-4c91-9c14-da28209fd375\") " pod="glance-kuttl-tests/glance-db-create-p4lj2" Dec 01 09:01:20 crc kubenswrapper[4813]: I1201 09:01:20.028370 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dr55r\" (UniqueName: \"kubernetes.io/projected/5dc86ea3-5e32-4c91-9c14-da28209fd375-kube-api-access-dr55r\") pod \"glance-db-create-p4lj2\" (UID: \"5dc86ea3-5e32-4c91-9c14-da28209fd375\") " pod="glance-kuttl-tests/glance-db-create-p4lj2" Dec 01 09:01:20 crc kubenswrapper[4813]: I1201 09:01:20.110041 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xr86d\" (UniqueName: \"kubernetes.io/projected/85bca524-2f65-4994-8471-8c45ac8cdd22-kube-api-access-xr86d\") pod \"glance-edab-account-create-update-gcfkg\" (UID: \"85bca524-2f65-4994-8471-8c45ac8cdd22\") " pod="glance-kuttl-tests/glance-edab-account-create-update-gcfkg" Dec 01 09:01:20 crc kubenswrapper[4813]: I1201 09:01:20.110339 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/85bca524-2f65-4994-8471-8c45ac8cdd22-operator-scripts\") pod \"glance-edab-account-create-update-gcfkg\" (UID: \"85bca524-2f65-4994-8471-8c45ac8cdd22\") " pod="glance-kuttl-tests/glance-edab-account-create-update-gcfkg" Dec 01 09:01:20 crc kubenswrapper[4813]: I1201 09:01:20.112679 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/85bca524-2f65-4994-8471-8c45ac8cdd22-operator-scripts\") pod \"glance-edab-account-create-update-gcfkg\" (UID: \"85bca524-2f65-4994-8471-8c45ac8cdd22\") " pod="glance-kuttl-tests/glance-edab-account-create-update-gcfkg" Dec 01 09:01:20 crc kubenswrapper[4813]: I1201 09:01:20.131791 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xr86d\" (UniqueName: \"kubernetes.io/projected/85bca524-2f65-4994-8471-8c45ac8cdd22-kube-api-access-xr86d\") pod \"glance-edab-account-create-update-gcfkg\" (UID: \"85bca524-2f65-4994-8471-8c45ac8cdd22\") " pod="glance-kuttl-tests/glance-edab-account-create-update-gcfkg" Dec 01 09:01:20 crc kubenswrapper[4813]: I1201 09:01:20.164686 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-p4lj2" Dec 01 09:01:20 crc kubenswrapper[4813]: I1201 09:01:20.174650 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-edab-account-create-update-gcfkg" Dec 01 09:01:20 crc kubenswrapper[4813]: I1201 09:01:20.404146 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="450e72b9-61aa-4251-b01b-e9c79833d45b" path="/var/lib/kubelet/pods/450e72b9-61aa-4251-b01b-e9c79833d45b/volumes" Dec 01 09:01:20 crc kubenswrapper[4813]: I1201 09:01:20.405241 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a53d021c-7061-4105-bed8-be1b1ad0d763" path="/var/lib/kubelet/pods/a53d021c-7061-4105-bed8-be1b1ad0d763/volumes" Dec 01 09:01:20 crc kubenswrapper[4813]: I1201 09:01:20.405816 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ab555f70-ee92-4aad-9324-0da0fb3770a1" path="/var/lib/kubelet/pods/ab555f70-ee92-4aad-9324-0da0fb3770a1/volumes" Dec 01 09:01:20 crc kubenswrapper[4813]: I1201 09:01:20.483051 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-edab-account-create-update-gcfkg"] Dec 01 09:01:20 crc kubenswrapper[4813]: I1201 09:01:20.639520 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-create-p4lj2"] Dec 01 09:01:20 crc kubenswrapper[4813]: W1201 09:01:20.641334 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5dc86ea3_5e32_4c91_9c14_da28209fd375.slice/crio-9c91d6c7ffa58d7c09fb3531877bfdc8cedb9aab720e7a382df5a0baece6910f WatchSource:0}: Error finding container 9c91d6c7ffa58d7c09fb3531877bfdc8cedb9aab720e7a382df5a0baece6910f: Status 404 returned error can't find the container with id 9c91d6c7ffa58d7c09fb3531877bfdc8cedb9aab720e7a382df5a0baece6910f Dec 01 09:01:21 crc kubenswrapper[4813]: I1201 09:01:21.254444 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-create-p4lj2" event={"ID":"5dc86ea3-5e32-4c91-9c14-da28209fd375","Type":"ContainerStarted","Data":"23b69e3281f74625e3c7674bf4a4b8ced6ff8e237e045d429a4b21c032a4c42c"} Dec 01 09:01:21 crc kubenswrapper[4813]: I1201 09:01:21.254899 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-create-p4lj2" event={"ID":"5dc86ea3-5e32-4c91-9c14-da28209fd375","Type":"ContainerStarted","Data":"9c91d6c7ffa58d7c09fb3531877bfdc8cedb9aab720e7a382df5a0baece6910f"} Dec 01 09:01:21 crc kubenswrapper[4813]: I1201 09:01:21.257234 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-edab-account-create-update-gcfkg" event={"ID":"85bca524-2f65-4994-8471-8c45ac8cdd22","Type":"ContainerStarted","Data":"cd655f585ece329a36be0668cdd4a160b0949ccdd5582302eda8d26fa6c29547"} Dec 01 09:01:21 crc kubenswrapper[4813]: I1201 09:01:21.257296 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-edab-account-create-update-gcfkg" event={"ID":"85bca524-2f65-4994-8471-8c45ac8cdd22","Type":"ContainerStarted","Data":"04f10f6f33b010a5e57750c8d718176e01ae918162659425df72e4bdfd1230f6"} Dec 01 09:01:21 crc kubenswrapper[4813]: I1201 09:01:21.279796 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-db-create-p4lj2" podStartSLOduration=2.279762512 podStartE2EDuration="2.279762512s" podCreationTimestamp="2025-12-01 09:01:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 09:01:21.26790492 +0000 UTC m=+1265.890826556" watchObservedRunningTime="2025-12-01 09:01:21.279762512 +0000 UTC m=+1265.902684138" Dec 01 09:01:21 crc kubenswrapper[4813]: I1201 09:01:21.310696 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-edab-account-create-update-gcfkg" podStartSLOduration=2.310662089 podStartE2EDuration="2.310662089s" podCreationTimestamp="2025-12-01 09:01:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 09:01:21.288437016 +0000 UTC m=+1265.911358602" watchObservedRunningTime="2025-12-01 09:01:21.310662089 +0000 UTC m=+1265.933583695" Dec 01 09:01:22 crc kubenswrapper[4813]: I1201 09:01:22.284860 4813 generic.go:334] "Generic (PLEG): container finished" podID="5dc86ea3-5e32-4c91-9c14-da28209fd375" containerID="23b69e3281f74625e3c7674bf4a4b8ced6ff8e237e045d429a4b21c032a4c42c" exitCode=0 Dec 01 09:01:22 crc kubenswrapper[4813]: I1201 09:01:22.285435 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-create-p4lj2" event={"ID":"5dc86ea3-5e32-4c91-9c14-da28209fd375","Type":"ContainerDied","Data":"23b69e3281f74625e3c7674bf4a4b8ced6ff8e237e045d429a4b21c032a4c42c"} Dec 01 09:01:22 crc kubenswrapper[4813]: I1201 09:01:22.291743 4813 generic.go:334] "Generic (PLEG): container finished" podID="85bca524-2f65-4994-8471-8c45ac8cdd22" containerID="cd655f585ece329a36be0668cdd4a160b0949ccdd5582302eda8d26fa6c29547" exitCode=0 Dec 01 09:01:22 crc kubenswrapper[4813]: I1201 09:01:22.291850 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-edab-account-create-update-gcfkg" event={"ID":"85bca524-2f65-4994-8471-8c45ac8cdd22","Type":"ContainerDied","Data":"cd655f585ece329a36be0668cdd4a160b0949ccdd5582302eda8d26fa6c29547"} Dec 01 09:01:23 crc kubenswrapper[4813]: I1201 09:01:23.720663 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-p4lj2" Dec 01 09:01:23 crc kubenswrapper[4813]: I1201 09:01:23.726270 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-edab-account-create-update-gcfkg" Dec 01 09:01:23 crc kubenswrapper[4813]: I1201 09:01:23.892175 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/85bca524-2f65-4994-8471-8c45ac8cdd22-operator-scripts\") pod \"85bca524-2f65-4994-8471-8c45ac8cdd22\" (UID: \"85bca524-2f65-4994-8471-8c45ac8cdd22\") " Dec 01 09:01:23 crc kubenswrapper[4813]: I1201 09:01:23.892319 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xr86d\" (UniqueName: \"kubernetes.io/projected/85bca524-2f65-4994-8471-8c45ac8cdd22-kube-api-access-xr86d\") pod \"85bca524-2f65-4994-8471-8c45ac8cdd22\" (UID: \"85bca524-2f65-4994-8471-8c45ac8cdd22\") " Dec 01 09:01:23 crc kubenswrapper[4813]: I1201 09:01:23.892368 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5dc86ea3-5e32-4c91-9c14-da28209fd375-operator-scripts\") pod \"5dc86ea3-5e32-4c91-9c14-da28209fd375\" (UID: \"5dc86ea3-5e32-4c91-9c14-da28209fd375\") " Dec 01 09:01:23 crc kubenswrapper[4813]: I1201 09:01:23.892409 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dr55r\" (UniqueName: \"kubernetes.io/projected/5dc86ea3-5e32-4c91-9c14-da28209fd375-kube-api-access-dr55r\") pod \"5dc86ea3-5e32-4c91-9c14-da28209fd375\" (UID: \"5dc86ea3-5e32-4c91-9c14-da28209fd375\") " Dec 01 09:01:23 crc kubenswrapper[4813]: I1201 09:01:23.893057 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/85bca524-2f65-4994-8471-8c45ac8cdd22-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "85bca524-2f65-4994-8471-8c45ac8cdd22" (UID: "85bca524-2f65-4994-8471-8c45ac8cdd22"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 09:01:23 crc kubenswrapper[4813]: I1201 09:01:23.893804 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5dc86ea3-5e32-4c91-9c14-da28209fd375-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "5dc86ea3-5e32-4c91-9c14-da28209fd375" (UID: "5dc86ea3-5e32-4c91-9c14-da28209fd375"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 09:01:23 crc kubenswrapper[4813]: I1201 09:01:23.900575 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/85bca524-2f65-4994-8471-8c45ac8cdd22-kube-api-access-xr86d" (OuterVolumeSpecName: "kube-api-access-xr86d") pod "85bca524-2f65-4994-8471-8c45ac8cdd22" (UID: "85bca524-2f65-4994-8471-8c45ac8cdd22"). InnerVolumeSpecName "kube-api-access-xr86d". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:01:23 crc kubenswrapper[4813]: I1201 09:01:23.900704 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5dc86ea3-5e32-4c91-9c14-da28209fd375-kube-api-access-dr55r" (OuterVolumeSpecName: "kube-api-access-dr55r") pod "5dc86ea3-5e32-4c91-9c14-da28209fd375" (UID: "5dc86ea3-5e32-4c91-9c14-da28209fd375"). InnerVolumeSpecName "kube-api-access-dr55r". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:01:23 crc kubenswrapper[4813]: I1201 09:01:23.993923 4813 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/85bca524-2f65-4994-8471-8c45ac8cdd22-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 09:01:23 crc kubenswrapper[4813]: I1201 09:01:23.994007 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xr86d\" (UniqueName: \"kubernetes.io/projected/85bca524-2f65-4994-8471-8c45ac8cdd22-kube-api-access-xr86d\") on node \"crc\" DevicePath \"\"" Dec 01 09:01:23 crc kubenswrapper[4813]: I1201 09:01:23.994030 4813 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5dc86ea3-5e32-4c91-9c14-da28209fd375-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 09:01:23 crc kubenswrapper[4813]: I1201 09:01:23.994046 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dr55r\" (UniqueName: \"kubernetes.io/projected/5dc86ea3-5e32-4c91-9c14-da28209fd375-kube-api-access-dr55r\") on node \"crc\" DevicePath \"\"" Dec 01 09:01:24 crc kubenswrapper[4813]: I1201 09:01:24.315070 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-create-p4lj2" event={"ID":"5dc86ea3-5e32-4c91-9c14-da28209fd375","Type":"ContainerDied","Data":"9c91d6c7ffa58d7c09fb3531877bfdc8cedb9aab720e7a382df5a0baece6910f"} Dec 01 09:01:24 crc kubenswrapper[4813]: I1201 09:01:24.315119 4813 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9c91d6c7ffa58d7c09fb3531877bfdc8cedb9aab720e7a382df5a0baece6910f" Dec 01 09:01:24 crc kubenswrapper[4813]: I1201 09:01:24.315121 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-p4lj2" Dec 01 09:01:24 crc kubenswrapper[4813]: I1201 09:01:24.317215 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-edab-account-create-update-gcfkg" event={"ID":"85bca524-2f65-4994-8471-8c45ac8cdd22","Type":"ContainerDied","Data":"04f10f6f33b010a5e57750c8d718176e01ae918162659425df72e4bdfd1230f6"} Dec 01 09:01:24 crc kubenswrapper[4813]: I1201 09:01:24.317249 4813 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="04f10f6f33b010a5e57750c8d718176e01ae918162659425df72e4bdfd1230f6" Dec 01 09:01:24 crc kubenswrapper[4813]: I1201 09:01:24.317295 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-edab-account-create-update-gcfkg" Dec 01 09:01:25 crc kubenswrapper[4813]: I1201 09:01:25.035738 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-db-sync-9lgcq"] Dec 01 09:01:25 crc kubenswrapper[4813]: E1201 09:01:25.036393 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5dc86ea3-5e32-4c91-9c14-da28209fd375" containerName="mariadb-database-create" Dec 01 09:01:25 crc kubenswrapper[4813]: I1201 09:01:25.036410 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="5dc86ea3-5e32-4c91-9c14-da28209fd375" containerName="mariadb-database-create" Dec 01 09:01:25 crc kubenswrapper[4813]: E1201 09:01:25.036427 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="85bca524-2f65-4994-8471-8c45ac8cdd22" containerName="mariadb-account-create-update" Dec 01 09:01:25 crc kubenswrapper[4813]: I1201 09:01:25.036433 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="85bca524-2f65-4994-8471-8c45ac8cdd22" containerName="mariadb-account-create-update" Dec 01 09:01:25 crc kubenswrapper[4813]: I1201 09:01:25.036591 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="5dc86ea3-5e32-4c91-9c14-da28209fd375" containerName="mariadb-database-create" Dec 01 09:01:25 crc kubenswrapper[4813]: I1201 09:01:25.036625 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="85bca524-2f65-4994-8471-8c45ac8cdd22" containerName="mariadb-account-create-update" Dec 01 09:01:25 crc kubenswrapper[4813]: I1201 09:01:25.037129 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-9lgcq" Dec 01 09:01:25 crc kubenswrapper[4813]: I1201 09:01:25.038700 4813 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-config-data" Dec 01 09:01:25 crc kubenswrapper[4813]: I1201 09:01:25.038780 4813 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-glance-dockercfg-dvqjg" Dec 01 09:01:25 crc kubenswrapper[4813]: I1201 09:01:25.044280 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-sync-9lgcq"] Dec 01 09:01:25 crc kubenswrapper[4813]: I1201 09:01:25.112732 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/cd1b1c1f-46a6-4c70-a583-45da831d2cb1-db-sync-config-data\") pod \"glance-db-sync-9lgcq\" (UID: \"cd1b1c1f-46a6-4c70-a583-45da831d2cb1\") " pod="glance-kuttl-tests/glance-db-sync-9lgcq" Dec 01 09:01:25 crc kubenswrapper[4813]: I1201 09:01:25.112801 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-plnq6\" (UniqueName: \"kubernetes.io/projected/cd1b1c1f-46a6-4c70-a583-45da831d2cb1-kube-api-access-plnq6\") pod \"glance-db-sync-9lgcq\" (UID: \"cd1b1c1f-46a6-4c70-a583-45da831d2cb1\") " pod="glance-kuttl-tests/glance-db-sync-9lgcq" Dec 01 09:01:25 crc kubenswrapper[4813]: I1201 09:01:25.112933 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cd1b1c1f-46a6-4c70-a583-45da831d2cb1-config-data\") pod \"glance-db-sync-9lgcq\" (UID: \"cd1b1c1f-46a6-4c70-a583-45da831d2cb1\") " pod="glance-kuttl-tests/glance-db-sync-9lgcq" Dec 01 09:01:25 crc kubenswrapper[4813]: I1201 09:01:25.214739 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cd1b1c1f-46a6-4c70-a583-45da831d2cb1-config-data\") pod \"glance-db-sync-9lgcq\" (UID: \"cd1b1c1f-46a6-4c70-a583-45da831d2cb1\") " pod="glance-kuttl-tests/glance-db-sync-9lgcq" Dec 01 09:01:25 crc kubenswrapper[4813]: I1201 09:01:25.214911 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/cd1b1c1f-46a6-4c70-a583-45da831d2cb1-db-sync-config-data\") pod \"glance-db-sync-9lgcq\" (UID: \"cd1b1c1f-46a6-4c70-a583-45da831d2cb1\") " pod="glance-kuttl-tests/glance-db-sync-9lgcq" Dec 01 09:01:25 crc kubenswrapper[4813]: I1201 09:01:25.214951 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-plnq6\" (UniqueName: \"kubernetes.io/projected/cd1b1c1f-46a6-4c70-a583-45da831d2cb1-kube-api-access-plnq6\") pod \"glance-db-sync-9lgcq\" (UID: \"cd1b1c1f-46a6-4c70-a583-45da831d2cb1\") " pod="glance-kuttl-tests/glance-db-sync-9lgcq" Dec 01 09:01:25 crc kubenswrapper[4813]: I1201 09:01:25.221035 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cd1b1c1f-46a6-4c70-a583-45da831d2cb1-config-data\") pod \"glance-db-sync-9lgcq\" (UID: \"cd1b1c1f-46a6-4c70-a583-45da831d2cb1\") " pod="glance-kuttl-tests/glance-db-sync-9lgcq" Dec 01 09:01:25 crc kubenswrapper[4813]: I1201 09:01:25.221447 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/cd1b1c1f-46a6-4c70-a583-45da831d2cb1-db-sync-config-data\") pod \"glance-db-sync-9lgcq\" (UID: \"cd1b1c1f-46a6-4c70-a583-45da831d2cb1\") " pod="glance-kuttl-tests/glance-db-sync-9lgcq" Dec 01 09:01:25 crc kubenswrapper[4813]: I1201 09:01:25.258915 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-plnq6\" (UniqueName: \"kubernetes.io/projected/cd1b1c1f-46a6-4c70-a583-45da831d2cb1-kube-api-access-plnq6\") pod \"glance-db-sync-9lgcq\" (UID: \"cd1b1c1f-46a6-4c70-a583-45da831d2cb1\") " pod="glance-kuttl-tests/glance-db-sync-9lgcq" Dec 01 09:01:25 crc kubenswrapper[4813]: I1201 09:01:25.356534 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-9lgcq" Dec 01 09:01:25 crc kubenswrapper[4813]: I1201 09:01:25.587298 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-sync-9lgcq"] Dec 01 09:01:26 crc kubenswrapper[4813]: I1201 09:01:26.338053 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-9lgcq" event={"ID":"cd1b1c1f-46a6-4c70-a583-45da831d2cb1","Type":"ContainerStarted","Data":"e52f3b09c24264a2aaf36a533f3da9ad43d8f094026fa71de0903de56e5e8ef6"} Dec 01 09:01:26 crc kubenswrapper[4813]: I1201 09:01:26.338375 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-9lgcq" event={"ID":"cd1b1c1f-46a6-4c70-a583-45da831d2cb1","Type":"ContainerStarted","Data":"e42e1b29c6b837e711204afdd6b3b4f799a8c382cee6f881b9481c4bc0662f13"} Dec 01 09:01:26 crc kubenswrapper[4813]: I1201 09:01:26.357740 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-db-sync-9lgcq" podStartSLOduration=1.357714917 podStartE2EDuration="1.357714917s" podCreationTimestamp="2025-12-01 09:01:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 09:01:26.351106432 +0000 UTC m=+1270.974028038" watchObservedRunningTime="2025-12-01 09:01:26.357714917 +0000 UTC m=+1270.980636523" Dec 01 09:01:29 crc kubenswrapper[4813]: I1201 09:01:29.371403 4813 generic.go:334] "Generic (PLEG): container finished" podID="cd1b1c1f-46a6-4c70-a583-45da831d2cb1" containerID="e52f3b09c24264a2aaf36a533f3da9ad43d8f094026fa71de0903de56e5e8ef6" exitCode=0 Dec 01 09:01:29 crc kubenswrapper[4813]: I1201 09:01:29.371488 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-9lgcq" event={"ID":"cd1b1c1f-46a6-4c70-a583-45da831d2cb1","Type":"ContainerDied","Data":"e52f3b09c24264a2aaf36a533f3da9ad43d8f094026fa71de0903de56e5e8ef6"} Dec 01 09:01:30 crc kubenswrapper[4813]: I1201 09:01:30.761113 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-9lgcq" Dec 01 09:01:30 crc kubenswrapper[4813]: I1201 09:01:30.873078 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-plnq6\" (UniqueName: \"kubernetes.io/projected/cd1b1c1f-46a6-4c70-a583-45da831d2cb1-kube-api-access-plnq6\") pod \"cd1b1c1f-46a6-4c70-a583-45da831d2cb1\" (UID: \"cd1b1c1f-46a6-4c70-a583-45da831d2cb1\") " Dec 01 09:01:30 crc kubenswrapper[4813]: I1201 09:01:30.873245 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/cd1b1c1f-46a6-4c70-a583-45da831d2cb1-db-sync-config-data\") pod \"cd1b1c1f-46a6-4c70-a583-45da831d2cb1\" (UID: \"cd1b1c1f-46a6-4c70-a583-45da831d2cb1\") " Dec 01 09:01:30 crc kubenswrapper[4813]: I1201 09:01:30.873287 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cd1b1c1f-46a6-4c70-a583-45da831d2cb1-config-data\") pod \"cd1b1c1f-46a6-4c70-a583-45da831d2cb1\" (UID: \"cd1b1c1f-46a6-4c70-a583-45da831d2cb1\") " Dec 01 09:01:30 crc kubenswrapper[4813]: I1201 09:01:30.879558 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cd1b1c1f-46a6-4c70-a583-45da831d2cb1-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "cd1b1c1f-46a6-4c70-a583-45da831d2cb1" (UID: "cd1b1c1f-46a6-4c70-a583-45da831d2cb1"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:01:30 crc kubenswrapper[4813]: I1201 09:01:30.879882 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd1b1c1f-46a6-4c70-a583-45da831d2cb1-kube-api-access-plnq6" (OuterVolumeSpecName: "kube-api-access-plnq6") pod "cd1b1c1f-46a6-4c70-a583-45da831d2cb1" (UID: "cd1b1c1f-46a6-4c70-a583-45da831d2cb1"). InnerVolumeSpecName "kube-api-access-plnq6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:01:30 crc kubenswrapper[4813]: I1201 09:01:30.911438 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cd1b1c1f-46a6-4c70-a583-45da831d2cb1-config-data" (OuterVolumeSpecName: "config-data") pod "cd1b1c1f-46a6-4c70-a583-45da831d2cb1" (UID: "cd1b1c1f-46a6-4c70-a583-45da831d2cb1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:01:30 crc kubenswrapper[4813]: I1201 09:01:30.975818 4813 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/cd1b1c1f-46a6-4c70-a583-45da831d2cb1-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 09:01:30 crc kubenswrapper[4813]: I1201 09:01:30.976606 4813 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cd1b1c1f-46a6-4c70-a583-45da831d2cb1-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 09:01:30 crc kubenswrapper[4813]: I1201 09:01:30.976640 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-plnq6\" (UniqueName: \"kubernetes.io/projected/cd1b1c1f-46a6-4c70-a583-45da831d2cb1-kube-api-access-plnq6\") on node \"crc\" DevicePath \"\"" Dec 01 09:01:31 crc kubenswrapper[4813]: I1201 09:01:31.390552 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-9lgcq" event={"ID":"cd1b1c1f-46a6-4c70-a583-45da831d2cb1","Type":"ContainerDied","Data":"e42e1b29c6b837e711204afdd6b3b4f799a8c382cee6f881b9481c4bc0662f13"} Dec 01 09:01:31 crc kubenswrapper[4813]: I1201 09:01:31.390596 4813 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e42e1b29c6b837e711204afdd6b3b4f799a8c382cee6f881b9481c4bc0662f13" Dec 01 09:01:31 crc kubenswrapper[4813]: I1201 09:01:31.390627 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-9lgcq" Dec 01 09:01:32 crc kubenswrapper[4813]: I1201 09:01:32.729986 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-external-api-0"] Dec 01 09:01:32 crc kubenswrapper[4813]: E1201 09:01:32.730568 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cd1b1c1f-46a6-4c70-a583-45da831d2cb1" containerName="glance-db-sync" Dec 01 09:01:32 crc kubenswrapper[4813]: I1201 09:01:32.730582 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="cd1b1c1f-46a6-4c70-a583-45da831d2cb1" containerName="glance-db-sync" Dec 01 09:01:32 crc kubenswrapper[4813]: I1201 09:01:32.730726 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="cd1b1c1f-46a6-4c70-a583-45da831d2cb1" containerName="glance-db-sync" Dec 01 09:01:32 crc kubenswrapper[4813]: I1201 09:01:32.731538 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-0" Dec 01 09:01:32 crc kubenswrapper[4813]: I1201 09:01:32.734553 4813 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-default-external-config-data" Dec 01 09:01:32 crc kubenswrapper[4813]: I1201 09:01:32.735378 4813 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-scripts" Dec 01 09:01:32 crc kubenswrapper[4813]: I1201 09:01:32.737318 4813 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-glance-dockercfg-dvqjg" Dec 01 09:01:32 crc kubenswrapper[4813]: I1201 09:01:32.755820 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-0"] Dec 01 09:01:32 crc kubenswrapper[4813]: I1201 09:01:32.810169 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Dec 01 09:01:32 crc kubenswrapper[4813]: I1201 09:01:32.811436 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:01:32 crc kubenswrapper[4813]: I1201 09:01:32.817363 4813 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-default-internal-config-data" Dec 01 09:01:32 crc kubenswrapper[4813]: I1201 09:01:32.839499 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Dec 01 09:01:32 crc kubenswrapper[4813]: I1201 09:01:32.903373 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7b8bf821-52ae-409b-8e17-fc8e1302ed10-scripts\") pod \"glance-default-external-api-0\" (UID: \"7b8bf821-52ae-409b-8e17-fc8e1302ed10\") " pod="glance-kuttl-tests/glance-default-external-api-0" Dec 01 09:01:32 crc kubenswrapper[4813]: I1201 09:01:32.903449 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/018e9026-a4ad-493c-9c2c-66d9ed3e36a5-lib-modules\") pod \"glance-default-internal-api-0\" (UID: \"018e9026-a4ad-493c-9c2c-66d9ed3e36a5\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:01:32 crc kubenswrapper[4813]: I1201 09:01:32.903471 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/018e9026-a4ad-493c-9c2c-66d9ed3e36a5-etc-nvme\") pod \"glance-default-internal-api-0\" (UID: \"018e9026-a4ad-493c-9c2c-66d9ed3e36a5\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:01:32 crc kubenswrapper[4813]: I1201 09:01:32.903531 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage16-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage16-crc\") pod \"glance-default-external-api-0\" (UID: \"7b8bf821-52ae-409b-8e17-fc8e1302ed10\") " pod="glance-kuttl-tests/glance-default-external-api-0" Dec 01 09:01:32 crc kubenswrapper[4813]: I1201 09:01:32.903556 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/018e9026-a4ad-493c-9c2c-66d9ed3e36a5-run\") pod \"glance-default-internal-api-0\" (UID: \"018e9026-a4ad-493c-9c2c-66d9ed3e36a5\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:01:32 crc kubenswrapper[4813]: I1201 09:01:32.903579 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-internal-api-0\" (UID: \"018e9026-a4ad-493c-9c2c-66d9ed3e36a5\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:01:32 crc kubenswrapper[4813]: I1201 09:01:32.903599 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/018e9026-a4ad-493c-9c2c-66d9ed3e36a5-scripts\") pod \"glance-default-internal-api-0\" (UID: \"018e9026-a4ad-493c-9c2c-66d9ed3e36a5\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:01:32 crc kubenswrapper[4813]: I1201 09:01:32.903650 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/7b8bf821-52ae-409b-8e17-fc8e1302ed10-etc-iscsi\") pod \"glance-default-external-api-0\" (UID: \"7b8bf821-52ae-409b-8e17-fc8e1302ed10\") " pod="glance-kuttl-tests/glance-default-external-api-0" Dec 01 09:01:32 crc kubenswrapper[4813]: I1201 09:01:32.903676 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/018e9026-a4ad-493c-9c2c-66d9ed3e36a5-sys\") pod \"glance-default-internal-api-0\" (UID: \"018e9026-a4ad-493c-9c2c-66d9ed3e36a5\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:01:32 crc kubenswrapper[4813]: I1201 09:01:32.903693 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/018e9026-a4ad-493c-9c2c-66d9ed3e36a5-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"018e9026-a4ad-493c-9c2c-66d9ed3e36a5\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:01:32 crc kubenswrapper[4813]: I1201 09:01:32.903708 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/018e9026-a4ad-493c-9c2c-66d9ed3e36a5-dev\") pod \"glance-default-internal-api-0\" (UID: \"018e9026-a4ad-493c-9c2c-66d9ed3e36a5\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:01:32 crc kubenswrapper[4813]: I1201 09:01:32.903725 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7b8bf821-52ae-409b-8e17-fc8e1302ed10-logs\") pod \"glance-default-external-api-0\" (UID: \"7b8bf821-52ae-409b-8e17-fc8e1302ed10\") " pod="glance-kuttl-tests/glance-default-external-api-0" Dec 01 09:01:32 crc kubenswrapper[4813]: I1201 09:01:32.903752 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7b8bf821-52ae-409b-8e17-fc8e1302ed10-config-data\") pod \"glance-default-external-api-0\" (UID: \"7b8bf821-52ae-409b-8e17-fc8e1302ed10\") " pod="glance-kuttl-tests/glance-default-external-api-0" Dec 01 09:01:32 crc kubenswrapper[4813]: I1201 09:01:32.903771 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/7b8bf821-52ae-409b-8e17-fc8e1302ed10-var-locks-brick\") pod \"glance-default-external-api-0\" (UID: \"7b8bf821-52ae-409b-8e17-fc8e1302ed10\") " pod="glance-kuttl-tests/glance-default-external-api-0" Dec 01 09:01:32 crc kubenswrapper[4813]: I1201 09:01:32.903792 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/018e9026-a4ad-493c-9c2c-66d9ed3e36a5-config-data\") pod \"glance-default-internal-api-0\" (UID: \"018e9026-a4ad-493c-9c2c-66d9ed3e36a5\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:01:32 crc kubenswrapper[4813]: I1201 09:01:32.903807 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/7b8bf821-52ae-409b-8e17-fc8e1302ed10-run\") pod \"glance-default-external-api-0\" (UID: \"7b8bf821-52ae-409b-8e17-fc8e1302ed10\") " pod="glance-kuttl-tests/glance-default-external-api-0" Dec 01 09:01:32 crc kubenswrapper[4813]: I1201 09:01:32.903826 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-internal-api-0\" (UID: \"018e9026-a4ad-493c-9c2c-66d9ed3e36a5\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:01:32 crc kubenswrapper[4813]: I1201 09:01:32.903855 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/018e9026-a4ad-493c-9c2c-66d9ed3e36a5-var-locks-brick\") pod \"glance-default-internal-api-0\" (UID: \"018e9026-a4ad-493c-9c2c-66d9ed3e36a5\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:01:32 crc kubenswrapper[4813]: I1201 09:01:32.903878 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/7b8bf821-52ae-409b-8e17-fc8e1302ed10-sys\") pod \"glance-default-external-api-0\" (UID: \"7b8bf821-52ae-409b-8e17-fc8e1302ed10\") " pod="glance-kuttl-tests/glance-default-external-api-0" Dec 01 09:01:32 crc kubenswrapper[4813]: I1201 09:01:32.903895 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/018e9026-a4ad-493c-9c2c-66d9ed3e36a5-logs\") pod \"glance-default-internal-api-0\" (UID: \"018e9026-a4ad-493c-9c2c-66d9ed3e36a5\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:01:32 crc kubenswrapper[4813]: I1201 09:01:32.903931 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7b8bf821-52ae-409b-8e17-fc8e1302ed10-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"7b8bf821-52ae-409b-8e17-fc8e1302ed10\") " pod="glance-kuttl-tests/glance-default-external-api-0" Dec 01 09:01:32 crc kubenswrapper[4813]: I1201 09:01:32.903956 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/7b8bf821-52ae-409b-8e17-fc8e1302ed10-dev\") pod \"glance-default-external-api-0\" (UID: \"7b8bf821-52ae-409b-8e17-fc8e1302ed10\") " pod="glance-kuttl-tests/glance-default-external-api-0" Dec 01 09:01:32 crc kubenswrapper[4813]: I1201 09:01:32.903982 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/018e9026-a4ad-493c-9c2c-66d9ed3e36a5-etc-iscsi\") pod \"glance-default-internal-api-0\" (UID: \"018e9026-a4ad-493c-9c2c-66d9ed3e36a5\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:01:32 crc kubenswrapper[4813]: I1201 09:01:32.904051 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-0\" (UID: \"7b8bf821-52ae-409b-8e17-fc8e1302ed10\") " pod="glance-kuttl-tests/glance-default-external-api-0" Dec 01 09:01:32 crc kubenswrapper[4813]: I1201 09:01:32.904077 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-79rlt\" (UniqueName: \"kubernetes.io/projected/018e9026-a4ad-493c-9c2c-66d9ed3e36a5-kube-api-access-79rlt\") pod \"glance-default-internal-api-0\" (UID: \"018e9026-a4ad-493c-9c2c-66d9ed3e36a5\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:01:32 crc kubenswrapper[4813]: I1201 09:01:32.904102 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/7b8bf821-52ae-409b-8e17-fc8e1302ed10-lib-modules\") pod \"glance-default-external-api-0\" (UID: \"7b8bf821-52ae-409b-8e17-fc8e1302ed10\") " pod="glance-kuttl-tests/glance-default-external-api-0" Dec 01 09:01:32 crc kubenswrapper[4813]: I1201 09:01:32.904126 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/7b8bf821-52ae-409b-8e17-fc8e1302ed10-etc-nvme\") pod \"glance-default-external-api-0\" (UID: \"7b8bf821-52ae-409b-8e17-fc8e1302ed10\") " pod="glance-kuttl-tests/glance-default-external-api-0" Dec 01 09:01:32 crc kubenswrapper[4813]: I1201 09:01:32.904154 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-km49c\" (UniqueName: \"kubernetes.io/projected/7b8bf821-52ae-409b-8e17-fc8e1302ed10-kube-api-access-km49c\") pod \"glance-default-external-api-0\" (UID: \"7b8bf821-52ae-409b-8e17-fc8e1302ed10\") " pod="glance-kuttl-tests/glance-default-external-api-0" Dec 01 09:01:33 crc kubenswrapper[4813]: I1201 09:01:33.005340 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage16-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage16-crc\") pod \"glance-default-external-api-0\" (UID: \"7b8bf821-52ae-409b-8e17-fc8e1302ed10\") " pod="glance-kuttl-tests/glance-default-external-api-0" Dec 01 09:01:33 crc kubenswrapper[4813]: I1201 09:01:33.005381 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/018e9026-a4ad-493c-9c2c-66d9ed3e36a5-run\") pod \"glance-default-internal-api-0\" (UID: \"018e9026-a4ad-493c-9c2c-66d9ed3e36a5\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:01:33 crc kubenswrapper[4813]: I1201 09:01:33.005401 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-internal-api-0\" (UID: \"018e9026-a4ad-493c-9c2c-66d9ed3e36a5\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:01:33 crc kubenswrapper[4813]: I1201 09:01:33.005417 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/018e9026-a4ad-493c-9c2c-66d9ed3e36a5-scripts\") pod \"glance-default-internal-api-0\" (UID: \"018e9026-a4ad-493c-9c2c-66d9ed3e36a5\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:01:33 crc kubenswrapper[4813]: I1201 09:01:33.005528 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/018e9026-a4ad-493c-9c2c-66d9ed3e36a5-run\") pod \"glance-default-internal-api-0\" (UID: \"018e9026-a4ad-493c-9c2c-66d9ed3e36a5\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:01:33 crc kubenswrapper[4813]: I1201 09:01:33.005781 4813 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-internal-api-0\" (UID: \"018e9026-a4ad-493c-9c2c-66d9ed3e36a5\") device mount path \"/mnt/openstack/pv04\"" pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:01:33 crc kubenswrapper[4813]: I1201 09:01:33.005781 4813 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage16-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage16-crc\") pod \"glance-default-external-api-0\" (UID: \"7b8bf821-52ae-409b-8e17-fc8e1302ed10\") device mount path \"/mnt/openstack/pv16\"" pod="glance-kuttl-tests/glance-default-external-api-0" Dec 01 09:01:33 crc kubenswrapper[4813]: I1201 09:01:33.005872 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/7b8bf821-52ae-409b-8e17-fc8e1302ed10-etc-iscsi\") pod \"glance-default-external-api-0\" (UID: \"7b8bf821-52ae-409b-8e17-fc8e1302ed10\") " pod="glance-kuttl-tests/glance-default-external-api-0" Dec 01 09:01:33 crc kubenswrapper[4813]: I1201 09:01:33.005892 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/018e9026-a4ad-493c-9c2c-66d9ed3e36a5-sys\") pod \"glance-default-internal-api-0\" (UID: \"018e9026-a4ad-493c-9c2c-66d9ed3e36a5\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:01:33 crc kubenswrapper[4813]: I1201 09:01:33.005939 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/7b8bf821-52ae-409b-8e17-fc8e1302ed10-etc-iscsi\") pod \"glance-default-external-api-0\" (UID: \"7b8bf821-52ae-409b-8e17-fc8e1302ed10\") " pod="glance-kuttl-tests/glance-default-external-api-0" Dec 01 09:01:33 crc kubenswrapper[4813]: I1201 09:01:33.006087 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/018e9026-a4ad-493c-9c2c-66d9ed3e36a5-sys\") pod \"glance-default-internal-api-0\" (UID: \"018e9026-a4ad-493c-9c2c-66d9ed3e36a5\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:01:33 crc kubenswrapper[4813]: I1201 09:01:33.010785 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/018e9026-a4ad-493c-9c2c-66d9ed3e36a5-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"018e9026-a4ad-493c-9c2c-66d9ed3e36a5\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:01:33 crc kubenswrapper[4813]: I1201 09:01:33.010816 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/018e9026-a4ad-493c-9c2c-66d9ed3e36a5-dev\") pod \"glance-default-internal-api-0\" (UID: \"018e9026-a4ad-493c-9c2c-66d9ed3e36a5\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:01:33 crc kubenswrapper[4813]: I1201 09:01:33.010832 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7b8bf821-52ae-409b-8e17-fc8e1302ed10-logs\") pod \"glance-default-external-api-0\" (UID: \"7b8bf821-52ae-409b-8e17-fc8e1302ed10\") " pod="glance-kuttl-tests/glance-default-external-api-0" Dec 01 09:01:33 crc kubenswrapper[4813]: I1201 09:01:33.010860 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7b8bf821-52ae-409b-8e17-fc8e1302ed10-config-data\") pod \"glance-default-external-api-0\" (UID: \"7b8bf821-52ae-409b-8e17-fc8e1302ed10\") " pod="glance-kuttl-tests/glance-default-external-api-0" Dec 01 09:01:33 crc kubenswrapper[4813]: I1201 09:01:33.010874 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/7b8bf821-52ae-409b-8e17-fc8e1302ed10-var-locks-brick\") pod \"glance-default-external-api-0\" (UID: \"7b8bf821-52ae-409b-8e17-fc8e1302ed10\") " pod="glance-kuttl-tests/glance-default-external-api-0" Dec 01 09:01:33 crc kubenswrapper[4813]: I1201 09:01:33.010893 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/7b8bf821-52ae-409b-8e17-fc8e1302ed10-run\") pod \"glance-default-external-api-0\" (UID: \"7b8bf821-52ae-409b-8e17-fc8e1302ed10\") " pod="glance-kuttl-tests/glance-default-external-api-0" Dec 01 09:01:33 crc kubenswrapper[4813]: I1201 09:01:33.010907 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/018e9026-a4ad-493c-9c2c-66d9ed3e36a5-config-data\") pod \"glance-default-internal-api-0\" (UID: \"018e9026-a4ad-493c-9c2c-66d9ed3e36a5\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:01:33 crc kubenswrapper[4813]: I1201 09:01:33.010922 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-internal-api-0\" (UID: \"018e9026-a4ad-493c-9c2c-66d9ed3e36a5\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:01:33 crc kubenswrapper[4813]: I1201 09:01:33.010952 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/018e9026-a4ad-493c-9c2c-66d9ed3e36a5-var-locks-brick\") pod \"glance-default-internal-api-0\" (UID: \"018e9026-a4ad-493c-9c2c-66d9ed3e36a5\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:01:33 crc kubenswrapper[4813]: I1201 09:01:33.010976 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/7b8bf821-52ae-409b-8e17-fc8e1302ed10-sys\") pod \"glance-default-external-api-0\" (UID: \"7b8bf821-52ae-409b-8e17-fc8e1302ed10\") " pod="glance-kuttl-tests/glance-default-external-api-0" Dec 01 09:01:33 crc kubenswrapper[4813]: I1201 09:01:33.011004 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/018e9026-a4ad-493c-9c2c-66d9ed3e36a5-logs\") pod \"glance-default-internal-api-0\" (UID: \"018e9026-a4ad-493c-9c2c-66d9ed3e36a5\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:01:33 crc kubenswrapper[4813]: I1201 09:01:33.011043 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7b8bf821-52ae-409b-8e17-fc8e1302ed10-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"7b8bf821-52ae-409b-8e17-fc8e1302ed10\") " pod="glance-kuttl-tests/glance-default-external-api-0" Dec 01 09:01:33 crc kubenswrapper[4813]: I1201 09:01:33.011061 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/7b8bf821-52ae-409b-8e17-fc8e1302ed10-dev\") pod \"glance-default-external-api-0\" (UID: \"7b8bf821-52ae-409b-8e17-fc8e1302ed10\") " pod="glance-kuttl-tests/glance-default-external-api-0" Dec 01 09:01:33 crc kubenswrapper[4813]: I1201 09:01:33.011101 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/018e9026-a4ad-493c-9c2c-66d9ed3e36a5-etc-iscsi\") pod \"glance-default-internal-api-0\" (UID: \"018e9026-a4ad-493c-9c2c-66d9ed3e36a5\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:01:33 crc kubenswrapper[4813]: I1201 09:01:33.011127 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-0\" (UID: \"7b8bf821-52ae-409b-8e17-fc8e1302ed10\") " pod="glance-kuttl-tests/glance-default-external-api-0" Dec 01 09:01:33 crc kubenswrapper[4813]: I1201 09:01:33.011159 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-79rlt\" (UniqueName: \"kubernetes.io/projected/018e9026-a4ad-493c-9c2c-66d9ed3e36a5-kube-api-access-79rlt\") pod \"glance-default-internal-api-0\" (UID: \"018e9026-a4ad-493c-9c2c-66d9ed3e36a5\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:01:33 crc kubenswrapper[4813]: I1201 09:01:33.011186 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/7b8bf821-52ae-409b-8e17-fc8e1302ed10-lib-modules\") pod \"glance-default-external-api-0\" (UID: \"7b8bf821-52ae-409b-8e17-fc8e1302ed10\") " pod="glance-kuttl-tests/glance-default-external-api-0" Dec 01 09:01:33 crc kubenswrapper[4813]: I1201 09:01:33.011208 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/7b8bf821-52ae-409b-8e17-fc8e1302ed10-etc-nvme\") pod \"glance-default-external-api-0\" (UID: \"7b8bf821-52ae-409b-8e17-fc8e1302ed10\") " pod="glance-kuttl-tests/glance-default-external-api-0" Dec 01 09:01:33 crc kubenswrapper[4813]: I1201 09:01:33.011239 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-km49c\" (UniqueName: \"kubernetes.io/projected/7b8bf821-52ae-409b-8e17-fc8e1302ed10-kube-api-access-km49c\") pod \"glance-default-external-api-0\" (UID: \"7b8bf821-52ae-409b-8e17-fc8e1302ed10\") " pod="glance-kuttl-tests/glance-default-external-api-0" Dec 01 09:01:33 crc kubenswrapper[4813]: I1201 09:01:33.011246 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/018e9026-a4ad-493c-9c2c-66d9ed3e36a5-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"018e9026-a4ad-493c-9c2c-66d9ed3e36a5\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:01:33 crc kubenswrapper[4813]: I1201 09:01:33.011263 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7b8bf821-52ae-409b-8e17-fc8e1302ed10-scripts\") pod \"glance-default-external-api-0\" (UID: \"7b8bf821-52ae-409b-8e17-fc8e1302ed10\") " pod="glance-kuttl-tests/glance-default-external-api-0" Dec 01 09:01:33 crc kubenswrapper[4813]: I1201 09:01:33.011374 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7b8bf821-52ae-409b-8e17-fc8e1302ed10-logs\") pod \"glance-default-external-api-0\" (UID: \"7b8bf821-52ae-409b-8e17-fc8e1302ed10\") " pod="glance-kuttl-tests/glance-default-external-api-0" Dec 01 09:01:33 crc kubenswrapper[4813]: I1201 09:01:33.011405 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/018e9026-a4ad-493c-9c2c-66d9ed3e36a5-lib-modules\") pod \"glance-default-internal-api-0\" (UID: \"018e9026-a4ad-493c-9c2c-66d9ed3e36a5\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:01:33 crc kubenswrapper[4813]: I1201 09:01:33.011419 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/018e9026-a4ad-493c-9c2c-66d9ed3e36a5-dev\") pod \"glance-default-internal-api-0\" (UID: \"018e9026-a4ad-493c-9c2c-66d9ed3e36a5\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:01:33 crc kubenswrapper[4813]: I1201 09:01:33.011436 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/018e9026-a4ad-493c-9c2c-66d9ed3e36a5-etc-nvme\") pod \"glance-default-internal-api-0\" (UID: \"018e9026-a4ad-493c-9c2c-66d9ed3e36a5\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:01:33 crc kubenswrapper[4813]: I1201 09:01:33.011569 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/018e9026-a4ad-493c-9c2c-66d9ed3e36a5-logs\") pod \"glance-default-internal-api-0\" (UID: \"018e9026-a4ad-493c-9c2c-66d9ed3e36a5\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:01:33 crc kubenswrapper[4813]: I1201 09:01:33.011664 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/018e9026-a4ad-493c-9c2c-66d9ed3e36a5-etc-nvme\") pod \"glance-default-internal-api-0\" (UID: \"018e9026-a4ad-493c-9c2c-66d9ed3e36a5\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:01:33 crc kubenswrapper[4813]: I1201 09:01:33.011677 4813 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-internal-api-0\" (UID: \"018e9026-a4ad-493c-9c2c-66d9ed3e36a5\") device mount path \"/mnt/openstack/pv03\"" pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:01:33 crc kubenswrapper[4813]: I1201 09:01:33.011700 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/018e9026-a4ad-493c-9c2c-66d9ed3e36a5-lib-modules\") pod \"glance-default-internal-api-0\" (UID: \"018e9026-a4ad-493c-9c2c-66d9ed3e36a5\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:01:33 crc kubenswrapper[4813]: I1201 09:01:33.011759 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/7b8bf821-52ae-409b-8e17-fc8e1302ed10-var-locks-brick\") pod \"glance-default-external-api-0\" (UID: \"7b8bf821-52ae-409b-8e17-fc8e1302ed10\") " pod="glance-kuttl-tests/glance-default-external-api-0" Dec 01 09:01:33 crc kubenswrapper[4813]: I1201 09:01:33.012048 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/018e9026-a4ad-493c-9c2c-66d9ed3e36a5-var-locks-brick\") pod \"glance-default-internal-api-0\" (UID: \"018e9026-a4ad-493c-9c2c-66d9ed3e36a5\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:01:33 crc kubenswrapper[4813]: I1201 09:01:33.012049 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/7b8bf821-52ae-409b-8e17-fc8e1302ed10-sys\") pod \"glance-default-external-api-0\" (UID: \"7b8bf821-52ae-409b-8e17-fc8e1302ed10\") " pod="glance-kuttl-tests/glance-default-external-api-0" Dec 01 09:01:33 crc kubenswrapper[4813]: I1201 09:01:33.012127 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/7b8bf821-52ae-409b-8e17-fc8e1302ed10-dev\") pod \"glance-default-external-api-0\" (UID: \"7b8bf821-52ae-409b-8e17-fc8e1302ed10\") " pod="glance-kuttl-tests/glance-default-external-api-0" Dec 01 09:01:33 crc kubenswrapper[4813]: I1201 09:01:33.012159 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/018e9026-a4ad-493c-9c2c-66d9ed3e36a5-etc-iscsi\") pod \"glance-default-internal-api-0\" (UID: \"018e9026-a4ad-493c-9c2c-66d9ed3e36a5\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:01:33 crc kubenswrapper[4813]: I1201 09:01:33.012194 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/7b8bf821-52ae-409b-8e17-fc8e1302ed10-lib-modules\") pod \"glance-default-external-api-0\" (UID: \"7b8bf821-52ae-409b-8e17-fc8e1302ed10\") " pod="glance-kuttl-tests/glance-default-external-api-0" Dec 01 09:01:33 crc kubenswrapper[4813]: I1201 09:01:33.012217 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/7b8bf821-52ae-409b-8e17-fc8e1302ed10-run\") pod \"glance-default-external-api-0\" (UID: \"7b8bf821-52ae-409b-8e17-fc8e1302ed10\") " pod="glance-kuttl-tests/glance-default-external-api-0" Dec 01 09:01:33 crc kubenswrapper[4813]: I1201 09:01:33.012276 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/7b8bf821-52ae-409b-8e17-fc8e1302ed10-etc-nvme\") pod \"glance-default-external-api-0\" (UID: \"7b8bf821-52ae-409b-8e17-fc8e1302ed10\") " pod="glance-kuttl-tests/glance-default-external-api-0" Dec 01 09:01:33 crc kubenswrapper[4813]: I1201 09:01:33.012341 4813 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-0\" (UID: \"7b8bf821-52ae-409b-8e17-fc8e1302ed10\") device mount path \"/mnt/openstack/pv10\"" pod="glance-kuttl-tests/glance-default-external-api-0" Dec 01 09:01:33 crc kubenswrapper[4813]: I1201 09:01:33.012419 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7b8bf821-52ae-409b-8e17-fc8e1302ed10-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"7b8bf821-52ae-409b-8e17-fc8e1302ed10\") " pod="glance-kuttl-tests/glance-default-external-api-0" Dec 01 09:01:33 crc kubenswrapper[4813]: I1201 09:01:33.013884 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/018e9026-a4ad-493c-9c2c-66d9ed3e36a5-scripts\") pod \"glance-default-internal-api-0\" (UID: \"018e9026-a4ad-493c-9c2c-66d9ed3e36a5\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:01:33 crc kubenswrapper[4813]: I1201 09:01:33.015147 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7b8bf821-52ae-409b-8e17-fc8e1302ed10-scripts\") pod \"glance-default-external-api-0\" (UID: \"7b8bf821-52ae-409b-8e17-fc8e1302ed10\") " pod="glance-kuttl-tests/glance-default-external-api-0" Dec 01 09:01:33 crc kubenswrapper[4813]: I1201 09:01:33.017202 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7b8bf821-52ae-409b-8e17-fc8e1302ed10-config-data\") pod \"glance-default-external-api-0\" (UID: \"7b8bf821-52ae-409b-8e17-fc8e1302ed10\") " pod="glance-kuttl-tests/glance-default-external-api-0" Dec 01 09:01:33 crc kubenswrapper[4813]: I1201 09:01:33.028890 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/018e9026-a4ad-493c-9c2c-66d9ed3e36a5-config-data\") pod \"glance-default-internal-api-0\" (UID: \"018e9026-a4ad-493c-9c2c-66d9ed3e36a5\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:01:33 crc kubenswrapper[4813]: I1201 09:01:33.032514 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-internal-api-0\" (UID: \"018e9026-a4ad-493c-9c2c-66d9ed3e36a5\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:01:33 crc kubenswrapper[4813]: I1201 09:01:33.038800 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-79rlt\" (UniqueName: \"kubernetes.io/projected/018e9026-a4ad-493c-9c2c-66d9ed3e36a5-kube-api-access-79rlt\") pod \"glance-default-internal-api-0\" (UID: \"018e9026-a4ad-493c-9c2c-66d9ed3e36a5\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:01:33 crc kubenswrapper[4813]: I1201 09:01:33.045190 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-km49c\" (UniqueName: \"kubernetes.io/projected/7b8bf821-52ae-409b-8e17-fc8e1302ed10-kube-api-access-km49c\") pod \"glance-default-external-api-0\" (UID: \"7b8bf821-52ae-409b-8e17-fc8e1302ed10\") " pod="glance-kuttl-tests/glance-default-external-api-0" Dec 01 09:01:33 crc kubenswrapper[4813]: I1201 09:01:33.053030 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-0\" (UID: \"7b8bf821-52ae-409b-8e17-fc8e1302ed10\") " pod="glance-kuttl-tests/glance-default-external-api-0" Dec 01 09:01:33 crc kubenswrapper[4813]: I1201 09:01:33.054386 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-internal-api-0\" (UID: \"018e9026-a4ad-493c-9c2c-66d9ed3e36a5\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:01:33 crc kubenswrapper[4813]: I1201 09:01:33.059665 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage16-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage16-crc\") pod \"glance-default-external-api-0\" (UID: \"7b8bf821-52ae-409b-8e17-fc8e1302ed10\") " pod="glance-kuttl-tests/glance-default-external-api-0" Dec 01 09:01:33 crc kubenswrapper[4813]: I1201 09:01:33.126594 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:01:33 crc kubenswrapper[4813]: I1201 09:01:33.345171 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-0" Dec 01 09:01:33 crc kubenswrapper[4813]: I1201 09:01:33.598245 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Dec 01 09:01:33 crc kubenswrapper[4813]: I1201 09:01:33.764009 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-0"] Dec 01 09:01:33 crc kubenswrapper[4813]: W1201 09:01:33.767129 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7b8bf821_52ae_409b_8e17_fc8e1302ed10.slice/crio-ee0e6ad4fc73cf0cefcbbef4c9024d0c4e23c891a852b9d53bae9d0ab1476118 WatchSource:0}: Error finding container ee0e6ad4fc73cf0cefcbbef4c9024d0c4e23c891a852b9d53bae9d0ab1476118: Status 404 returned error can't find the container with id ee0e6ad4fc73cf0cefcbbef4c9024d0c4e23c891a852b9d53bae9d0ab1476118 Dec 01 09:01:33 crc kubenswrapper[4813]: I1201 09:01:33.930888 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Dec 01 09:01:34 crc kubenswrapper[4813]: I1201 09:01:34.416618 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"018e9026-a4ad-493c-9c2c-66d9ed3e36a5","Type":"ContainerStarted","Data":"a873a3860dd04d40ab66031010f614cdd1860c405a9a5504d2cbeb521feb7210"} Dec 01 09:01:34 crc kubenswrapper[4813]: I1201 09:01:34.417046 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"018e9026-a4ad-493c-9c2c-66d9ed3e36a5","Type":"ContainerStarted","Data":"08d90d01c529c5419aad930c125e2644e493e7185a0d36892250941689688ffc"} Dec 01 09:01:34 crc kubenswrapper[4813]: I1201 09:01:34.418217 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-0" event={"ID":"7b8bf821-52ae-409b-8e17-fc8e1302ed10","Type":"ContainerStarted","Data":"56c363c42bb8f1be549639c72070e76cf11757d4ca387266e8486456ffae3383"} Dec 01 09:01:34 crc kubenswrapper[4813]: I1201 09:01:34.418270 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-0" event={"ID":"7b8bf821-52ae-409b-8e17-fc8e1302ed10","Type":"ContainerStarted","Data":"ee0e6ad4fc73cf0cefcbbef4c9024d0c4e23c891a852b9d53bae9d0ab1476118"} Dec 01 09:01:35 crc kubenswrapper[4813]: I1201 09:01:35.428451 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"018e9026-a4ad-493c-9c2c-66d9ed3e36a5","Type":"ContainerStarted","Data":"dde2dcff10fe5a6481898c71c15762a3789d9b276021fd4f11415ad475ac9567"} Dec 01 09:01:35 crc kubenswrapper[4813]: I1201 09:01:35.428579 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-internal-api-0" podUID="018e9026-a4ad-493c-9c2c-66d9ed3e36a5" containerName="glance-log" containerID="cri-o://a873a3860dd04d40ab66031010f614cdd1860c405a9a5504d2cbeb521feb7210" gracePeriod=30 Dec 01 09:01:35 crc kubenswrapper[4813]: I1201 09:01:35.428638 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-internal-api-0" podUID="018e9026-a4ad-493c-9c2c-66d9ed3e36a5" containerName="glance-httpd" containerID="cri-o://dde2dcff10fe5a6481898c71c15762a3789d9b276021fd4f11415ad475ac9567" gracePeriod=30 Dec 01 09:01:35 crc kubenswrapper[4813]: I1201 09:01:35.431735 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-0" event={"ID":"7b8bf821-52ae-409b-8e17-fc8e1302ed10","Type":"ContainerStarted","Data":"a22c275b1abdf242357c033d2a976a1e0bf84f3b733c91633972190ba9de2c51"} Dec 01 09:01:35 crc kubenswrapper[4813]: I1201 09:01:35.449813 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-default-internal-api-0" podStartSLOduration=4.449796175 podStartE2EDuration="4.449796175s" podCreationTimestamp="2025-12-01 09:01:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 09:01:35.44855351 +0000 UTC m=+1280.071475096" watchObservedRunningTime="2025-12-01 09:01:35.449796175 +0000 UTC m=+1280.072717761" Dec 01 09:01:36 crc kubenswrapper[4813]: I1201 09:01:36.443834 4813 generic.go:334] "Generic (PLEG): container finished" podID="018e9026-a4ad-493c-9c2c-66d9ed3e36a5" containerID="dde2dcff10fe5a6481898c71c15762a3789d9b276021fd4f11415ad475ac9567" exitCode=143 Dec 01 09:01:36 crc kubenswrapper[4813]: I1201 09:01:36.444440 4813 generic.go:334] "Generic (PLEG): container finished" podID="018e9026-a4ad-493c-9c2c-66d9ed3e36a5" containerID="a873a3860dd04d40ab66031010f614cdd1860c405a9a5504d2cbeb521feb7210" exitCode=143 Dec 01 09:01:36 crc kubenswrapper[4813]: I1201 09:01:36.443880 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"018e9026-a4ad-493c-9c2c-66d9ed3e36a5","Type":"ContainerDied","Data":"dde2dcff10fe5a6481898c71c15762a3789d9b276021fd4f11415ad475ac9567"} Dec 01 09:01:36 crc kubenswrapper[4813]: I1201 09:01:36.444610 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"018e9026-a4ad-493c-9c2c-66d9ed3e36a5","Type":"ContainerDied","Data":"a873a3860dd04d40ab66031010f614cdd1860c405a9a5504d2cbeb521feb7210"} Dec 01 09:01:36 crc kubenswrapper[4813]: I1201 09:01:36.796568 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:01:36 crc kubenswrapper[4813]: I1201 09:01:36.828016 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-default-external-api-0" podStartSLOduration=4.827998402 podStartE2EDuration="4.827998402s" podCreationTimestamp="2025-12-01 09:01:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 09:01:35.482533714 +0000 UTC m=+1280.105455300" watchObservedRunningTime="2025-12-01 09:01:36.827998402 +0000 UTC m=+1281.450919988" Dec 01 09:01:36 crc kubenswrapper[4813]: I1201 09:01:36.867792 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/018e9026-a4ad-493c-9c2c-66d9ed3e36a5-sys\") pod \"018e9026-a4ad-493c-9c2c-66d9ed3e36a5\" (UID: \"018e9026-a4ad-493c-9c2c-66d9ed3e36a5\") " Dec 01 09:01:36 crc kubenswrapper[4813]: I1201 09:01:36.867857 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/018e9026-a4ad-493c-9c2c-66d9ed3e36a5-config-data\") pod \"018e9026-a4ad-493c-9c2c-66d9ed3e36a5\" (UID: \"018e9026-a4ad-493c-9c2c-66d9ed3e36a5\") " Dec 01 09:01:36 crc kubenswrapper[4813]: I1201 09:01:36.867881 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/018e9026-a4ad-493c-9c2c-66d9ed3e36a5-var-locks-brick\") pod \"018e9026-a4ad-493c-9c2c-66d9ed3e36a5\" (UID: \"018e9026-a4ad-493c-9c2c-66d9ed3e36a5\") " Dec 01 09:01:36 crc kubenswrapper[4813]: I1201 09:01:36.867919 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-79rlt\" (UniqueName: \"kubernetes.io/projected/018e9026-a4ad-493c-9c2c-66d9ed3e36a5-kube-api-access-79rlt\") pod \"018e9026-a4ad-493c-9c2c-66d9ed3e36a5\" (UID: \"018e9026-a4ad-493c-9c2c-66d9ed3e36a5\") " Dec 01 09:01:36 crc kubenswrapper[4813]: I1201 09:01:36.867942 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/018e9026-a4ad-493c-9c2c-66d9ed3e36a5-httpd-run\") pod \"018e9026-a4ad-493c-9c2c-66d9ed3e36a5\" (UID: \"018e9026-a4ad-493c-9c2c-66d9ed3e36a5\") " Dec 01 09:01:36 crc kubenswrapper[4813]: I1201 09:01:36.867984 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/018e9026-a4ad-493c-9c2c-66d9ed3e36a5-scripts\") pod \"018e9026-a4ad-493c-9c2c-66d9ed3e36a5\" (UID: \"018e9026-a4ad-493c-9c2c-66d9ed3e36a5\") " Dec 01 09:01:36 crc kubenswrapper[4813]: I1201 09:01:36.868017 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/018e9026-a4ad-493c-9c2c-66d9ed3e36a5-dev\") pod \"018e9026-a4ad-493c-9c2c-66d9ed3e36a5\" (UID: \"018e9026-a4ad-493c-9c2c-66d9ed3e36a5\") " Dec 01 09:01:36 crc kubenswrapper[4813]: I1201 09:01:36.868042 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance-cache\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"018e9026-a4ad-493c-9c2c-66d9ed3e36a5\" (UID: \"018e9026-a4ad-493c-9c2c-66d9ed3e36a5\") " Dec 01 09:01:36 crc kubenswrapper[4813]: I1201 09:01:36.868060 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"018e9026-a4ad-493c-9c2c-66d9ed3e36a5\" (UID: \"018e9026-a4ad-493c-9c2c-66d9ed3e36a5\") " Dec 01 09:01:36 crc kubenswrapper[4813]: I1201 09:01:36.868097 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/018e9026-a4ad-493c-9c2c-66d9ed3e36a5-etc-nvme\") pod \"018e9026-a4ad-493c-9c2c-66d9ed3e36a5\" (UID: \"018e9026-a4ad-493c-9c2c-66d9ed3e36a5\") " Dec 01 09:01:36 crc kubenswrapper[4813]: I1201 09:01:36.868119 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/018e9026-a4ad-493c-9c2c-66d9ed3e36a5-etc-iscsi\") pod \"018e9026-a4ad-493c-9c2c-66d9ed3e36a5\" (UID: \"018e9026-a4ad-493c-9c2c-66d9ed3e36a5\") " Dec 01 09:01:36 crc kubenswrapper[4813]: I1201 09:01:36.868147 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/018e9026-a4ad-493c-9c2c-66d9ed3e36a5-logs\") pod \"018e9026-a4ad-493c-9c2c-66d9ed3e36a5\" (UID: \"018e9026-a4ad-493c-9c2c-66d9ed3e36a5\") " Dec 01 09:01:36 crc kubenswrapper[4813]: I1201 09:01:36.868174 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/018e9026-a4ad-493c-9c2c-66d9ed3e36a5-run\") pod \"018e9026-a4ad-493c-9c2c-66d9ed3e36a5\" (UID: \"018e9026-a4ad-493c-9c2c-66d9ed3e36a5\") " Dec 01 09:01:36 crc kubenswrapper[4813]: I1201 09:01:36.868194 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/018e9026-a4ad-493c-9c2c-66d9ed3e36a5-lib-modules\") pod \"018e9026-a4ad-493c-9c2c-66d9ed3e36a5\" (UID: \"018e9026-a4ad-493c-9c2c-66d9ed3e36a5\") " Dec 01 09:01:36 crc kubenswrapper[4813]: I1201 09:01:36.868498 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/018e9026-a4ad-493c-9c2c-66d9ed3e36a5-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "018e9026-a4ad-493c-9c2c-66d9ed3e36a5" (UID: "018e9026-a4ad-493c-9c2c-66d9ed3e36a5"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 09:01:36 crc kubenswrapper[4813]: I1201 09:01:36.868545 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/018e9026-a4ad-493c-9c2c-66d9ed3e36a5-sys" (OuterVolumeSpecName: "sys") pod "018e9026-a4ad-493c-9c2c-66d9ed3e36a5" (UID: "018e9026-a4ad-493c-9c2c-66d9ed3e36a5"). InnerVolumeSpecName "sys". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 09:01:36 crc kubenswrapper[4813]: I1201 09:01:36.876422 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/018e9026-a4ad-493c-9c2c-66d9ed3e36a5-etc-iscsi" (OuterVolumeSpecName: "etc-iscsi") pod "018e9026-a4ad-493c-9c2c-66d9ed3e36a5" (UID: "018e9026-a4ad-493c-9c2c-66d9ed3e36a5"). InnerVolumeSpecName "etc-iscsi". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 09:01:36 crc kubenswrapper[4813]: I1201 09:01:36.876637 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/018e9026-a4ad-493c-9c2c-66d9ed3e36a5-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "018e9026-a4ad-493c-9c2c-66d9ed3e36a5" (UID: "018e9026-a4ad-493c-9c2c-66d9ed3e36a5"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 09:01:36 crc kubenswrapper[4813]: I1201 09:01:36.876700 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/018e9026-a4ad-493c-9c2c-66d9ed3e36a5-var-locks-brick" (OuterVolumeSpecName: "var-locks-brick") pod "018e9026-a4ad-493c-9c2c-66d9ed3e36a5" (UID: "018e9026-a4ad-493c-9c2c-66d9ed3e36a5"). InnerVolumeSpecName "var-locks-brick". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 09:01:36 crc kubenswrapper[4813]: I1201 09:01:36.877232 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage03-crc" (OuterVolumeSpecName: "glance-cache") pod "018e9026-a4ad-493c-9c2c-66d9ed3e36a5" (UID: "018e9026-a4ad-493c-9c2c-66d9ed3e36a5"). InnerVolumeSpecName "local-storage03-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 01 09:01:36 crc kubenswrapper[4813]: I1201 09:01:36.878838 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/018e9026-a4ad-493c-9c2c-66d9ed3e36a5-etc-nvme" (OuterVolumeSpecName: "etc-nvme") pod "018e9026-a4ad-493c-9c2c-66d9ed3e36a5" (UID: "018e9026-a4ad-493c-9c2c-66d9ed3e36a5"). InnerVolumeSpecName "etc-nvme". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 09:01:36 crc kubenswrapper[4813]: I1201 09:01:36.878915 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/018e9026-a4ad-493c-9c2c-66d9ed3e36a5-dev" (OuterVolumeSpecName: "dev") pod "018e9026-a4ad-493c-9c2c-66d9ed3e36a5" (UID: "018e9026-a4ad-493c-9c2c-66d9ed3e36a5"). InnerVolumeSpecName "dev". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 09:01:36 crc kubenswrapper[4813]: I1201 09:01:36.878938 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/018e9026-a4ad-493c-9c2c-66d9ed3e36a5-run" (OuterVolumeSpecName: "run") pod "018e9026-a4ad-493c-9c2c-66d9ed3e36a5" (UID: "018e9026-a4ad-493c-9c2c-66d9ed3e36a5"). InnerVolumeSpecName "run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 09:01:36 crc kubenswrapper[4813]: I1201 09:01:36.879027 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/018e9026-a4ad-493c-9c2c-66d9ed3e36a5-logs" (OuterVolumeSpecName: "logs") pod "018e9026-a4ad-493c-9c2c-66d9ed3e36a5" (UID: "018e9026-a4ad-493c-9c2c-66d9ed3e36a5"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 09:01:36 crc kubenswrapper[4813]: I1201 09:01:36.879792 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage04-crc" (OuterVolumeSpecName: "glance") pod "018e9026-a4ad-493c-9c2c-66d9ed3e36a5" (UID: "018e9026-a4ad-493c-9c2c-66d9ed3e36a5"). InnerVolumeSpecName "local-storage04-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 01 09:01:36 crc kubenswrapper[4813]: I1201 09:01:36.882147 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/018e9026-a4ad-493c-9c2c-66d9ed3e36a5-scripts" (OuterVolumeSpecName: "scripts") pod "018e9026-a4ad-493c-9c2c-66d9ed3e36a5" (UID: "018e9026-a4ad-493c-9c2c-66d9ed3e36a5"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:01:36 crc kubenswrapper[4813]: I1201 09:01:36.883136 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/018e9026-a4ad-493c-9c2c-66d9ed3e36a5-kube-api-access-79rlt" (OuterVolumeSpecName: "kube-api-access-79rlt") pod "018e9026-a4ad-493c-9c2c-66d9ed3e36a5" (UID: "018e9026-a4ad-493c-9c2c-66d9ed3e36a5"). InnerVolumeSpecName "kube-api-access-79rlt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:01:36 crc kubenswrapper[4813]: I1201 09:01:36.906832 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/018e9026-a4ad-493c-9c2c-66d9ed3e36a5-config-data" (OuterVolumeSpecName: "config-data") pod "018e9026-a4ad-493c-9c2c-66d9ed3e36a5" (UID: "018e9026-a4ad-493c-9c2c-66d9ed3e36a5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:01:36 crc kubenswrapper[4813]: I1201 09:01:36.969745 4813 reconciler_common.go:293] "Volume detached for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/018e9026-a4ad-493c-9c2c-66d9ed3e36a5-etc-nvme\") on node \"crc\" DevicePath \"\"" Dec 01 09:01:36 crc kubenswrapper[4813]: I1201 09:01:36.969783 4813 reconciler_common.go:293] "Volume detached for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/018e9026-a4ad-493c-9c2c-66d9ed3e36a5-etc-iscsi\") on node \"crc\" DevicePath \"\"" Dec 01 09:01:36 crc kubenswrapper[4813]: I1201 09:01:36.969794 4813 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/018e9026-a4ad-493c-9c2c-66d9ed3e36a5-logs\") on node \"crc\" DevicePath \"\"" Dec 01 09:01:36 crc kubenswrapper[4813]: I1201 09:01:36.969803 4813 reconciler_common.go:293] "Volume detached for volume \"run\" (UniqueName: \"kubernetes.io/host-path/018e9026-a4ad-493c-9c2c-66d9ed3e36a5-run\") on node \"crc\" DevicePath \"\"" Dec 01 09:01:36 crc kubenswrapper[4813]: I1201 09:01:36.969811 4813 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/018e9026-a4ad-493c-9c2c-66d9ed3e36a5-lib-modules\") on node \"crc\" DevicePath \"\"" Dec 01 09:01:36 crc kubenswrapper[4813]: I1201 09:01:36.969820 4813 reconciler_common.go:293] "Volume detached for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/018e9026-a4ad-493c-9c2c-66d9ed3e36a5-sys\") on node \"crc\" DevicePath \"\"" Dec 01 09:01:36 crc kubenswrapper[4813]: I1201 09:01:36.969829 4813 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/018e9026-a4ad-493c-9c2c-66d9ed3e36a5-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 09:01:36 crc kubenswrapper[4813]: I1201 09:01:36.969837 4813 reconciler_common.go:293] "Volume detached for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/018e9026-a4ad-493c-9c2c-66d9ed3e36a5-var-locks-brick\") on node \"crc\" DevicePath \"\"" Dec 01 09:01:36 crc kubenswrapper[4813]: I1201 09:01:36.969846 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-79rlt\" (UniqueName: \"kubernetes.io/projected/018e9026-a4ad-493c-9c2c-66d9ed3e36a5-kube-api-access-79rlt\") on node \"crc\" DevicePath \"\"" Dec 01 09:01:36 crc kubenswrapper[4813]: I1201 09:01:36.969856 4813 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/018e9026-a4ad-493c-9c2c-66d9ed3e36a5-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 01 09:01:36 crc kubenswrapper[4813]: I1201 09:01:36.969864 4813 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/018e9026-a4ad-493c-9c2c-66d9ed3e36a5-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 09:01:36 crc kubenswrapper[4813]: I1201 09:01:36.969871 4813 reconciler_common.go:293] "Volume detached for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/018e9026-a4ad-493c-9c2c-66d9ed3e36a5-dev\") on node \"crc\" DevicePath \"\"" Dec 01 09:01:36 crc kubenswrapper[4813]: I1201 09:01:36.969904 4813 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" " Dec 01 09:01:36 crc kubenswrapper[4813]: I1201 09:01:36.969917 4813 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" " Dec 01 09:01:36 crc kubenswrapper[4813]: I1201 09:01:36.983051 4813 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage04-crc" (UniqueName: "kubernetes.io/local-volume/local-storage04-crc") on node "crc" Dec 01 09:01:36 crc kubenswrapper[4813]: I1201 09:01:36.992446 4813 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage03-crc" (UniqueName: "kubernetes.io/local-volume/local-storage03-crc") on node "crc" Dec 01 09:01:37 crc kubenswrapper[4813]: I1201 09:01:37.071482 4813 reconciler_common.go:293] "Volume detached for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" DevicePath \"\"" Dec 01 09:01:37 crc kubenswrapper[4813]: I1201 09:01:37.071519 4813 reconciler_common.go:293] "Volume detached for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" DevicePath \"\"" Dec 01 09:01:37 crc kubenswrapper[4813]: I1201 09:01:37.455264 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"018e9026-a4ad-493c-9c2c-66d9ed3e36a5","Type":"ContainerDied","Data":"08d90d01c529c5419aad930c125e2644e493e7185a0d36892250941689688ffc"} Dec 01 09:01:37 crc kubenswrapper[4813]: I1201 09:01:37.455321 4813 scope.go:117] "RemoveContainer" containerID="dde2dcff10fe5a6481898c71c15762a3789d9b276021fd4f11415ad475ac9567" Dec 01 09:01:37 crc kubenswrapper[4813]: I1201 09:01:37.455377 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:01:37 crc kubenswrapper[4813]: I1201 09:01:37.474638 4813 scope.go:117] "RemoveContainer" containerID="a873a3860dd04d40ab66031010f614cdd1860c405a9a5504d2cbeb521feb7210" Dec 01 09:01:37 crc kubenswrapper[4813]: I1201 09:01:37.501290 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Dec 01 09:01:37 crc kubenswrapper[4813]: I1201 09:01:37.506498 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Dec 01 09:01:37 crc kubenswrapper[4813]: I1201 09:01:37.542684 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Dec 01 09:01:37 crc kubenswrapper[4813]: E1201 09:01:37.543322 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="018e9026-a4ad-493c-9c2c-66d9ed3e36a5" containerName="glance-httpd" Dec 01 09:01:37 crc kubenswrapper[4813]: I1201 09:01:37.543344 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="018e9026-a4ad-493c-9c2c-66d9ed3e36a5" containerName="glance-httpd" Dec 01 09:01:37 crc kubenswrapper[4813]: E1201 09:01:37.543369 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="018e9026-a4ad-493c-9c2c-66d9ed3e36a5" containerName="glance-log" Dec 01 09:01:37 crc kubenswrapper[4813]: I1201 09:01:37.543380 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="018e9026-a4ad-493c-9c2c-66d9ed3e36a5" containerName="glance-log" Dec 01 09:01:37 crc kubenswrapper[4813]: I1201 09:01:37.543658 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="018e9026-a4ad-493c-9c2c-66d9ed3e36a5" containerName="glance-log" Dec 01 09:01:37 crc kubenswrapper[4813]: I1201 09:01:37.543694 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="018e9026-a4ad-493c-9c2c-66d9ed3e36a5" containerName="glance-httpd" Dec 01 09:01:37 crc kubenswrapper[4813]: I1201 09:01:37.545003 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:01:37 crc kubenswrapper[4813]: I1201 09:01:37.551483 4813 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-default-internal-config-data" Dec 01 09:01:37 crc kubenswrapper[4813]: I1201 09:01:37.577190 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-internal-api-0\" (UID: \"a3fd4804-b7ff-42d4-aa58-999cb59efbc7\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:01:37 crc kubenswrapper[4813]: I1201 09:01:37.577475 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Dec 01 09:01:37 crc kubenswrapper[4813]: I1201 09:01:37.678474 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/a3fd4804-b7ff-42d4-aa58-999cb59efbc7-var-locks-brick\") pod \"glance-default-internal-api-0\" (UID: \"a3fd4804-b7ff-42d4-aa58-999cb59efbc7\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:01:37 crc kubenswrapper[4813]: I1201 09:01:37.678787 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/a3fd4804-b7ff-42d4-aa58-999cb59efbc7-lib-modules\") pod \"glance-default-internal-api-0\" (UID: \"a3fd4804-b7ff-42d4-aa58-999cb59efbc7\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:01:37 crc kubenswrapper[4813]: I1201 09:01:37.678820 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a3fd4804-b7ff-42d4-aa58-999cb59efbc7-scripts\") pod \"glance-default-internal-api-0\" (UID: \"a3fd4804-b7ff-42d4-aa58-999cb59efbc7\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:01:37 crc kubenswrapper[4813]: I1201 09:01:37.678871 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/a3fd4804-b7ff-42d4-aa58-999cb59efbc7-etc-nvme\") pod \"glance-default-internal-api-0\" (UID: \"a3fd4804-b7ff-42d4-aa58-999cb59efbc7\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:01:37 crc kubenswrapper[4813]: I1201 09:01:37.678895 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/a3fd4804-b7ff-42d4-aa58-999cb59efbc7-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"a3fd4804-b7ff-42d4-aa58-999cb59efbc7\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:01:37 crc kubenswrapper[4813]: I1201 09:01:37.678919 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/a3fd4804-b7ff-42d4-aa58-999cb59efbc7-dev\") pod \"glance-default-internal-api-0\" (UID: \"a3fd4804-b7ff-42d4-aa58-999cb59efbc7\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:01:37 crc kubenswrapper[4813]: I1201 09:01:37.678973 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-internal-api-0\" (UID: \"a3fd4804-b7ff-42d4-aa58-999cb59efbc7\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:01:37 crc kubenswrapper[4813]: I1201 09:01:37.679012 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/a3fd4804-b7ff-42d4-aa58-999cb59efbc7-sys\") pod \"glance-default-internal-api-0\" (UID: \"a3fd4804-b7ff-42d4-aa58-999cb59efbc7\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:01:37 crc kubenswrapper[4813]: I1201 09:01:37.679048 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a3fd4804-b7ff-42d4-aa58-999cb59efbc7-logs\") pod \"glance-default-internal-api-0\" (UID: \"a3fd4804-b7ff-42d4-aa58-999cb59efbc7\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:01:37 crc kubenswrapper[4813]: I1201 09:01:37.679085 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zlg5m\" (UniqueName: \"kubernetes.io/projected/a3fd4804-b7ff-42d4-aa58-999cb59efbc7-kube-api-access-zlg5m\") pod \"glance-default-internal-api-0\" (UID: \"a3fd4804-b7ff-42d4-aa58-999cb59efbc7\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:01:37 crc kubenswrapper[4813]: I1201 09:01:37.679122 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/a3fd4804-b7ff-42d4-aa58-999cb59efbc7-etc-iscsi\") pod \"glance-default-internal-api-0\" (UID: \"a3fd4804-b7ff-42d4-aa58-999cb59efbc7\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:01:37 crc kubenswrapper[4813]: I1201 09:01:37.679152 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a3fd4804-b7ff-42d4-aa58-999cb59efbc7-config-data\") pod \"glance-default-internal-api-0\" (UID: \"a3fd4804-b7ff-42d4-aa58-999cb59efbc7\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:01:37 crc kubenswrapper[4813]: I1201 09:01:37.679178 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-internal-api-0\" (UID: \"a3fd4804-b7ff-42d4-aa58-999cb59efbc7\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:01:37 crc kubenswrapper[4813]: I1201 09:01:37.679246 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/a3fd4804-b7ff-42d4-aa58-999cb59efbc7-run\") pod \"glance-default-internal-api-0\" (UID: \"a3fd4804-b7ff-42d4-aa58-999cb59efbc7\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:01:37 crc kubenswrapper[4813]: I1201 09:01:37.679447 4813 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-internal-api-0\" (UID: \"a3fd4804-b7ff-42d4-aa58-999cb59efbc7\") device mount path \"/mnt/openstack/pv04\"" pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:01:37 crc kubenswrapper[4813]: I1201 09:01:37.705677 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-internal-api-0\" (UID: \"a3fd4804-b7ff-42d4-aa58-999cb59efbc7\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:01:37 crc kubenswrapper[4813]: I1201 09:01:37.780750 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/a3fd4804-b7ff-42d4-aa58-999cb59efbc7-run\") pod \"glance-default-internal-api-0\" (UID: \"a3fd4804-b7ff-42d4-aa58-999cb59efbc7\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:01:37 crc kubenswrapper[4813]: I1201 09:01:37.780805 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/a3fd4804-b7ff-42d4-aa58-999cb59efbc7-var-locks-brick\") pod \"glance-default-internal-api-0\" (UID: \"a3fd4804-b7ff-42d4-aa58-999cb59efbc7\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:01:37 crc kubenswrapper[4813]: I1201 09:01:37.780838 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/a3fd4804-b7ff-42d4-aa58-999cb59efbc7-lib-modules\") pod \"glance-default-internal-api-0\" (UID: \"a3fd4804-b7ff-42d4-aa58-999cb59efbc7\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:01:37 crc kubenswrapper[4813]: I1201 09:01:37.780860 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a3fd4804-b7ff-42d4-aa58-999cb59efbc7-scripts\") pod \"glance-default-internal-api-0\" (UID: \"a3fd4804-b7ff-42d4-aa58-999cb59efbc7\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:01:37 crc kubenswrapper[4813]: I1201 09:01:37.780874 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/a3fd4804-b7ff-42d4-aa58-999cb59efbc7-run\") pod \"glance-default-internal-api-0\" (UID: \"a3fd4804-b7ff-42d4-aa58-999cb59efbc7\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:01:37 crc kubenswrapper[4813]: I1201 09:01:37.780903 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/a3fd4804-b7ff-42d4-aa58-999cb59efbc7-etc-nvme\") pod \"glance-default-internal-api-0\" (UID: \"a3fd4804-b7ff-42d4-aa58-999cb59efbc7\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:01:37 crc kubenswrapper[4813]: I1201 09:01:37.780922 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/a3fd4804-b7ff-42d4-aa58-999cb59efbc7-lib-modules\") pod \"glance-default-internal-api-0\" (UID: \"a3fd4804-b7ff-42d4-aa58-999cb59efbc7\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:01:37 crc kubenswrapper[4813]: I1201 09:01:37.780927 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/a3fd4804-b7ff-42d4-aa58-999cb59efbc7-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"a3fd4804-b7ff-42d4-aa58-999cb59efbc7\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:01:37 crc kubenswrapper[4813]: I1201 09:01:37.780952 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/a3fd4804-b7ff-42d4-aa58-999cb59efbc7-dev\") pod \"glance-default-internal-api-0\" (UID: \"a3fd4804-b7ff-42d4-aa58-999cb59efbc7\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:01:37 crc kubenswrapper[4813]: I1201 09:01:37.780975 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/a3fd4804-b7ff-42d4-aa58-999cb59efbc7-etc-nvme\") pod \"glance-default-internal-api-0\" (UID: \"a3fd4804-b7ff-42d4-aa58-999cb59efbc7\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:01:37 crc kubenswrapper[4813]: I1201 09:01:37.780927 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/a3fd4804-b7ff-42d4-aa58-999cb59efbc7-var-locks-brick\") pod \"glance-default-internal-api-0\" (UID: \"a3fd4804-b7ff-42d4-aa58-999cb59efbc7\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:01:37 crc kubenswrapper[4813]: I1201 09:01:37.781024 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/a3fd4804-b7ff-42d4-aa58-999cb59efbc7-sys\") pod \"glance-default-internal-api-0\" (UID: \"a3fd4804-b7ff-42d4-aa58-999cb59efbc7\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:01:37 crc kubenswrapper[4813]: I1201 09:01:37.781065 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a3fd4804-b7ff-42d4-aa58-999cb59efbc7-logs\") pod \"glance-default-internal-api-0\" (UID: \"a3fd4804-b7ff-42d4-aa58-999cb59efbc7\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:01:37 crc kubenswrapper[4813]: I1201 09:01:37.781088 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/a3fd4804-b7ff-42d4-aa58-999cb59efbc7-dev\") pod \"glance-default-internal-api-0\" (UID: \"a3fd4804-b7ff-42d4-aa58-999cb59efbc7\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:01:37 crc kubenswrapper[4813]: I1201 09:01:37.781116 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zlg5m\" (UniqueName: \"kubernetes.io/projected/a3fd4804-b7ff-42d4-aa58-999cb59efbc7-kube-api-access-zlg5m\") pod \"glance-default-internal-api-0\" (UID: \"a3fd4804-b7ff-42d4-aa58-999cb59efbc7\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:01:37 crc kubenswrapper[4813]: I1201 09:01:37.781170 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/a3fd4804-b7ff-42d4-aa58-999cb59efbc7-etc-iscsi\") pod \"glance-default-internal-api-0\" (UID: \"a3fd4804-b7ff-42d4-aa58-999cb59efbc7\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:01:37 crc kubenswrapper[4813]: I1201 09:01:37.781182 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/a3fd4804-b7ff-42d4-aa58-999cb59efbc7-sys\") pod \"glance-default-internal-api-0\" (UID: \"a3fd4804-b7ff-42d4-aa58-999cb59efbc7\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:01:37 crc kubenswrapper[4813]: I1201 09:01:37.781200 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a3fd4804-b7ff-42d4-aa58-999cb59efbc7-config-data\") pod \"glance-default-internal-api-0\" (UID: \"a3fd4804-b7ff-42d4-aa58-999cb59efbc7\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:01:37 crc kubenswrapper[4813]: I1201 09:01:37.781275 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-internal-api-0\" (UID: \"a3fd4804-b7ff-42d4-aa58-999cb59efbc7\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:01:37 crc kubenswrapper[4813]: I1201 09:01:37.781472 4813 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-internal-api-0\" (UID: \"a3fd4804-b7ff-42d4-aa58-999cb59efbc7\") device mount path \"/mnt/openstack/pv03\"" pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:01:37 crc kubenswrapper[4813]: I1201 09:01:37.781472 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/a3fd4804-b7ff-42d4-aa58-999cb59efbc7-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"a3fd4804-b7ff-42d4-aa58-999cb59efbc7\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:01:37 crc kubenswrapper[4813]: I1201 09:01:37.781516 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a3fd4804-b7ff-42d4-aa58-999cb59efbc7-logs\") pod \"glance-default-internal-api-0\" (UID: \"a3fd4804-b7ff-42d4-aa58-999cb59efbc7\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:01:37 crc kubenswrapper[4813]: I1201 09:01:37.781522 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/a3fd4804-b7ff-42d4-aa58-999cb59efbc7-etc-iscsi\") pod \"glance-default-internal-api-0\" (UID: \"a3fd4804-b7ff-42d4-aa58-999cb59efbc7\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:01:37 crc kubenswrapper[4813]: I1201 09:01:37.784701 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a3fd4804-b7ff-42d4-aa58-999cb59efbc7-scripts\") pod \"glance-default-internal-api-0\" (UID: \"a3fd4804-b7ff-42d4-aa58-999cb59efbc7\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:01:37 crc kubenswrapper[4813]: I1201 09:01:37.786023 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a3fd4804-b7ff-42d4-aa58-999cb59efbc7-config-data\") pod \"glance-default-internal-api-0\" (UID: \"a3fd4804-b7ff-42d4-aa58-999cb59efbc7\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:01:37 crc kubenswrapper[4813]: I1201 09:01:37.808255 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zlg5m\" (UniqueName: \"kubernetes.io/projected/a3fd4804-b7ff-42d4-aa58-999cb59efbc7-kube-api-access-zlg5m\") pod \"glance-default-internal-api-0\" (UID: \"a3fd4804-b7ff-42d4-aa58-999cb59efbc7\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:01:37 crc kubenswrapper[4813]: I1201 09:01:37.810082 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-internal-api-0\" (UID: \"a3fd4804-b7ff-42d4-aa58-999cb59efbc7\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:01:37 crc kubenswrapper[4813]: I1201 09:01:37.868360 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:01:38 crc kubenswrapper[4813]: I1201 09:01:38.313275 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Dec 01 09:01:38 crc kubenswrapper[4813]: W1201 09:01:38.325480 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda3fd4804_b7ff_42d4_aa58_999cb59efbc7.slice/crio-dbd0a66eff1c44932b1223799123a9e56a244df412f20f7aa8661fe3c1c12d26 WatchSource:0}: Error finding container dbd0a66eff1c44932b1223799123a9e56a244df412f20f7aa8661fe3c1c12d26: Status 404 returned error can't find the container with id dbd0a66eff1c44932b1223799123a9e56a244df412f20f7aa8661fe3c1c12d26 Dec 01 09:01:38 crc kubenswrapper[4813]: I1201 09:01:38.404355 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="018e9026-a4ad-493c-9c2c-66d9ed3e36a5" path="/var/lib/kubelet/pods/018e9026-a4ad-493c-9c2c-66d9ed3e36a5/volumes" Dec 01 09:01:38 crc kubenswrapper[4813]: I1201 09:01:38.464661 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"a3fd4804-b7ff-42d4-aa58-999cb59efbc7","Type":"ContainerStarted","Data":"dbd0a66eff1c44932b1223799123a9e56a244df412f20f7aa8661fe3c1c12d26"} Dec 01 09:01:39 crc kubenswrapper[4813]: I1201 09:01:39.477678 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"a3fd4804-b7ff-42d4-aa58-999cb59efbc7","Type":"ContainerStarted","Data":"d89dec11e5c1011779b313ec517b6cceb51e641055e65c21caf627d7b47703b2"} Dec 01 09:01:39 crc kubenswrapper[4813]: I1201 09:01:39.478128 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"a3fd4804-b7ff-42d4-aa58-999cb59efbc7","Type":"ContainerStarted","Data":"6781e4dd9f6569955b120e287571b0574a25a313f8bbb4c49c326fc1bd841f7d"} Dec 01 09:01:39 crc kubenswrapper[4813]: I1201 09:01:39.510117 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-default-internal-api-0" podStartSLOduration=2.510087931 podStartE2EDuration="2.510087931s" podCreationTimestamp="2025-12-01 09:01:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 09:01:39.502861438 +0000 UTC m=+1284.125783024" watchObservedRunningTime="2025-12-01 09:01:39.510087931 +0000 UTC m=+1284.133009517" Dec 01 09:01:43 crc kubenswrapper[4813]: I1201 09:01:43.346009 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-external-api-0" Dec 01 09:01:43 crc kubenswrapper[4813]: I1201 09:01:43.346384 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-external-api-0" Dec 01 09:01:43 crc kubenswrapper[4813]: I1201 09:01:43.373623 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-external-api-0" Dec 01 09:01:43 crc kubenswrapper[4813]: I1201 09:01:43.392115 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-external-api-0" Dec 01 09:01:43 crc kubenswrapper[4813]: I1201 09:01:43.516467 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-external-api-0" Dec 01 09:01:43 crc kubenswrapper[4813]: I1201 09:01:43.516535 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-external-api-0" Dec 01 09:01:45 crc kubenswrapper[4813]: I1201 09:01:45.400560 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-external-api-0" Dec 01 09:01:45 crc kubenswrapper[4813]: I1201 09:01:45.403435 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-external-api-0" Dec 01 09:01:47 crc kubenswrapper[4813]: I1201 09:01:47.222023 4813 patch_prober.go:28] interesting pod/machine-config-daemon-9wn4q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 09:01:47 crc kubenswrapper[4813]: I1201 09:01:47.222401 4813 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9wn4q" podUID="3d423402-e08f-42e9-b60c-6eb9bf2f6d48" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 09:01:47 crc kubenswrapper[4813]: I1201 09:01:47.222478 4813 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-9wn4q" Dec 01 09:01:47 crc kubenswrapper[4813]: I1201 09:01:47.223511 4813 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"e5115101af56b8d932bd5a35078afaf20c64fdf9b230d78f43cbc4c834a68bf5"} pod="openshift-machine-config-operator/machine-config-daemon-9wn4q" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 01 09:01:47 crc kubenswrapper[4813]: I1201 09:01:47.223606 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-9wn4q" podUID="3d423402-e08f-42e9-b60c-6eb9bf2f6d48" containerName="machine-config-daemon" containerID="cri-o://e5115101af56b8d932bd5a35078afaf20c64fdf9b230d78f43cbc4c834a68bf5" gracePeriod=600 Dec 01 09:01:47 crc kubenswrapper[4813]: I1201 09:01:47.570910 4813 generic.go:334] "Generic (PLEG): container finished" podID="3d423402-e08f-42e9-b60c-6eb9bf2f6d48" containerID="e5115101af56b8d932bd5a35078afaf20c64fdf9b230d78f43cbc4c834a68bf5" exitCode=0 Dec 01 09:01:47 crc kubenswrapper[4813]: I1201 09:01:47.570982 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-9wn4q" event={"ID":"3d423402-e08f-42e9-b60c-6eb9bf2f6d48","Type":"ContainerDied","Data":"e5115101af56b8d932bd5a35078afaf20c64fdf9b230d78f43cbc4c834a68bf5"} Dec 01 09:01:47 crc kubenswrapper[4813]: I1201 09:01:47.571032 4813 scope.go:117] "RemoveContainer" containerID="f46458d1e4ce790f7980bd9d4f88fad10c8bcf27bf165b2631ed78f8dc68e1b3" Dec 01 09:01:47 crc kubenswrapper[4813]: I1201 09:01:47.869600 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:01:47 crc kubenswrapper[4813]: I1201 09:01:47.869739 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:01:47 crc kubenswrapper[4813]: I1201 09:01:47.894503 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:01:47 crc kubenswrapper[4813]: I1201 09:01:47.936310 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:01:48 crc kubenswrapper[4813]: I1201 09:01:48.582036 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-9wn4q" event={"ID":"3d423402-e08f-42e9-b60c-6eb9bf2f6d48","Type":"ContainerStarted","Data":"09b49517239f156b4ac50c72e4d184fde70e2daaf7b602adcfc3b42ec90a71e0"} Dec 01 09:01:48 crc kubenswrapper[4813]: I1201 09:01:48.582648 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:01:48 crc kubenswrapper[4813]: I1201 09:01:48.582687 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:01:50 crc kubenswrapper[4813]: I1201 09:01:50.551925 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:01:50 crc kubenswrapper[4813]: I1201 09:01:50.553932 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:01:52 crc kubenswrapper[4813]: I1201 09:01:52.621706 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-external-api-2"] Dec 01 09:01:52 crc kubenswrapper[4813]: I1201 09:01:52.624366 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-2" Dec 01 09:01:52 crc kubenswrapper[4813]: I1201 09:01:52.636341 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-external-api-1"] Dec 01 09:01:52 crc kubenswrapper[4813]: I1201 09:01:52.642246 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-1" Dec 01 09:01:52 crc kubenswrapper[4813]: I1201 09:01:52.663298 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-1"] Dec 01 09:01:52 crc kubenswrapper[4813]: I1201 09:01:52.682924 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-2"] Dec 01 09:01:52 crc kubenswrapper[4813]: I1201 09:01:52.731731 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-2"] Dec 01 09:01:52 crc kubenswrapper[4813]: I1201 09:01:52.732869 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-2" Dec 01 09:01:52 crc kubenswrapper[4813]: I1201 09:01:52.742350 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-1"] Dec 01 09:01:52 crc kubenswrapper[4813]: I1201 09:01:52.744228 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-1" Dec 01 09:01:52 crc kubenswrapper[4813]: I1201 09:01:52.746162 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/0bb179fc-fb0c-420a-a432-5a0d28884e23-var-locks-brick\") pod \"glance-default-external-api-2\" (UID: \"0bb179fc-fb0c-420a-a432-5a0d28884e23\") " pod="glance-kuttl-tests/glance-default-external-api-2" Dec 01 09:01:52 crc kubenswrapper[4813]: I1201 09:01:52.746212 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cdzkz\" (UniqueName: \"kubernetes.io/projected/0bb179fc-fb0c-420a-a432-5a0d28884e23-kube-api-access-cdzkz\") pod \"glance-default-external-api-2\" (UID: \"0bb179fc-fb0c-420a-a432-5a0d28884e23\") " pod="glance-kuttl-tests/glance-default-external-api-2" Dec 01 09:01:52 crc kubenswrapper[4813]: I1201 09:01:52.746249 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-external-api-2\" (UID: \"0bb179fc-fb0c-420a-a432-5a0d28884e23\") " pod="glance-kuttl-tests/glance-default-external-api-2" Dec 01 09:01:52 crc kubenswrapper[4813]: I1201 09:01:52.746271 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/0bb179fc-fb0c-420a-a432-5a0d28884e23-sys\") pod \"glance-default-external-api-2\" (UID: \"0bb179fc-fb0c-420a-a432-5a0d28884e23\") " pod="glance-kuttl-tests/glance-default-external-api-2" Dec 01 09:01:52 crc kubenswrapper[4813]: I1201 09:01:52.746310 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/0bb179fc-fb0c-420a-a432-5a0d28884e23-etc-iscsi\") pod \"glance-default-external-api-2\" (UID: \"0bb179fc-fb0c-420a-a432-5a0d28884e23\") " pod="glance-kuttl-tests/glance-default-external-api-2" Dec 01 09:01:52 crc kubenswrapper[4813]: I1201 09:01:52.746347 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/0bb179fc-fb0c-420a-a432-5a0d28884e23-lib-modules\") pod \"glance-default-external-api-2\" (UID: \"0bb179fc-fb0c-420a-a432-5a0d28884e23\") " pod="glance-kuttl-tests/glance-default-external-api-2" Dec 01 09:01:52 crc kubenswrapper[4813]: I1201 09:01:52.746371 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0bb179fc-fb0c-420a-a432-5a0d28884e23-logs\") pod \"glance-default-external-api-2\" (UID: \"0bb179fc-fb0c-420a-a432-5a0d28884e23\") " pod="glance-kuttl-tests/glance-default-external-api-2" Dec 01 09:01:52 crc kubenswrapper[4813]: I1201 09:01:52.746430 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0bb179fc-fb0c-420a-a432-5a0d28884e23-scripts\") pod \"glance-default-external-api-2\" (UID: \"0bb179fc-fb0c-420a-a432-5a0d28884e23\") " pod="glance-kuttl-tests/glance-default-external-api-2" Dec 01 09:01:52 crc kubenswrapper[4813]: I1201 09:01:52.746458 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/0bb179fc-fb0c-420a-a432-5a0d28884e23-dev\") pod \"glance-default-external-api-2\" (UID: \"0bb179fc-fb0c-420a-a432-5a0d28884e23\") " pod="glance-kuttl-tests/glance-default-external-api-2" Dec 01 09:01:52 crc kubenswrapper[4813]: I1201 09:01:52.746480 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/0bb179fc-fb0c-420a-a432-5a0d28884e23-httpd-run\") pod \"glance-default-external-api-2\" (UID: \"0bb179fc-fb0c-420a-a432-5a0d28884e23\") " pod="glance-kuttl-tests/glance-default-external-api-2" Dec 01 09:01:52 crc kubenswrapper[4813]: I1201 09:01:52.746510 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/0bb179fc-fb0c-420a-a432-5a0d28884e23-run\") pod \"glance-default-external-api-2\" (UID: \"0bb179fc-fb0c-420a-a432-5a0d28884e23\") " pod="glance-kuttl-tests/glance-default-external-api-2" Dec 01 09:01:52 crc kubenswrapper[4813]: I1201 09:01:52.746547 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0bb179fc-fb0c-420a-a432-5a0d28884e23-config-data\") pod \"glance-default-external-api-2\" (UID: \"0bb179fc-fb0c-420a-a432-5a0d28884e23\") " pod="glance-kuttl-tests/glance-default-external-api-2" Dec 01 09:01:52 crc kubenswrapper[4813]: I1201 09:01:52.746580 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-external-api-2\" (UID: \"0bb179fc-fb0c-420a-a432-5a0d28884e23\") " pod="glance-kuttl-tests/glance-default-external-api-2" Dec 01 09:01:52 crc kubenswrapper[4813]: I1201 09:01:52.746602 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/0bb179fc-fb0c-420a-a432-5a0d28884e23-etc-nvme\") pod \"glance-default-external-api-2\" (UID: \"0bb179fc-fb0c-420a-a432-5a0d28884e23\") " pod="glance-kuttl-tests/glance-default-external-api-2" Dec 01 09:01:52 crc kubenswrapper[4813]: I1201 09:01:52.758894 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-1"] Dec 01 09:01:52 crc kubenswrapper[4813]: I1201 09:01:52.767930 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-2"] Dec 01 09:01:52 crc kubenswrapper[4813]: I1201 09:01:52.847664 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/43861577-d77d-45ed-95ea-25a5dffa1b46-lib-modules\") pod \"glance-default-internal-api-2\" (UID: \"43861577-d77d-45ed-95ea-25a5dffa1b46\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Dec 01 09:01:52 crc kubenswrapper[4813]: I1201 09:01:52.847720 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/0bb179fc-fb0c-420a-a432-5a0d28884e23-run\") pod \"glance-default-external-api-2\" (UID: \"0bb179fc-fb0c-420a-a432-5a0d28884e23\") " pod="glance-kuttl-tests/glance-default-external-api-2" Dec 01 09:01:52 crc kubenswrapper[4813]: I1201 09:01:52.847744 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/43861577-d77d-45ed-95ea-25a5dffa1b46-config-data\") pod \"glance-default-internal-api-2\" (UID: \"43861577-d77d-45ed-95ea-25a5dffa1b46\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Dec 01 09:01:52 crc kubenswrapper[4813]: I1201 09:01:52.847766 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/7790813b-eccc-442b-a400-7fb5863599a7-dev\") pod \"glance-default-external-api-1\" (UID: \"7790813b-eccc-442b-a400-7fb5863599a7\") " pod="glance-kuttl-tests/glance-default-external-api-1" Dec 01 09:01:52 crc kubenswrapper[4813]: I1201 09:01:52.847784 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage20-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage20-crc\") pod \"glance-default-internal-api-1\" (UID: \"e58d9e76-7271-47c3-a7aa-a5a802bfdb41\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Dec 01 09:01:52 crc kubenswrapper[4813]: I1201 09:01:52.847804 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0bb179fc-fb0c-420a-a432-5a0d28884e23-config-data\") pod \"glance-default-external-api-2\" (UID: \"0bb179fc-fb0c-420a-a432-5a0d28884e23\") " pod="glance-kuttl-tests/glance-default-external-api-2" Dec 01 09:01:52 crc kubenswrapper[4813]: I1201 09:01:52.847823 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage14-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage14-crc\") pod \"glance-default-internal-api-1\" (UID: \"e58d9e76-7271-47c3-a7aa-a5a802bfdb41\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Dec 01 09:01:52 crc kubenswrapper[4813]: I1201 09:01:52.847840 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dhtk6\" (UniqueName: \"kubernetes.io/projected/43861577-d77d-45ed-95ea-25a5dffa1b46-kube-api-access-dhtk6\") pod \"glance-default-internal-api-2\" (UID: \"43861577-d77d-45ed-95ea-25a5dffa1b46\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Dec 01 09:01:52 crc kubenswrapper[4813]: I1201 09:01:52.847860 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/0bb179fc-fb0c-420a-a432-5a0d28884e23-run\") pod \"glance-default-external-api-2\" (UID: \"0bb179fc-fb0c-420a-a432-5a0d28884e23\") " pod="glance-kuttl-tests/glance-default-external-api-2" Dec 01 09:01:52 crc kubenswrapper[4813]: I1201 09:01:52.847934 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e58d9e76-7271-47c3-a7aa-a5a802bfdb41-config-data\") pod \"glance-default-internal-api-1\" (UID: \"e58d9e76-7271-47c3-a7aa-a5a802bfdb41\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Dec 01 09:01:52 crc kubenswrapper[4813]: I1201 09:01:52.848032 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-external-api-2\" (UID: \"0bb179fc-fb0c-420a-a432-5a0d28884e23\") " pod="glance-kuttl-tests/glance-default-external-api-2" Dec 01 09:01:52 crc kubenswrapper[4813]: I1201 09:01:52.848060 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/0bb179fc-fb0c-420a-a432-5a0d28884e23-etc-nvme\") pod \"glance-default-external-api-2\" (UID: \"0bb179fc-fb0c-420a-a432-5a0d28884e23\") " pod="glance-kuttl-tests/glance-default-external-api-2" Dec 01 09:01:52 crc kubenswrapper[4813]: I1201 09:01:52.848089 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/0bb179fc-fb0c-420a-a432-5a0d28884e23-var-locks-brick\") pod \"glance-default-external-api-2\" (UID: \"0bb179fc-fb0c-420a-a432-5a0d28884e23\") " pod="glance-kuttl-tests/glance-default-external-api-2" Dec 01 09:01:52 crc kubenswrapper[4813]: I1201 09:01:52.848136 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/e58d9e76-7271-47c3-a7aa-a5a802bfdb41-run\") pod \"glance-default-internal-api-1\" (UID: \"e58d9e76-7271-47c3-a7aa-a5a802bfdb41\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Dec 01 09:01:52 crc kubenswrapper[4813]: I1201 09:01:52.848158 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/43861577-d77d-45ed-95ea-25a5dffa1b46-etc-iscsi\") pod \"glance-default-internal-api-2\" (UID: \"43861577-d77d-45ed-95ea-25a5dffa1b46\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Dec 01 09:01:52 crc kubenswrapper[4813]: I1201 09:01:52.848173 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/0bb179fc-fb0c-420a-a432-5a0d28884e23-etc-nvme\") pod \"glance-default-external-api-2\" (UID: \"0bb179fc-fb0c-420a-a432-5a0d28884e23\") " pod="glance-kuttl-tests/glance-default-external-api-2" Dec 01 09:01:52 crc kubenswrapper[4813]: I1201 09:01:52.848204 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cdzkz\" (UniqueName: \"kubernetes.io/projected/0bb179fc-fb0c-420a-a432-5a0d28884e23-kube-api-access-cdzkz\") pod \"glance-default-external-api-2\" (UID: \"0bb179fc-fb0c-420a-a432-5a0d28884e23\") " pod="glance-kuttl-tests/glance-default-external-api-2" Dec 01 09:01:52 crc kubenswrapper[4813]: I1201 09:01:52.848229 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/0bb179fc-fb0c-420a-a432-5a0d28884e23-var-locks-brick\") pod \"glance-default-external-api-2\" (UID: \"0bb179fc-fb0c-420a-a432-5a0d28884e23\") " pod="glance-kuttl-tests/glance-default-external-api-2" Dec 01 09:01:52 crc kubenswrapper[4813]: I1201 09:01:52.848233 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/7790813b-eccc-442b-a400-7fb5863599a7-run\") pod \"glance-default-external-api-1\" (UID: \"7790813b-eccc-442b-a400-7fb5863599a7\") " pod="glance-kuttl-tests/glance-default-external-api-1" Dec 01 09:01:52 crc kubenswrapper[4813]: I1201 09:01:52.848268 4813 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-external-api-2\" (UID: \"0bb179fc-fb0c-420a-a432-5a0d28884e23\") device mount path \"/mnt/openstack/pv01\"" pod="glance-kuttl-tests/glance-default-external-api-2" Dec 01 09:01:52 crc kubenswrapper[4813]: I1201 09:01:52.848296 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/43861577-d77d-45ed-95ea-25a5dffa1b46-logs\") pod \"glance-default-internal-api-2\" (UID: \"43861577-d77d-45ed-95ea-25a5dffa1b46\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Dec 01 09:01:52 crc kubenswrapper[4813]: I1201 09:01:52.848339 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e58d9e76-7271-47c3-a7aa-a5a802bfdb41-scripts\") pod \"glance-default-internal-api-1\" (UID: \"e58d9e76-7271-47c3-a7aa-a5a802bfdb41\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Dec 01 09:01:52 crc kubenswrapper[4813]: I1201 09:01:52.848369 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-external-api-2\" (UID: \"0bb179fc-fb0c-420a-a432-5a0d28884e23\") " pod="glance-kuttl-tests/glance-default-external-api-2" Dec 01 09:01:52 crc kubenswrapper[4813]: I1201 09:01:52.848396 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/0bb179fc-fb0c-420a-a432-5a0d28884e23-sys\") pod \"glance-default-external-api-2\" (UID: \"0bb179fc-fb0c-420a-a432-5a0d28884e23\") " pod="glance-kuttl-tests/glance-default-external-api-2" Dec 01 09:01:52 crc kubenswrapper[4813]: I1201 09:01:52.848424 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/e58d9e76-7271-47c3-a7aa-a5a802bfdb41-etc-nvme\") pod \"glance-default-internal-api-1\" (UID: \"e58d9e76-7271-47c3-a7aa-a5a802bfdb41\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Dec 01 09:01:52 crc kubenswrapper[4813]: I1201 09:01:52.848458 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e58d9e76-7271-47c3-a7aa-a5a802bfdb41-httpd-run\") pod \"glance-default-internal-api-1\" (UID: \"e58d9e76-7271-47c3-a7aa-a5a802bfdb41\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Dec 01 09:01:52 crc kubenswrapper[4813]: I1201 09:01:52.848490 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/43861577-d77d-45ed-95ea-25a5dffa1b46-var-locks-brick\") pod \"glance-default-internal-api-2\" (UID: \"43861577-d77d-45ed-95ea-25a5dffa1b46\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Dec 01 09:01:52 crc kubenswrapper[4813]: I1201 09:01:52.848534 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/7790813b-eccc-442b-a400-7fb5863599a7-lib-modules\") pod \"glance-default-external-api-1\" (UID: \"7790813b-eccc-442b-a400-7fb5863599a7\") " pod="glance-kuttl-tests/glance-default-external-api-1" Dec 01 09:01:52 crc kubenswrapper[4813]: I1201 09:01:52.848556 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/0bb179fc-fb0c-420a-a432-5a0d28884e23-etc-iscsi\") pod \"glance-default-external-api-2\" (UID: \"0bb179fc-fb0c-420a-a432-5a0d28884e23\") " pod="glance-kuttl-tests/glance-default-external-api-2" Dec 01 09:01:52 crc kubenswrapper[4813]: I1201 09:01:52.848570 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/0bb179fc-fb0c-420a-a432-5a0d28884e23-sys\") pod \"glance-default-external-api-2\" (UID: \"0bb179fc-fb0c-420a-a432-5a0d28884e23\") " pod="glance-kuttl-tests/glance-default-external-api-2" Dec 01 09:01:52 crc kubenswrapper[4813]: I1201 09:01:52.848581 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/7790813b-eccc-442b-a400-7fb5863599a7-var-locks-brick\") pod \"glance-default-external-api-1\" (UID: \"7790813b-eccc-442b-a400-7fb5863599a7\") " pod="glance-kuttl-tests/glance-default-external-api-1" Dec 01 09:01:52 crc kubenswrapper[4813]: I1201 09:01:52.848533 4813 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-external-api-2\" (UID: \"0bb179fc-fb0c-420a-a432-5a0d28884e23\") device mount path \"/mnt/openstack/pv05\"" pod="glance-kuttl-tests/glance-default-external-api-2" Dec 01 09:01:52 crc kubenswrapper[4813]: I1201 09:01:52.848622 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/0bb179fc-fb0c-420a-a432-5a0d28884e23-etc-iscsi\") pod \"glance-default-external-api-2\" (UID: \"0bb179fc-fb0c-420a-a432-5a0d28884e23\") " pod="glance-kuttl-tests/glance-default-external-api-2" Dec 01 09:01:52 crc kubenswrapper[4813]: I1201 09:01:52.848687 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/e58d9e76-7271-47c3-a7aa-a5a802bfdb41-etc-iscsi\") pod \"glance-default-internal-api-1\" (UID: \"e58d9e76-7271-47c3-a7aa-a5a802bfdb41\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Dec 01 09:01:52 crc kubenswrapper[4813]: I1201 09:01:52.848707 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-2\" (UID: \"43861577-d77d-45ed-95ea-25a5dffa1b46\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Dec 01 09:01:52 crc kubenswrapper[4813]: I1201 09:01:52.848723 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7790813b-eccc-442b-a400-7fb5863599a7-logs\") pod \"glance-default-external-api-1\" (UID: \"7790813b-eccc-442b-a400-7fb5863599a7\") " pod="glance-kuttl-tests/glance-default-external-api-1" Dec 01 09:01:52 crc kubenswrapper[4813]: I1201 09:01:52.848744 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/0bb179fc-fb0c-420a-a432-5a0d28884e23-lib-modules\") pod \"glance-default-external-api-2\" (UID: \"0bb179fc-fb0c-420a-a432-5a0d28884e23\") " pod="glance-kuttl-tests/glance-default-external-api-2" Dec 01 09:01:52 crc kubenswrapper[4813]: I1201 09:01:52.848760 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/43861577-d77d-45ed-95ea-25a5dffa1b46-run\") pod \"glance-default-internal-api-2\" (UID: \"43861577-d77d-45ed-95ea-25a5dffa1b46\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Dec 01 09:01:52 crc kubenswrapper[4813]: I1201 09:01:52.848787 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/43861577-d77d-45ed-95ea-25a5dffa1b46-httpd-run\") pod \"glance-default-internal-api-2\" (UID: \"43861577-d77d-45ed-95ea-25a5dffa1b46\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Dec 01 09:01:52 crc kubenswrapper[4813]: I1201 09:01:52.848801 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/e58d9e76-7271-47c3-a7aa-a5a802bfdb41-sys\") pod \"glance-default-internal-api-1\" (UID: \"e58d9e76-7271-47c3-a7aa-a5a802bfdb41\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Dec 01 09:01:52 crc kubenswrapper[4813]: I1201 09:01:52.848824 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0bb179fc-fb0c-420a-a432-5a0d28884e23-logs\") pod \"glance-default-external-api-2\" (UID: \"0bb179fc-fb0c-420a-a432-5a0d28884e23\") " pod="glance-kuttl-tests/glance-default-external-api-2" Dec 01 09:01:52 crc kubenswrapper[4813]: I1201 09:01:52.848845 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/e58d9e76-7271-47c3-a7aa-a5a802bfdb41-var-locks-brick\") pod \"glance-default-internal-api-1\" (UID: \"e58d9e76-7271-47c3-a7aa-a5a802bfdb41\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Dec 01 09:01:52 crc kubenswrapper[4813]: I1201 09:01:52.848861 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e58d9e76-7271-47c3-a7aa-a5a802bfdb41-logs\") pod \"glance-default-internal-api-1\" (UID: \"e58d9e76-7271-47c3-a7aa-a5a802bfdb41\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Dec 01 09:01:52 crc kubenswrapper[4813]: I1201 09:01:52.848893 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/7790813b-eccc-442b-a400-7fb5863599a7-etc-nvme\") pod \"glance-default-external-api-1\" (UID: \"7790813b-eccc-442b-a400-7fb5863599a7\") " pod="glance-kuttl-tests/glance-default-external-api-1" Dec 01 09:01:52 crc kubenswrapper[4813]: I1201 09:01:52.848917 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tkp6h\" (UniqueName: \"kubernetes.io/projected/7790813b-eccc-442b-a400-7fb5863599a7-kube-api-access-tkp6h\") pod \"glance-default-external-api-1\" (UID: \"7790813b-eccc-442b-a400-7fb5863599a7\") " pod="glance-kuttl-tests/glance-default-external-api-1" Dec 01 09:01:52 crc kubenswrapper[4813]: I1201 09:01:52.848941 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/43861577-d77d-45ed-95ea-25a5dffa1b46-scripts\") pod \"glance-default-internal-api-2\" (UID: \"43861577-d77d-45ed-95ea-25a5dffa1b46\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Dec 01 09:01:52 crc kubenswrapper[4813]: I1201 09:01:52.848977 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-njccq\" (UniqueName: \"kubernetes.io/projected/e58d9e76-7271-47c3-a7aa-a5a802bfdb41-kube-api-access-njccq\") pod \"glance-default-internal-api-1\" (UID: \"e58d9e76-7271-47c3-a7aa-a5a802bfdb41\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Dec 01 09:01:52 crc kubenswrapper[4813]: I1201 09:01:52.848995 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/43861577-d77d-45ed-95ea-25a5dffa1b46-etc-nvme\") pod \"glance-default-internal-api-2\" (UID: \"43861577-d77d-45ed-95ea-25a5dffa1b46\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Dec 01 09:01:52 crc kubenswrapper[4813]: I1201 09:01:52.849009 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7790813b-eccc-442b-a400-7fb5863599a7-httpd-run\") pod \"glance-default-external-api-1\" (UID: \"7790813b-eccc-442b-a400-7fb5863599a7\") " pod="glance-kuttl-tests/glance-default-external-api-1" Dec 01 09:01:52 crc kubenswrapper[4813]: I1201 09:01:52.849028 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage17-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage17-crc\") pod \"glance-default-external-api-1\" (UID: \"7790813b-eccc-442b-a400-7fb5863599a7\") " pod="glance-kuttl-tests/glance-default-external-api-1" Dec 01 09:01:52 crc kubenswrapper[4813]: I1201 09:01:52.849043 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/43861577-d77d-45ed-95ea-25a5dffa1b46-sys\") pod \"glance-default-internal-api-2\" (UID: \"43861577-d77d-45ed-95ea-25a5dffa1b46\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Dec 01 09:01:52 crc kubenswrapper[4813]: I1201 09:01:52.849059 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/43861577-d77d-45ed-95ea-25a5dffa1b46-dev\") pod \"glance-default-internal-api-2\" (UID: \"43861577-d77d-45ed-95ea-25a5dffa1b46\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Dec 01 09:01:52 crc kubenswrapper[4813]: I1201 09:01:52.849081 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7790813b-eccc-442b-a400-7fb5863599a7-config-data\") pod \"glance-default-external-api-1\" (UID: \"7790813b-eccc-442b-a400-7fb5863599a7\") " pod="glance-kuttl-tests/glance-default-external-api-1" Dec 01 09:01:52 crc kubenswrapper[4813]: I1201 09:01:52.849108 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-external-api-1\" (UID: \"7790813b-eccc-442b-a400-7fb5863599a7\") " pod="glance-kuttl-tests/glance-default-external-api-1" Dec 01 09:01:52 crc kubenswrapper[4813]: I1201 09:01:52.849175 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/e58d9e76-7271-47c3-a7aa-a5a802bfdb41-lib-modules\") pod \"glance-default-internal-api-1\" (UID: \"e58d9e76-7271-47c3-a7aa-a5a802bfdb41\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Dec 01 09:01:52 crc kubenswrapper[4813]: I1201 09:01:52.849194 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0bb179fc-fb0c-420a-a432-5a0d28884e23-scripts\") pod \"glance-default-external-api-2\" (UID: \"0bb179fc-fb0c-420a-a432-5a0d28884e23\") " pod="glance-kuttl-tests/glance-default-external-api-2" Dec 01 09:01:52 crc kubenswrapper[4813]: I1201 09:01:52.849214 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage13-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage13-crc\") pod \"glance-default-internal-api-2\" (UID: \"43861577-d77d-45ed-95ea-25a5dffa1b46\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Dec 01 09:01:52 crc kubenswrapper[4813]: I1201 09:01:52.849233 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/7790813b-eccc-442b-a400-7fb5863599a7-sys\") pod \"glance-default-external-api-1\" (UID: \"7790813b-eccc-442b-a400-7fb5863599a7\") " pod="glance-kuttl-tests/glance-default-external-api-1" Dec 01 09:01:52 crc kubenswrapper[4813]: I1201 09:01:52.849260 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/7790813b-eccc-442b-a400-7fb5863599a7-etc-iscsi\") pod \"glance-default-external-api-1\" (UID: \"7790813b-eccc-442b-a400-7fb5863599a7\") " pod="glance-kuttl-tests/glance-default-external-api-1" Dec 01 09:01:52 crc kubenswrapper[4813]: I1201 09:01:52.849462 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/0bb179fc-fb0c-420a-a432-5a0d28884e23-lib-modules\") pod \"glance-default-external-api-2\" (UID: \"0bb179fc-fb0c-420a-a432-5a0d28884e23\") " pod="glance-kuttl-tests/glance-default-external-api-2" Dec 01 09:01:52 crc kubenswrapper[4813]: I1201 09:01:52.849559 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/0bb179fc-fb0c-420a-a432-5a0d28884e23-dev\") pod \"glance-default-external-api-2\" (UID: \"0bb179fc-fb0c-420a-a432-5a0d28884e23\") " pod="glance-kuttl-tests/glance-default-external-api-2" Dec 01 09:01:52 crc kubenswrapper[4813]: I1201 09:01:52.849619 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/e58d9e76-7271-47c3-a7aa-a5a802bfdb41-dev\") pod \"glance-default-internal-api-1\" (UID: \"e58d9e76-7271-47c3-a7aa-a5a802bfdb41\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Dec 01 09:01:52 crc kubenswrapper[4813]: I1201 09:01:52.850134 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/0bb179fc-fb0c-420a-a432-5a0d28884e23-dev\") pod \"glance-default-external-api-2\" (UID: \"0bb179fc-fb0c-420a-a432-5a0d28884e23\") " pod="glance-kuttl-tests/glance-default-external-api-2" Dec 01 09:01:52 crc kubenswrapper[4813]: I1201 09:01:52.850190 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0bb179fc-fb0c-420a-a432-5a0d28884e23-logs\") pod \"glance-default-external-api-2\" (UID: \"0bb179fc-fb0c-420a-a432-5a0d28884e23\") " pod="glance-kuttl-tests/glance-default-external-api-2" Dec 01 09:01:52 crc kubenswrapper[4813]: I1201 09:01:52.850212 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/0bb179fc-fb0c-420a-a432-5a0d28884e23-httpd-run\") pod \"glance-default-external-api-2\" (UID: \"0bb179fc-fb0c-420a-a432-5a0d28884e23\") " pod="glance-kuttl-tests/glance-default-external-api-2" Dec 01 09:01:52 crc kubenswrapper[4813]: I1201 09:01:52.850346 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7790813b-eccc-442b-a400-7fb5863599a7-scripts\") pod \"glance-default-external-api-1\" (UID: \"7790813b-eccc-442b-a400-7fb5863599a7\") " pod="glance-kuttl-tests/glance-default-external-api-1" Dec 01 09:01:52 crc kubenswrapper[4813]: I1201 09:01:52.850530 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/0bb179fc-fb0c-420a-a432-5a0d28884e23-httpd-run\") pod \"glance-default-external-api-2\" (UID: \"0bb179fc-fb0c-420a-a432-5a0d28884e23\") " pod="glance-kuttl-tests/glance-default-external-api-2" Dec 01 09:01:52 crc kubenswrapper[4813]: I1201 09:01:52.855514 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0bb179fc-fb0c-420a-a432-5a0d28884e23-config-data\") pod \"glance-default-external-api-2\" (UID: \"0bb179fc-fb0c-420a-a432-5a0d28884e23\") " pod="glance-kuttl-tests/glance-default-external-api-2" Dec 01 09:01:52 crc kubenswrapper[4813]: I1201 09:01:52.856952 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0bb179fc-fb0c-420a-a432-5a0d28884e23-scripts\") pod \"glance-default-external-api-2\" (UID: \"0bb179fc-fb0c-420a-a432-5a0d28884e23\") " pod="glance-kuttl-tests/glance-default-external-api-2" Dec 01 09:01:52 crc kubenswrapper[4813]: I1201 09:01:52.868252 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-external-api-2\" (UID: \"0bb179fc-fb0c-420a-a432-5a0d28884e23\") " pod="glance-kuttl-tests/glance-default-external-api-2" Dec 01 09:01:52 crc kubenswrapper[4813]: I1201 09:01:52.871047 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cdzkz\" (UniqueName: \"kubernetes.io/projected/0bb179fc-fb0c-420a-a432-5a0d28884e23-kube-api-access-cdzkz\") pod \"glance-default-external-api-2\" (UID: \"0bb179fc-fb0c-420a-a432-5a0d28884e23\") " pod="glance-kuttl-tests/glance-default-external-api-2" Dec 01 09:01:52 crc kubenswrapper[4813]: I1201 09:01:52.872317 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-external-api-2\" (UID: \"0bb179fc-fb0c-420a-a432-5a0d28884e23\") " pod="glance-kuttl-tests/glance-default-external-api-2" Dec 01 09:01:52 crc kubenswrapper[4813]: I1201 09:01:52.952912 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/43861577-d77d-45ed-95ea-25a5dffa1b46-var-locks-brick\") pod \"glance-default-internal-api-2\" (UID: \"43861577-d77d-45ed-95ea-25a5dffa1b46\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Dec 01 09:01:52 crc kubenswrapper[4813]: I1201 09:01:52.952956 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/7790813b-eccc-442b-a400-7fb5863599a7-lib-modules\") pod \"glance-default-external-api-1\" (UID: \"7790813b-eccc-442b-a400-7fb5863599a7\") " pod="glance-kuttl-tests/glance-default-external-api-1" Dec 01 09:01:52 crc kubenswrapper[4813]: I1201 09:01:52.952996 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/7790813b-eccc-442b-a400-7fb5863599a7-var-locks-brick\") pod \"glance-default-external-api-1\" (UID: \"7790813b-eccc-442b-a400-7fb5863599a7\") " pod="glance-kuttl-tests/glance-default-external-api-1" Dec 01 09:01:52 crc kubenswrapper[4813]: I1201 09:01:52.953029 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-2\" (UID: \"43861577-d77d-45ed-95ea-25a5dffa1b46\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Dec 01 09:01:52 crc kubenswrapper[4813]: I1201 09:01:52.953051 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/e58d9e76-7271-47c3-a7aa-a5a802bfdb41-etc-iscsi\") pod \"glance-default-internal-api-1\" (UID: \"e58d9e76-7271-47c3-a7aa-a5a802bfdb41\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Dec 01 09:01:52 crc kubenswrapper[4813]: I1201 09:01:52.953048 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/43861577-d77d-45ed-95ea-25a5dffa1b46-var-locks-brick\") pod \"glance-default-internal-api-2\" (UID: \"43861577-d77d-45ed-95ea-25a5dffa1b46\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Dec 01 09:01:52 crc kubenswrapper[4813]: I1201 09:01:52.953131 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/7790813b-eccc-442b-a400-7fb5863599a7-var-locks-brick\") pod \"glance-default-external-api-1\" (UID: \"7790813b-eccc-442b-a400-7fb5863599a7\") " pod="glance-kuttl-tests/glance-default-external-api-1" Dec 01 09:01:52 crc kubenswrapper[4813]: I1201 09:01:52.953067 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7790813b-eccc-442b-a400-7fb5863599a7-logs\") pod \"glance-default-external-api-1\" (UID: \"7790813b-eccc-442b-a400-7fb5863599a7\") " pod="glance-kuttl-tests/glance-default-external-api-1" Dec 01 09:01:52 crc kubenswrapper[4813]: I1201 09:01:52.953179 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/e58d9e76-7271-47c3-a7aa-a5a802bfdb41-etc-iscsi\") pod \"glance-default-internal-api-1\" (UID: \"e58d9e76-7271-47c3-a7aa-a5a802bfdb41\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Dec 01 09:01:52 crc kubenswrapper[4813]: I1201 09:01:52.953200 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/43861577-d77d-45ed-95ea-25a5dffa1b46-run\") pod \"glance-default-internal-api-2\" (UID: \"43861577-d77d-45ed-95ea-25a5dffa1b46\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Dec 01 09:01:52 crc kubenswrapper[4813]: I1201 09:01:52.953240 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/e58d9e76-7271-47c3-a7aa-a5a802bfdb41-sys\") pod \"glance-default-internal-api-1\" (UID: \"e58d9e76-7271-47c3-a7aa-a5a802bfdb41\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Dec 01 09:01:52 crc kubenswrapper[4813]: I1201 09:01:52.953280 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/43861577-d77d-45ed-95ea-25a5dffa1b46-httpd-run\") pod \"glance-default-internal-api-2\" (UID: \"43861577-d77d-45ed-95ea-25a5dffa1b46\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Dec 01 09:01:52 crc kubenswrapper[4813]: I1201 09:01:52.953241 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/7790813b-eccc-442b-a400-7fb5863599a7-lib-modules\") pod \"glance-default-external-api-1\" (UID: \"7790813b-eccc-442b-a400-7fb5863599a7\") " pod="glance-kuttl-tests/glance-default-external-api-1" Dec 01 09:01:52 crc kubenswrapper[4813]: I1201 09:01:52.953321 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/e58d9e76-7271-47c3-a7aa-a5a802bfdb41-var-locks-brick\") pod \"glance-default-internal-api-1\" (UID: \"e58d9e76-7271-47c3-a7aa-a5a802bfdb41\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Dec 01 09:01:52 crc kubenswrapper[4813]: I1201 09:01:52.953334 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/e58d9e76-7271-47c3-a7aa-a5a802bfdb41-sys\") pod \"glance-default-internal-api-1\" (UID: \"e58d9e76-7271-47c3-a7aa-a5a802bfdb41\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Dec 01 09:01:52 crc kubenswrapper[4813]: I1201 09:01:52.953352 4813 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-2\" (UID: \"43861577-d77d-45ed-95ea-25a5dffa1b46\") device mount path \"/mnt/openstack/pv12\"" pod="glance-kuttl-tests/glance-default-internal-api-2" Dec 01 09:01:52 crc kubenswrapper[4813]: I1201 09:01:52.953357 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e58d9e76-7271-47c3-a7aa-a5a802bfdb41-logs\") pod \"glance-default-internal-api-1\" (UID: \"e58d9e76-7271-47c3-a7aa-a5a802bfdb41\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Dec 01 09:01:52 crc kubenswrapper[4813]: I1201 09:01:52.953393 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/e58d9e76-7271-47c3-a7aa-a5a802bfdb41-var-locks-brick\") pod \"glance-default-internal-api-1\" (UID: \"e58d9e76-7271-47c3-a7aa-a5a802bfdb41\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Dec 01 09:01:52 crc kubenswrapper[4813]: I1201 09:01:52.953409 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7790813b-eccc-442b-a400-7fb5863599a7-logs\") pod \"glance-default-external-api-1\" (UID: \"7790813b-eccc-442b-a400-7fb5863599a7\") " pod="glance-kuttl-tests/glance-default-external-api-1" Dec 01 09:01:52 crc kubenswrapper[4813]: I1201 09:01:52.953454 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/7790813b-eccc-442b-a400-7fb5863599a7-etc-nvme\") pod \"glance-default-external-api-1\" (UID: \"7790813b-eccc-442b-a400-7fb5863599a7\") " pod="glance-kuttl-tests/glance-default-external-api-1" Dec 01 09:01:52 crc kubenswrapper[4813]: I1201 09:01:52.953502 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tkp6h\" (UniqueName: \"kubernetes.io/projected/7790813b-eccc-442b-a400-7fb5863599a7-kube-api-access-tkp6h\") pod \"glance-default-external-api-1\" (UID: \"7790813b-eccc-442b-a400-7fb5863599a7\") " pod="glance-kuttl-tests/glance-default-external-api-1" Dec 01 09:01:52 crc kubenswrapper[4813]: I1201 09:01:52.953534 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/43861577-d77d-45ed-95ea-25a5dffa1b46-scripts\") pod \"glance-default-internal-api-2\" (UID: \"43861577-d77d-45ed-95ea-25a5dffa1b46\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Dec 01 09:01:52 crc kubenswrapper[4813]: I1201 09:01:52.953756 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-njccq\" (UniqueName: \"kubernetes.io/projected/e58d9e76-7271-47c3-a7aa-a5a802bfdb41-kube-api-access-njccq\") pod \"glance-default-internal-api-1\" (UID: \"e58d9e76-7271-47c3-a7aa-a5a802bfdb41\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Dec 01 09:01:52 crc kubenswrapper[4813]: I1201 09:01:52.953776 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/7790813b-eccc-442b-a400-7fb5863599a7-etc-nvme\") pod \"glance-default-external-api-1\" (UID: \"7790813b-eccc-442b-a400-7fb5863599a7\") " pod="glance-kuttl-tests/glance-default-external-api-1" Dec 01 09:01:52 crc kubenswrapper[4813]: I1201 09:01:52.953819 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/43861577-d77d-45ed-95ea-25a5dffa1b46-etc-nvme\") pod \"glance-default-internal-api-2\" (UID: \"43861577-d77d-45ed-95ea-25a5dffa1b46\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Dec 01 09:01:52 crc kubenswrapper[4813]: I1201 09:01:52.953779 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/43861577-d77d-45ed-95ea-25a5dffa1b46-etc-nvme\") pod \"glance-default-internal-api-2\" (UID: \"43861577-d77d-45ed-95ea-25a5dffa1b46\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Dec 01 09:01:52 crc kubenswrapper[4813]: I1201 09:01:52.953853 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7790813b-eccc-442b-a400-7fb5863599a7-httpd-run\") pod \"glance-default-external-api-1\" (UID: \"7790813b-eccc-442b-a400-7fb5863599a7\") " pod="glance-kuttl-tests/glance-default-external-api-1" Dec 01 09:01:52 crc kubenswrapper[4813]: I1201 09:01:52.953878 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage17-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage17-crc\") pod \"glance-default-external-api-1\" (UID: \"7790813b-eccc-442b-a400-7fb5863599a7\") " pod="glance-kuttl-tests/glance-default-external-api-1" Dec 01 09:01:52 crc kubenswrapper[4813]: I1201 09:01:52.953902 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/43861577-d77d-45ed-95ea-25a5dffa1b46-sys\") pod \"glance-default-internal-api-2\" (UID: \"43861577-d77d-45ed-95ea-25a5dffa1b46\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Dec 01 09:01:52 crc kubenswrapper[4813]: I1201 09:01:52.953922 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/43861577-d77d-45ed-95ea-25a5dffa1b46-dev\") pod \"glance-default-internal-api-2\" (UID: \"43861577-d77d-45ed-95ea-25a5dffa1b46\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Dec 01 09:01:52 crc kubenswrapper[4813]: I1201 09:01:52.953947 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7790813b-eccc-442b-a400-7fb5863599a7-config-data\") pod \"glance-default-external-api-1\" (UID: \"7790813b-eccc-442b-a400-7fb5863599a7\") " pod="glance-kuttl-tests/glance-default-external-api-1" Dec 01 09:01:52 crc kubenswrapper[4813]: I1201 09:01:52.954006 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-external-api-1\" (UID: \"7790813b-eccc-442b-a400-7fb5863599a7\") " pod="glance-kuttl-tests/glance-default-external-api-1" Dec 01 09:01:52 crc kubenswrapper[4813]: I1201 09:01:52.954054 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/e58d9e76-7271-47c3-a7aa-a5a802bfdb41-lib-modules\") pod \"glance-default-internal-api-1\" (UID: \"e58d9e76-7271-47c3-a7aa-a5a802bfdb41\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Dec 01 09:01:52 crc kubenswrapper[4813]: I1201 09:01:52.954080 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/7790813b-eccc-442b-a400-7fb5863599a7-sys\") pod \"glance-default-external-api-1\" (UID: \"7790813b-eccc-442b-a400-7fb5863599a7\") " pod="glance-kuttl-tests/glance-default-external-api-1" Dec 01 09:01:52 crc kubenswrapper[4813]: I1201 09:01:52.954097 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e58d9e76-7271-47c3-a7aa-a5a802bfdb41-logs\") pod \"glance-default-internal-api-1\" (UID: \"e58d9e76-7271-47c3-a7aa-a5a802bfdb41\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Dec 01 09:01:52 crc kubenswrapper[4813]: I1201 09:01:52.954100 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage13-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage13-crc\") pod \"glance-default-internal-api-2\" (UID: \"43861577-d77d-45ed-95ea-25a5dffa1b46\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Dec 01 09:01:52 crc kubenswrapper[4813]: I1201 09:01:52.954161 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/7790813b-eccc-442b-a400-7fb5863599a7-etc-iscsi\") pod \"glance-default-external-api-1\" (UID: \"7790813b-eccc-442b-a400-7fb5863599a7\") " pod="glance-kuttl-tests/glance-default-external-api-1" Dec 01 09:01:52 crc kubenswrapper[4813]: I1201 09:01:52.954186 4813 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage13-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage13-crc\") pod \"glance-default-internal-api-2\" (UID: \"43861577-d77d-45ed-95ea-25a5dffa1b46\") device mount path \"/mnt/openstack/pv13\"" pod="glance-kuttl-tests/glance-default-internal-api-2" Dec 01 09:01:52 crc kubenswrapper[4813]: I1201 09:01:52.954225 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/e58d9e76-7271-47c3-a7aa-a5a802bfdb41-dev\") pod \"glance-default-internal-api-1\" (UID: \"e58d9e76-7271-47c3-a7aa-a5a802bfdb41\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Dec 01 09:01:52 crc kubenswrapper[4813]: I1201 09:01:52.954303 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/7790813b-eccc-442b-a400-7fb5863599a7-etc-iscsi\") pod \"glance-default-external-api-1\" (UID: \"7790813b-eccc-442b-a400-7fb5863599a7\") " pod="glance-kuttl-tests/glance-default-external-api-1" Dec 01 09:01:52 crc kubenswrapper[4813]: I1201 09:01:52.954339 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/e58d9e76-7271-47c3-a7aa-a5a802bfdb41-lib-modules\") pod \"glance-default-internal-api-1\" (UID: \"e58d9e76-7271-47c3-a7aa-a5a802bfdb41\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Dec 01 09:01:52 crc kubenswrapper[4813]: I1201 09:01:52.954471 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7790813b-eccc-442b-a400-7fb5863599a7-httpd-run\") pod \"glance-default-external-api-1\" (UID: \"7790813b-eccc-442b-a400-7fb5863599a7\") " pod="glance-kuttl-tests/glance-default-external-api-1" Dec 01 09:01:52 crc kubenswrapper[4813]: I1201 09:01:52.954494 4813 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-external-api-1\" (UID: \"7790813b-eccc-442b-a400-7fb5863599a7\") device mount path \"/mnt/openstack/pv08\"" pod="glance-kuttl-tests/glance-default-external-api-1" Dec 01 09:01:52 crc kubenswrapper[4813]: I1201 09:01:52.954051 4813 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage17-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage17-crc\") pod \"glance-default-external-api-1\" (UID: \"7790813b-eccc-442b-a400-7fb5863599a7\") device mount path \"/mnt/openstack/pv17\"" pod="glance-kuttl-tests/glance-default-external-api-1" Dec 01 09:01:52 crc kubenswrapper[4813]: I1201 09:01:52.954560 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/43861577-d77d-45ed-95ea-25a5dffa1b46-sys\") pod \"glance-default-internal-api-2\" (UID: \"43861577-d77d-45ed-95ea-25a5dffa1b46\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Dec 01 09:01:52 crc kubenswrapper[4813]: I1201 09:01:52.954609 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/43861577-d77d-45ed-95ea-25a5dffa1b46-dev\") pod \"glance-default-internal-api-2\" (UID: \"43861577-d77d-45ed-95ea-25a5dffa1b46\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Dec 01 09:01:52 crc kubenswrapper[4813]: I1201 09:01:52.954630 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/7790813b-eccc-442b-a400-7fb5863599a7-sys\") pod \"glance-default-external-api-1\" (UID: \"7790813b-eccc-442b-a400-7fb5863599a7\") " pod="glance-kuttl-tests/glance-default-external-api-1" Dec 01 09:01:52 crc kubenswrapper[4813]: I1201 09:01:52.954193 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/e58d9e76-7271-47c3-a7aa-a5a802bfdb41-dev\") pod \"glance-default-internal-api-1\" (UID: \"e58d9e76-7271-47c3-a7aa-a5a802bfdb41\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Dec 01 09:01:52 crc kubenswrapper[4813]: I1201 09:01:52.954698 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7790813b-eccc-442b-a400-7fb5863599a7-scripts\") pod \"glance-default-external-api-1\" (UID: \"7790813b-eccc-442b-a400-7fb5863599a7\") " pod="glance-kuttl-tests/glance-default-external-api-1" Dec 01 09:01:52 crc kubenswrapper[4813]: I1201 09:01:52.954719 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/43861577-d77d-45ed-95ea-25a5dffa1b46-lib-modules\") pod \"glance-default-internal-api-2\" (UID: \"43861577-d77d-45ed-95ea-25a5dffa1b46\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Dec 01 09:01:52 crc kubenswrapper[4813]: I1201 09:01:52.954762 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/43861577-d77d-45ed-95ea-25a5dffa1b46-config-data\") pod \"glance-default-internal-api-2\" (UID: \"43861577-d77d-45ed-95ea-25a5dffa1b46\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Dec 01 09:01:52 crc kubenswrapper[4813]: I1201 09:01:52.954786 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/7790813b-eccc-442b-a400-7fb5863599a7-dev\") pod \"glance-default-external-api-1\" (UID: \"7790813b-eccc-442b-a400-7fb5863599a7\") " pod="glance-kuttl-tests/glance-default-external-api-1" Dec 01 09:01:52 crc kubenswrapper[4813]: I1201 09:01:52.954805 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage20-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage20-crc\") pod \"glance-default-internal-api-1\" (UID: \"e58d9e76-7271-47c3-a7aa-a5a802bfdb41\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Dec 01 09:01:52 crc kubenswrapper[4813]: I1201 09:01:52.954850 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage14-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage14-crc\") pod \"glance-default-internal-api-1\" (UID: \"e58d9e76-7271-47c3-a7aa-a5a802bfdb41\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Dec 01 09:01:52 crc kubenswrapper[4813]: I1201 09:01:52.954867 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dhtk6\" (UniqueName: \"kubernetes.io/projected/43861577-d77d-45ed-95ea-25a5dffa1b46-kube-api-access-dhtk6\") pod \"glance-default-internal-api-2\" (UID: \"43861577-d77d-45ed-95ea-25a5dffa1b46\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Dec 01 09:01:52 crc kubenswrapper[4813]: I1201 09:01:52.954888 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e58d9e76-7271-47c3-a7aa-a5a802bfdb41-config-data\") pod \"glance-default-internal-api-1\" (UID: \"e58d9e76-7271-47c3-a7aa-a5a802bfdb41\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Dec 01 09:01:52 crc kubenswrapper[4813]: I1201 09:01:52.954934 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/e58d9e76-7271-47c3-a7aa-a5a802bfdb41-run\") pod \"glance-default-internal-api-1\" (UID: \"e58d9e76-7271-47c3-a7aa-a5a802bfdb41\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Dec 01 09:01:52 crc kubenswrapper[4813]: I1201 09:01:52.954950 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/43861577-d77d-45ed-95ea-25a5dffa1b46-etc-iscsi\") pod \"glance-default-internal-api-2\" (UID: \"43861577-d77d-45ed-95ea-25a5dffa1b46\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Dec 01 09:01:52 crc kubenswrapper[4813]: I1201 09:01:52.955002 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/7790813b-eccc-442b-a400-7fb5863599a7-run\") pod \"glance-default-external-api-1\" (UID: \"7790813b-eccc-442b-a400-7fb5863599a7\") " pod="glance-kuttl-tests/glance-default-external-api-1" Dec 01 09:01:52 crc kubenswrapper[4813]: I1201 09:01:52.955026 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/43861577-d77d-45ed-95ea-25a5dffa1b46-logs\") pod \"glance-default-internal-api-2\" (UID: \"43861577-d77d-45ed-95ea-25a5dffa1b46\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Dec 01 09:01:52 crc kubenswrapper[4813]: I1201 09:01:52.955042 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e58d9e76-7271-47c3-a7aa-a5a802bfdb41-scripts\") pod \"glance-default-internal-api-1\" (UID: \"e58d9e76-7271-47c3-a7aa-a5a802bfdb41\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Dec 01 09:01:52 crc kubenswrapper[4813]: I1201 09:01:52.955090 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/e58d9e76-7271-47c3-a7aa-a5a802bfdb41-etc-nvme\") pod \"glance-default-internal-api-1\" (UID: \"e58d9e76-7271-47c3-a7aa-a5a802bfdb41\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Dec 01 09:01:52 crc kubenswrapper[4813]: I1201 09:01:52.955107 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e58d9e76-7271-47c3-a7aa-a5a802bfdb41-httpd-run\") pod \"glance-default-internal-api-1\" (UID: \"e58d9e76-7271-47c3-a7aa-a5a802bfdb41\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Dec 01 09:01:52 crc kubenswrapper[4813]: I1201 09:01:52.955466 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e58d9e76-7271-47c3-a7aa-a5a802bfdb41-httpd-run\") pod \"glance-default-internal-api-1\" (UID: \"e58d9e76-7271-47c3-a7aa-a5a802bfdb41\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Dec 01 09:01:52 crc kubenswrapper[4813]: I1201 09:01:52.956269 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/7790813b-eccc-442b-a400-7fb5863599a7-dev\") pod \"glance-default-external-api-1\" (UID: \"7790813b-eccc-442b-a400-7fb5863599a7\") " pod="glance-kuttl-tests/glance-default-external-api-1" Dec 01 09:01:52 crc kubenswrapper[4813]: I1201 09:01:52.956322 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/43861577-d77d-45ed-95ea-25a5dffa1b46-lib-modules\") pod \"glance-default-internal-api-2\" (UID: \"43861577-d77d-45ed-95ea-25a5dffa1b46\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Dec 01 09:01:52 crc kubenswrapper[4813]: I1201 09:01:52.956359 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/7790813b-eccc-442b-a400-7fb5863599a7-run\") pod \"glance-default-external-api-1\" (UID: \"7790813b-eccc-442b-a400-7fb5863599a7\") " pod="glance-kuttl-tests/glance-default-external-api-1" Dec 01 09:01:52 crc kubenswrapper[4813]: I1201 09:01:52.956390 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/e58d9e76-7271-47c3-a7aa-a5a802bfdb41-run\") pod \"glance-default-internal-api-1\" (UID: \"e58d9e76-7271-47c3-a7aa-a5a802bfdb41\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Dec 01 09:01:52 crc kubenswrapper[4813]: I1201 09:01:52.956421 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/43861577-d77d-45ed-95ea-25a5dffa1b46-etc-iscsi\") pod \"glance-default-internal-api-2\" (UID: \"43861577-d77d-45ed-95ea-25a5dffa1b46\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Dec 01 09:01:52 crc kubenswrapper[4813]: I1201 09:01:52.956539 4813 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage20-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage20-crc\") pod \"glance-default-internal-api-1\" (UID: \"e58d9e76-7271-47c3-a7aa-a5a802bfdb41\") device mount path \"/mnt/openstack/pv20\"" pod="glance-kuttl-tests/glance-default-internal-api-1" Dec 01 09:01:52 crc kubenswrapper[4813]: I1201 09:01:52.956707 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/43861577-d77d-45ed-95ea-25a5dffa1b46-httpd-run\") pod \"glance-default-internal-api-2\" (UID: \"43861577-d77d-45ed-95ea-25a5dffa1b46\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Dec 01 09:01:52 crc kubenswrapper[4813]: I1201 09:01:52.957754 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/43861577-d77d-45ed-95ea-25a5dffa1b46-logs\") pod \"glance-default-internal-api-2\" (UID: \"43861577-d77d-45ed-95ea-25a5dffa1b46\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Dec 01 09:01:52 crc kubenswrapper[4813]: I1201 09:01:52.957858 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/e58d9e76-7271-47c3-a7aa-a5a802bfdb41-etc-nvme\") pod \"glance-default-internal-api-1\" (UID: \"e58d9e76-7271-47c3-a7aa-a5a802bfdb41\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Dec 01 09:01:52 crc kubenswrapper[4813]: I1201 09:01:52.958531 4813 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage14-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage14-crc\") pod \"glance-default-internal-api-1\" (UID: \"e58d9e76-7271-47c3-a7aa-a5a802bfdb41\") device mount path \"/mnt/openstack/pv14\"" pod="glance-kuttl-tests/glance-default-internal-api-1" Dec 01 09:01:52 crc kubenswrapper[4813]: I1201 09:01:52.958739 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/43861577-d77d-45ed-95ea-25a5dffa1b46-run\") pod \"glance-default-internal-api-2\" (UID: \"43861577-d77d-45ed-95ea-25a5dffa1b46\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Dec 01 09:01:52 crc kubenswrapper[4813]: I1201 09:01:52.960564 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-2" Dec 01 09:01:52 crc kubenswrapper[4813]: I1201 09:01:52.960851 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7790813b-eccc-442b-a400-7fb5863599a7-scripts\") pod \"glance-default-external-api-1\" (UID: \"7790813b-eccc-442b-a400-7fb5863599a7\") " pod="glance-kuttl-tests/glance-default-external-api-1" Dec 01 09:01:52 crc kubenswrapper[4813]: I1201 09:01:52.962425 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/43861577-d77d-45ed-95ea-25a5dffa1b46-scripts\") pod \"glance-default-internal-api-2\" (UID: \"43861577-d77d-45ed-95ea-25a5dffa1b46\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Dec 01 09:01:52 crc kubenswrapper[4813]: I1201 09:01:52.962471 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e58d9e76-7271-47c3-a7aa-a5a802bfdb41-scripts\") pod \"glance-default-internal-api-1\" (UID: \"e58d9e76-7271-47c3-a7aa-a5a802bfdb41\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Dec 01 09:01:52 crc kubenswrapper[4813]: I1201 09:01:52.965093 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7790813b-eccc-442b-a400-7fb5863599a7-config-data\") pod \"glance-default-external-api-1\" (UID: \"7790813b-eccc-442b-a400-7fb5863599a7\") " pod="glance-kuttl-tests/glance-default-external-api-1" Dec 01 09:01:52 crc kubenswrapper[4813]: I1201 09:01:52.967226 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/43861577-d77d-45ed-95ea-25a5dffa1b46-config-data\") pod \"glance-default-internal-api-2\" (UID: \"43861577-d77d-45ed-95ea-25a5dffa1b46\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Dec 01 09:01:52 crc kubenswrapper[4813]: I1201 09:01:52.972878 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e58d9e76-7271-47c3-a7aa-a5a802bfdb41-config-data\") pod \"glance-default-internal-api-1\" (UID: \"e58d9e76-7271-47c3-a7aa-a5a802bfdb41\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Dec 01 09:01:52 crc kubenswrapper[4813]: I1201 09:01:52.979096 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-njccq\" (UniqueName: \"kubernetes.io/projected/e58d9e76-7271-47c3-a7aa-a5a802bfdb41-kube-api-access-njccq\") pod \"glance-default-internal-api-1\" (UID: \"e58d9e76-7271-47c3-a7aa-a5a802bfdb41\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Dec 01 09:01:52 crc kubenswrapper[4813]: I1201 09:01:52.979443 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tkp6h\" (UniqueName: \"kubernetes.io/projected/7790813b-eccc-442b-a400-7fb5863599a7-kube-api-access-tkp6h\") pod \"glance-default-external-api-1\" (UID: \"7790813b-eccc-442b-a400-7fb5863599a7\") " pod="glance-kuttl-tests/glance-default-external-api-1" Dec 01 09:01:52 crc kubenswrapper[4813]: I1201 09:01:52.981331 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dhtk6\" (UniqueName: \"kubernetes.io/projected/43861577-d77d-45ed-95ea-25a5dffa1b46-kube-api-access-dhtk6\") pod \"glance-default-internal-api-2\" (UID: \"43861577-d77d-45ed-95ea-25a5dffa1b46\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Dec 01 09:01:52 crc kubenswrapper[4813]: I1201 09:01:52.990215 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage20-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage20-crc\") pod \"glance-default-internal-api-1\" (UID: \"e58d9e76-7271-47c3-a7aa-a5a802bfdb41\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Dec 01 09:01:52 crc kubenswrapper[4813]: I1201 09:01:52.993205 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage13-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage13-crc\") pod \"glance-default-internal-api-2\" (UID: \"43861577-d77d-45ed-95ea-25a5dffa1b46\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Dec 01 09:01:53 crc kubenswrapper[4813]: I1201 09:01:53.000505 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-2\" (UID: \"43861577-d77d-45ed-95ea-25a5dffa1b46\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Dec 01 09:01:53 crc kubenswrapper[4813]: I1201 09:01:53.025671 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage14-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage14-crc\") pod \"glance-default-internal-api-1\" (UID: \"e58d9e76-7271-47c3-a7aa-a5a802bfdb41\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Dec 01 09:01:53 crc kubenswrapper[4813]: I1201 09:01:53.036109 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-external-api-1\" (UID: \"7790813b-eccc-442b-a400-7fb5863599a7\") " pod="glance-kuttl-tests/glance-default-external-api-1" Dec 01 09:01:53 crc kubenswrapper[4813]: I1201 09:01:53.052587 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-2" Dec 01 09:01:53 crc kubenswrapper[4813]: I1201 09:01:53.068131 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-1" Dec 01 09:01:53 crc kubenswrapper[4813]: I1201 09:01:53.081292 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage17-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage17-crc\") pod \"glance-default-external-api-1\" (UID: \"7790813b-eccc-442b-a400-7fb5863599a7\") " pod="glance-kuttl-tests/glance-default-external-api-1" Dec 01 09:01:53 crc kubenswrapper[4813]: I1201 09:01:53.275131 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-1" Dec 01 09:01:53 crc kubenswrapper[4813]: I1201 09:01:53.444014 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-2"] Dec 01 09:01:53 crc kubenswrapper[4813]: W1201 09:01:53.450487 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0bb179fc_fb0c_420a_a432_5a0d28884e23.slice/crio-4f734a074b74139ba6affc141d884d25fad6a85dd48c9b655f11b259d73c68a6 WatchSource:0}: Error finding container 4f734a074b74139ba6affc141d884d25fad6a85dd48c9b655f11b259d73c68a6: Status 404 returned error can't find the container with id 4f734a074b74139ba6affc141d884d25fad6a85dd48c9b655f11b259d73c68a6 Dec 01 09:01:53 crc kubenswrapper[4813]: I1201 09:01:53.482585 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-1"] Dec 01 09:01:53 crc kubenswrapper[4813]: W1201 09:01:53.487712 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7790813b_eccc_442b_a400_7fb5863599a7.slice/crio-40dd0b7da1121b814b593d0717d87acc08858d5045aebf43e888488cafad2f25 WatchSource:0}: Error finding container 40dd0b7da1121b814b593d0717d87acc08858d5045aebf43e888488cafad2f25: Status 404 returned error can't find the container with id 40dd0b7da1121b814b593d0717d87acc08858d5045aebf43e888488cafad2f25 Dec 01 09:01:53 crc kubenswrapper[4813]: I1201 09:01:53.503246 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-2"] Dec 01 09:01:53 crc kubenswrapper[4813]: I1201 09:01:53.565930 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-1"] Dec 01 09:01:53 crc kubenswrapper[4813]: I1201 09:01:53.624939 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-1" event={"ID":"e58d9e76-7271-47c3-a7aa-a5a802bfdb41","Type":"ContainerStarted","Data":"a07606be2b91ee3510badb11590d95baadd2564242ac6b630cdc16395498dd27"} Dec 01 09:01:53 crc kubenswrapper[4813]: I1201 09:01:53.626040 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-2" event={"ID":"43861577-d77d-45ed-95ea-25a5dffa1b46","Type":"ContainerStarted","Data":"5b61076ea0b84ec21636b478ed4ea8081d507abb0a8eb7925b449e58e195f906"} Dec 01 09:01:53 crc kubenswrapper[4813]: I1201 09:01:53.627093 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-1" event={"ID":"7790813b-eccc-442b-a400-7fb5863599a7","Type":"ContainerStarted","Data":"40dd0b7da1121b814b593d0717d87acc08858d5045aebf43e888488cafad2f25"} Dec 01 09:01:53 crc kubenswrapper[4813]: I1201 09:01:53.628396 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-2" event={"ID":"0bb179fc-fb0c-420a-a432-5a0d28884e23","Type":"ContainerStarted","Data":"4f734a074b74139ba6affc141d884d25fad6a85dd48c9b655f11b259d73c68a6"} Dec 01 09:01:54 crc kubenswrapper[4813]: I1201 09:01:54.657196 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-1" event={"ID":"e58d9e76-7271-47c3-a7aa-a5a802bfdb41","Type":"ContainerStarted","Data":"6999bd053c70cefb09a4011183136ca0736cf8688cd73cedf3627e8fb0b87dd7"} Dec 01 09:01:54 crc kubenswrapper[4813]: I1201 09:01:54.657506 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-1" event={"ID":"e58d9e76-7271-47c3-a7aa-a5a802bfdb41","Type":"ContainerStarted","Data":"61191cf315a8ec0efc4e03441826363ac77f4ce803b8cd8107d188390367161b"} Dec 01 09:01:54 crc kubenswrapper[4813]: I1201 09:01:54.661699 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-2" event={"ID":"43861577-d77d-45ed-95ea-25a5dffa1b46","Type":"ContainerStarted","Data":"16e8434757ae8743d0e044be613fba36879ad53137caa53097886a44032e9949"} Dec 01 09:01:54 crc kubenswrapper[4813]: I1201 09:01:54.661758 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-2" event={"ID":"43861577-d77d-45ed-95ea-25a5dffa1b46","Type":"ContainerStarted","Data":"72494af8d33fe162c335935341321833cde5cf4fce403b9ba7eaf72165deb984"} Dec 01 09:01:54 crc kubenswrapper[4813]: I1201 09:01:54.663402 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-1" event={"ID":"7790813b-eccc-442b-a400-7fb5863599a7","Type":"ContainerStarted","Data":"d1cabd26ab94cbb2f1d690a779a0d599930f7e7ab3d89783325610a597410b3f"} Dec 01 09:01:54 crc kubenswrapper[4813]: I1201 09:01:54.663441 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-1" event={"ID":"7790813b-eccc-442b-a400-7fb5863599a7","Type":"ContainerStarted","Data":"a1e8ceb6fe1d2778c4ba07d2f6328f0b7a52d25bf8d2de9445e17eb86a2d71ab"} Dec 01 09:01:54 crc kubenswrapper[4813]: I1201 09:01:54.666077 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-2" event={"ID":"0bb179fc-fb0c-420a-a432-5a0d28884e23","Type":"ContainerStarted","Data":"a2ff5f4871bec4b409097a90d5929c75feb2f96c17f8cf1b75cbfa57f68a9374"} Dec 01 09:01:54 crc kubenswrapper[4813]: I1201 09:01:54.666105 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-2" event={"ID":"0bb179fc-fb0c-420a-a432-5a0d28884e23","Type":"ContainerStarted","Data":"39ba0adee13915c1669441022a9cf3d86928a608ad457ed24bc0ab7ca527f640"} Dec 01 09:01:54 crc kubenswrapper[4813]: I1201 09:01:54.682348 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-default-internal-api-1" podStartSLOduration=3.682326646 podStartE2EDuration="3.682326646s" podCreationTimestamp="2025-12-01 09:01:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 09:01:54.681639717 +0000 UTC m=+1299.304561303" watchObservedRunningTime="2025-12-01 09:01:54.682326646 +0000 UTC m=+1299.305248232" Dec 01 09:01:54 crc kubenswrapper[4813]: I1201 09:01:54.708492 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-default-external-api-2" podStartSLOduration=3.7084703340000003 podStartE2EDuration="3.708470334s" podCreationTimestamp="2025-12-01 09:01:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 09:01:54.70087898 +0000 UTC m=+1299.323800566" watchObservedRunningTime="2025-12-01 09:01:54.708470334 +0000 UTC m=+1299.331391920" Dec 01 09:01:54 crc kubenswrapper[4813]: I1201 09:01:54.733207 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-default-internal-api-2" podStartSLOduration=3.733180662 podStartE2EDuration="3.733180662s" podCreationTimestamp="2025-12-01 09:01:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 09:01:54.723428777 +0000 UTC m=+1299.346350363" watchObservedRunningTime="2025-12-01 09:01:54.733180662 +0000 UTC m=+1299.356102248" Dec 01 09:01:54 crc kubenswrapper[4813]: I1201 09:01:54.752407 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-default-external-api-1" podStartSLOduration=3.752382945 podStartE2EDuration="3.752382945s" podCreationTimestamp="2025-12-01 09:01:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 09:01:54.742881477 +0000 UTC m=+1299.365803063" watchObservedRunningTime="2025-12-01 09:01:54.752382945 +0000 UTC m=+1299.375304541" Dec 01 09:02:02 crc kubenswrapper[4813]: I1201 09:02:02.962170 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-external-api-2" Dec 01 09:02:02 crc kubenswrapper[4813]: I1201 09:02:02.962768 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-external-api-2" Dec 01 09:02:02 crc kubenswrapper[4813]: I1201 09:02:02.985530 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-external-api-2" Dec 01 09:02:03 crc kubenswrapper[4813]: I1201 09:02:03.005714 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-external-api-2" Dec 01 09:02:03 crc kubenswrapper[4813]: I1201 09:02:03.053498 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-internal-api-2" Dec 01 09:02:03 crc kubenswrapper[4813]: I1201 09:02:03.053558 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-internal-api-2" Dec 01 09:02:03 crc kubenswrapper[4813]: I1201 09:02:03.069732 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-internal-api-1" Dec 01 09:02:03 crc kubenswrapper[4813]: I1201 09:02:03.069777 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-internal-api-1" Dec 01 09:02:03 crc kubenswrapper[4813]: I1201 09:02:03.081620 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-internal-api-2" Dec 01 09:02:03 crc kubenswrapper[4813]: I1201 09:02:03.096079 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-internal-api-2" Dec 01 09:02:03 crc kubenswrapper[4813]: I1201 09:02:03.103201 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-internal-api-1" Dec 01 09:02:03 crc kubenswrapper[4813]: I1201 09:02:03.124826 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-internal-api-1" Dec 01 09:02:03 crc kubenswrapper[4813]: I1201 09:02:03.276073 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-external-api-1" Dec 01 09:02:03 crc kubenswrapper[4813]: I1201 09:02:03.276160 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-external-api-1" Dec 01 09:02:03 crc kubenswrapper[4813]: I1201 09:02:03.419826 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-external-api-1" Dec 01 09:02:03 crc kubenswrapper[4813]: I1201 09:02:03.425165 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-external-api-1" Dec 01 09:02:03 crc kubenswrapper[4813]: I1201 09:02:03.783820 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-external-api-2" Dec 01 09:02:03 crc kubenswrapper[4813]: I1201 09:02:03.783946 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-external-api-1" Dec 01 09:02:03 crc kubenswrapper[4813]: I1201 09:02:03.783975 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-external-api-1" Dec 01 09:02:03 crc kubenswrapper[4813]: I1201 09:02:03.783987 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-internal-api-2" Dec 01 09:02:03 crc kubenswrapper[4813]: I1201 09:02:03.783999 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-external-api-2" Dec 01 09:02:03 crc kubenswrapper[4813]: I1201 09:02:03.784011 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-internal-api-1" Dec 01 09:02:03 crc kubenswrapper[4813]: I1201 09:02:03.784022 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-internal-api-1" Dec 01 09:02:03 crc kubenswrapper[4813]: I1201 09:02:03.784459 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-internal-api-2" Dec 01 09:02:05 crc kubenswrapper[4813]: I1201 09:02:05.878096 4813 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 01 09:02:05 crc kubenswrapper[4813]: I1201 09:02:05.878497 4813 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 01 09:02:05 crc kubenswrapper[4813]: I1201 09:02:05.878136 4813 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 01 09:02:05 crc kubenswrapper[4813]: I1201 09:02:05.878821 4813 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 01 09:02:05 crc kubenswrapper[4813]: I1201 09:02:05.878188 4813 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 01 09:02:05 crc kubenswrapper[4813]: I1201 09:02:05.879509 4813 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 01 09:02:05 crc kubenswrapper[4813]: I1201 09:02:05.878208 4813 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 01 09:02:05 crc kubenswrapper[4813]: I1201 09:02:05.879866 4813 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 01 09:02:06 crc kubenswrapper[4813]: I1201 09:02:06.167584 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-internal-api-2" Dec 01 09:02:06 crc kubenswrapper[4813]: I1201 09:02:06.340654 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-internal-api-2" Dec 01 09:02:06 crc kubenswrapper[4813]: I1201 09:02:06.530565 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-external-api-2" Dec 01 09:02:06 crc kubenswrapper[4813]: I1201 09:02:06.638116 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-internal-api-1" Dec 01 09:02:06 crc kubenswrapper[4813]: I1201 09:02:06.639033 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-internal-api-1" Dec 01 09:02:06 crc kubenswrapper[4813]: I1201 09:02:06.694969 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-external-api-2" Dec 01 09:02:06 crc kubenswrapper[4813]: I1201 09:02:06.734295 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-external-api-1" Dec 01 09:02:06 crc kubenswrapper[4813]: I1201 09:02:06.769321 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-external-api-1" Dec 01 09:02:07 crc kubenswrapper[4813]: I1201 09:02:07.466589 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-2"] Dec 01 09:02:07 crc kubenswrapper[4813]: I1201 09:02:07.489912 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-1"] Dec 01 09:02:07 crc kubenswrapper[4813]: I1201 09:02:07.709817 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-2"] Dec 01 09:02:07 crc kubenswrapper[4813]: I1201 09:02:07.720834 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-1"] Dec 01 09:02:08 crc kubenswrapper[4813]: I1201 09:02:08.942543 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-internal-api-1" podUID="e58d9e76-7271-47c3-a7aa-a5a802bfdb41" containerName="glance-log" containerID="cri-o://61191cf315a8ec0efc4e03441826363ac77f4ce803b8cd8107d188390367161b" gracePeriod=30 Dec 01 09:02:08 crc kubenswrapper[4813]: I1201 09:02:08.942573 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-internal-api-1" podUID="e58d9e76-7271-47c3-a7aa-a5a802bfdb41" containerName="glance-httpd" containerID="cri-o://6999bd053c70cefb09a4011183136ca0736cf8688cd73cedf3627e8fb0b87dd7" gracePeriod=30 Dec 01 09:02:08 crc kubenswrapper[4813]: I1201 09:02:08.942647 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-external-api-1" podUID="7790813b-eccc-442b-a400-7fb5863599a7" containerName="glance-httpd" containerID="cri-o://d1cabd26ab94cbb2f1d690a779a0d599930f7e7ab3d89783325610a597410b3f" gracePeriod=30 Dec 01 09:02:08 crc kubenswrapper[4813]: I1201 09:02:08.942578 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-external-api-1" podUID="7790813b-eccc-442b-a400-7fb5863599a7" containerName="glance-log" containerID="cri-o://a1e8ceb6fe1d2778c4ba07d2f6328f0b7a52d25bf8d2de9445e17eb86a2d71ab" gracePeriod=30 Dec 01 09:02:08 crc kubenswrapper[4813]: I1201 09:02:08.942957 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-internal-api-2" podUID="43861577-d77d-45ed-95ea-25a5dffa1b46" containerName="glance-log" containerID="cri-o://72494af8d33fe162c335935341321833cde5cf4fce403b9ba7eaf72165deb984" gracePeriod=30 Dec 01 09:02:08 crc kubenswrapper[4813]: I1201 09:02:08.943180 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-external-api-2" podUID="0bb179fc-fb0c-420a-a432-5a0d28884e23" containerName="glance-log" containerID="cri-o://39ba0adee13915c1669441022a9cf3d86928a608ad457ed24bc0ab7ca527f640" gracePeriod=30 Dec 01 09:02:08 crc kubenswrapper[4813]: I1201 09:02:08.943289 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-internal-api-2" podUID="43861577-d77d-45ed-95ea-25a5dffa1b46" containerName="glance-httpd" containerID="cri-o://16e8434757ae8743d0e044be613fba36879ad53137caa53097886a44032e9949" gracePeriod=30 Dec 01 09:02:08 crc kubenswrapper[4813]: I1201 09:02:08.943606 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-external-api-2" podUID="0bb179fc-fb0c-420a-a432-5a0d28884e23" containerName="glance-httpd" containerID="cri-o://a2ff5f4871bec4b409097a90d5929c75feb2f96c17f8cf1b75cbfa57f68a9374" gracePeriod=30 Dec 01 09:02:08 crc kubenswrapper[4813]: I1201 09:02:08.949538 4813 prober.go:107] "Probe failed" probeType="Readiness" pod="glance-kuttl-tests/glance-default-internal-api-1" podUID="e58d9e76-7271-47c3-a7aa-a5a802bfdb41" containerName="glance-httpd" probeResult="failure" output="Get \"http://10.217.0.131:9292/healthcheck\": EOF" Dec 01 09:02:08 crc kubenswrapper[4813]: I1201 09:02:08.951699 4813 prober.go:107] "Probe failed" probeType="Readiness" pod="glance-kuttl-tests/glance-default-internal-api-1" podUID="e58d9e76-7271-47c3-a7aa-a5a802bfdb41" containerName="glance-log" probeResult="failure" output="Get \"http://10.217.0.131:9292/healthcheck\": EOF" Dec 01 09:02:08 crc kubenswrapper[4813]: I1201 09:02:08.952130 4813 prober.go:107] "Probe failed" probeType="Readiness" pod="glance-kuttl-tests/glance-default-external-api-2" podUID="0bb179fc-fb0c-420a-a432-5a0d28884e23" containerName="glance-httpd" probeResult="failure" output="Get \"http://10.217.0.128:9292/healthcheck\": EOF" Dec 01 09:02:08 crc kubenswrapper[4813]: I1201 09:02:08.960805 4813 prober.go:107] "Probe failed" probeType="Readiness" pod="glance-kuttl-tests/glance-default-internal-api-2" podUID="43861577-d77d-45ed-95ea-25a5dffa1b46" containerName="glance-httpd" probeResult="failure" output="Get \"http://10.217.0.130:9292/healthcheck\": EOF" Dec 01 09:02:09 crc kubenswrapper[4813]: I1201 09:02:09.953926 4813 generic.go:334] "Generic (PLEG): container finished" podID="e58d9e76-7271-47c3-a7aa-a5a802bfdb41" containerID="61191cf315a8ec0efc4e03441826363ac77f4ce803b8cd8107d188390367161b" exitCode=143 Dec 01 09:02:09 crc kubenswrapper[4813]: I1201 09:02:09.954019 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-1" event={"ID":"e58d9e76-7271-47c3-a7aa-a5a802bfdb41","Type":"ContainerDied","Data":"61191cf315a8ec0efc4e03441826363ac77f4ce803b8cd8107d188390367161b"} Dec 01 09:02:09 crc kubenswrapper[4813]: I1201 09:02:09.956355 4813 generic.go:334] "Generic (PLEG): container finished" podID="43861577-d77d-45ed-95ea-25a5dffa1b46" containerID="72494af8d33fe162c335935341321833cde5cf4fce403b9ba7eaf72165deb984" exitCode=143 Dec 01 09:02:09 crc kubenswrapper[4813]: I1201 09:02:09.956449 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-2" event={"ID":"43861577-d77d-45ed-95ea-25a5dffa1b46","Type":"ContainerDied","Data":"72494af8d33fe162c335935341321833cde5cf4fce403b9ba7eaf72165deb984"} Dec 01 09:02:09 crc kubenswrapper[4813]: I1201 09:02:09.958333 4813 generic.go:334] "Generic (PLEG): container finished" podID="7790813b-eccc-442b-a400-7fb5863599a7" containerID="a1e8ceb6fe1d2778c4ba07d2f6328f0b7a52d25bf8d2de9445e17eb86a2d71ab" exitCode=143 Dec 01 09:02:09 crc kubenswrapper[4813]: I1201 09:02:09.958401 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-1" event={"ID":"7790813b-eccc-442b-a400-7fb5863599a7","Type":"ContainerDied","Data":"a1e8ceb6fe1d2778c4ba07d2f6328f0b7a52d25bf8d2de9445e17eb86a2d71ab"} Dec 01 09:02:09 crc kubenswrapper[4813]: I1201 09:02:09.960786 4813 generic.go:334] "Generic (PLEG): container finished" podID="0bb179fc-fb0c-420a-a432-5a0d28884e23" containerID="39ba0adee13915c1669441022a9cf3d86928a608ad457ed24bc0ab7ca527f640" exitCode=143 Dec 01 09:02:09 crc kubenswrapper[4813]: I1201 09:02:09.960825 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-2" event={"ID":"0bb179fc-fb0c-420a-a432-5a0d28884e23","Type":"ContainerDied","Data":"39ba0adee13915c1669441022a9cf3d86928a608ad457ed24bc0ab7ca527f640"} Dec 01 09:02:12 crc kubenswrapper[4813]: I1201 09:02:12.792779 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-2" Dec 01 09:02:12 crc kubenswrapper[4813]: I1201 09:02:12.977111 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage13-crc\") pod \"43861577-d77d-45ed-95ea-25a5dffa1b46\" (UID: \"43861577-d77d-45ed-95ea-25a5dffa1b46\") " Dec 01 09:02:12 crc kubenswrapper[4813]: I1201 09:02:12.977176 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dhtk6\" (UniqueName: \"kubernetes.io/projected/43861577-d77d-45ed-95ea-25a5dffa1b46-kube-api-access-dhtk6\") pod \"43861577-d77d-45ed-95ea-25a5dffa1b46\" (UID: \"43861577-d77d-45ed-95ea-25a5dffa1b46\") " Dec 01 09:02:12 crc kubenswrapper[4813]: I1201 09:02:12.977253 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/43861577-d77d-45ed-95ea-25a5dffa1b46-run\") pod \"43861577-d77d-45ed-95ea-25a5dffa1b46\" (UID: \"43861577-d77d-45ed-95ea-25a5dffa1b46\") " Dec 01 09:02:12 crc kubenswrapper[4813]: I1201 09:02:12.977294 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/43861577-d77d-45ed-95ea-25a5dffa1b46-var-locks-brick\") pod \"43861577-d77d-45ed-95ea-25a5dffa1b46\" (UID: \"43861577-d77d-45ed-95ea-25a5dffa1b46\") " Dec 01 09:02:12 crc kubenswrapper[4813]: I1201 09:02:12.977344 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/43861577-d77d-45ed-95ea-25a5dffa1b46-lib-modules\") pod \"43861577-d77d-45ed-95ea-25a5dffa1b46\" (UID: \"43861577-d77d-45ed-95ea-25a5dffa1b46\") " Dec 01 09:02:12 crc kubenswrapper[4813]: I1201 09:02:12.977397 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/43861577-d77d-45ed-95ea-25a5dffa1b46-scripts\") pod \"43861577-d77d-45ed-95ea-25a5dffa1b46\" (UID: \"43861577-d77d-45ed-95ea-25a5dffa1b46\") " Dec 01 09:02:12 crc kubenswrapper[4813]: I1201 09:02:12.977446 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/43861577-d77d-45ed-95ea-25a5dffa1b46-config-data\") pod \"43861577-d77d-45ed-95ea-25a5dffa1b46\" (UID: \"43861577-d77d-45ed-95ea-25a5dffa1b46\") " Dec 01 09:02:12 crc kubenswrapper[4813]: I1201 09:02:12.977611 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/43861577-d77d-45ed-95ea-25a5dffa1b46-var-locks-brick" (OuterVolumeSpecName: "var-locks-brick") pod "43861577-d77d-45ed-95ea-25a5dffa1b46" (UID: "43861577-d77d-45ed-95ea-25a5dffa1b46"). InnerVolumeSpecName "var-locks-brick". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 09:02:12 crc kubenswrapper[4813]: I1201 09:02:12.977715 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/43861577-d77d-45ed-95ea-25a5dffa1b46-run" (OuterVolumeSpecName: "run") pod "43861577-d77d-45ed-95ea-25a5dffa1b46" (UID: "43861577-d77d-45ed-95ea-25a5dffa1b46"). InnerVolumeSpecName "run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 09:02:12 crc kubenswrapper[4813]: I1201 09:02:12.977749 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/43861577-d77d-45ed-95ea-25a5dffa1b46-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "43861577-d77d-45ed-95ea-25a5dffa1b46" (UID: "43861577-d77d-45ed-95ea-25a5dffa1b46"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 09:02:12 crc kubenswrapper[4813]: I1201 09:02:12.977913 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/43861577-d77d-45ed-95ea-25a5dffa1b46-etc-nvme" (OuterVolumeSpecName: "etc-nvme") pod "43861577-d77d-45ed-95ea-25a5dffa1b46" (UID: "43861577-d77d-45ed-95ea-25a5dffa1b46"). InnerVolumeSpecName "etc-nvme". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 09:02:12 crc kubenswrapper[4813]: I1201 09:02:12.977508 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/43861577-d77d-45ed-95ea-25a5dffa1b46-etc-nvme\") pod \"43861577-d77d-45ed-95ea-25a5dffa1b46\" (UID: \"43861577-d77d-45ed-95ea-25a5dffa1b46\") " Dec 01 09:02:12 crc kubenswrapper[4813]: I1201 09:02:12.978217 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/43861577-d77d-45ed-95ea-25a5dffa1b46-dev\") pod \"43861577-d77d-45ed-95ea-25a5dffa1b46\" (UID: \"43861577-d77d-45ed-95ea-25a5dffa1b46\") " Dec 01 09:02:12 crc kubenswrapper[4813]: I1201 09:02:12.978275 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/43861577-d77d-45ed-95ea-25a5dffa1b46-logs\") pod \"43861577-d77d-45ed-95ea-25a5dffa1b46\" (UID: \"43861577-d77d-45ed-95ea-25a5dffa1b46\") " Dec 01 09:02:12 crc kubenswrapper[4813]: I1201 09:02:12.978315 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance-cache\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"43861577-d77d-45ed-95ea-25a5dffa1b46\" (UID: \"43861577-d77d-45ed-95ea-25a5dffa1b46\") " Dec 01 09:02:12 crc kubenswrapper[4813]: I1201 09:02:12.978336 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/43861577-d77d-45ed-95ea-25a5dffa1b46-etc-iscsi\") pod \"43861577-d77d-45ed-95ea-25a5dffa1b46\" (UID: \"43861577-d77d-45ed-95ea-25a5dffa1b46\") " Dec 01 09:02:12 crc kubenswrapper[4813]: I1201 09:02:12.978363 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/43861577-d77d-45ed-95ea-25a5dffa1b46-sys\") pod \"43861577-d77d-45ed-95ea-25a5dffa1b46\" (UID: \"43861577-d77d-45ed-95ea-25a5dffa1b46\") " Dec 01 09:02:12 crc kubenswrapper[4813]: I1201 09:02:12.978381 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/43861577-d77d-45ed-95ea-25a5dffa1b46-dev" (OuterVolumeSpecName: "dev") pod "43861577-d77d-45ed-95ea-25a5dffa1b46" (UID: "43861577-d77d-45ed-95ea-25a5dffa1b46"). InnerVolumeSpecName "dev". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 09:02:12 crc kubenswrapper[4813]: I1201 09:02:12.978443 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/43861577-d77d-45ed-95ea-25a5dffa1b46-httpd-run\") pod \"43861577-d77d-45ed-95ea-25a5dffa1b46\" (UID: \"43861577-d77d-45ed-95ea-25a5dffa1b46\") " Dec 01 09:02:12 crc kubenswrapper[4813]: I1201 09:02:12.978660 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/43861577-d77d-45ed-95ea-25a5dffa1b46-etc-iscsi" (OuterVolumeSpecName: "etc-iscsi") pod "43861577-d77d-45ed-95ea-25a5dffa1b46" (UID: "43861577-d77d-45ed-95ea-25a5dffa1b46"). InnerVolumeSpecName "etc-iscsi". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 09:02:12 crc kubenswrapper[4813]: I1201 09:02:12.978703 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/43861577-d77d-45ed-95ea-25a5dffa1b46-sys" (OuterVolumeSpecName: "sys") pod "43861577-d77d-45ed-95ea-25a5dffa1b46" (UID: "43861577-d77d-45ed-95ea-25a5dffa1b46"). InnerVolumeSpecName "sys". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 09:02:12 crc kubenswrapper[4813]: I1201 09:02:12.978935 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/43861577-d77d-45ed-95ea-25a5dffa1b46-logs" (OuterVolumeSpecName: "logs") pod "43861577-d77d-45ed-95ea-25a5dffa1b46" (UID: "43861577-d77d-45ed-95ea-25a5dffa1b46"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 09:02:12 crc kubenswrapper[4813]: I1201 09:02:12.979179 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/43861577-d77d-45ed-95ea-25a5dffa1b46-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "43861577-d77d-45ed-95ea-25a5dffa1b46" (UID: "43861577-d77d-45ed-95ea-25a5dffa1b46"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 09:02:12 crc kubenswrapper[4813]: I1201 09:02:12.979369 4813 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/43861577-d77d-45ed-95ea-25a5dffa1b46-logs\") on node \"crc\" DevicePath \"\"" Dec 01 09:02:12 crc kubenswrapper[4813]: I1201 09:02:12.979397 4813 reconciler_common.go:293] "Volume detached for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/43861577-d77d-45ed-95ea-25a5dffa1b46-etc-iscsi\") on node \"crc\" DevicePath \"\"" Dec 01 09:02:12 crc kubenswrapper[4813]: I1201 09:02:12.979408 4813 reconciler_common.go:293] "Volume detached for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/43861577-d77d-45ed-95ea-25a5dffa1b46-sys\") on node \"crc\" DevicePath \"\"" Dec 01 09:02:12 crc kubenswrapper[4813]: I1201 09:02:12.979418 4813 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/43861577-d77d-45ed-95ea-25a5dffa1b46-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 01 09:02:12 crc kubenswrapper[4813]: I1201 09:02:12.979428 4813 reconciler_common.go:293] "Volume detached for volume \"run\" (UniqueName: \"kubernetes.io/host-path/43861577-d77d-45ed-95ea-25a5dffa1b46-run\") on node \"crc\" DevicePath \"\"" Dec 01 09:02:12 crc kubenswrapper[4813]: I1201 09:02:12.979437 4813 reconciler_common.go:293] "Volume detached for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/43861577-d77d-45ed-95ea-25a5dffa1b46-var-locks-brick\") on node \"crc\" DevicePath \"\"" Dec 01 09:02:12 crc kubenswrapper[4813]: I1201 09:02:12.979448 4813 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/43861577-d77d-45ed-95ea-25a5dffa1b46-lib-modules\") on node \"crc\" DevicePath \"\"" Dec 01 09:02:12 crc kubenswrapper[4813]: I1201 09:02:12.979458 4813 reconciler_common.go:293] "Volume detached for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/43861577-d77d-45ed-95ea-25a5dffa1b46-etc-nvme\") on node \"crc\" DevicePath \"\"" Dec 01 09:02:12 crc kubenswrapper[4813]: I1201 09:02:12.979468 4813 reconciler_common.go:293] "Volume detached for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/43861577-d77d-45ed-95ea-25a5dffa1b46-dev\") on node \"crc\" DevicePath \"\"" Dec 01 09:02:12 crc kubenswrapper[4813]: I1201 09:02:12.984368 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage13-crc" (OuterVolumeSpecName: "glance") pod "43861577-d77d-45ed-95ea-25a5dffa1b46" (UID: "43861577-d77d-45ed-95ea-25a5dffa1b46"). InnerVolumeSpecName "local-storage13-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 01 09:02:12 crc kubenswrapper[4813]: I1201 09:02:12.984529 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43861577-d77d-45ed-95ea-25a5dffa1b46-scripts" (OuterVolumeSpecName: "scripts") pod "43861577-d77d-45ed-95ea-25a5dffa1b46" (UID: "43861577-d77d-45ed-95ea-25a5dffa1b46"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:02:12 crc kubenswrapper[4813]: I1201 09:02:12.984956 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage12-crc" (OuterVolumeSpecName: "glance-cache") pod "43861577-d77d-45ed-95ea-25a5dffa1b46" (UID: "43861577-d77d-45ed-95ea-25a5dffa1b46"). InnerVolumeSpecName "local-storage12-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 01 09:02:12 crc kubenswrapper[4813]: I1201 09:02:12.985065 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43861577-d77d-45ed-95ea-25a5dffa1b46-kube-api-access-dhtk6" (OuterVolumeSpecName: "kube-api-access-dhtk6") pod "43861577-d77d-45ed-95ea-25a5dffa1b46" (UID: "43861577-d77d-45ed-95ea-25a5dffa1b46"). InnerVolumeSpecName "kube-api-access-dhtk6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:02:13 crc kubenswrapper[4813]: I1201 09:02:13.005177 4813 generic.go:334] "Generic (PLEG): container finished" podID="7790813b-eccc-442b-a400-7fb5863599a7" containerID="d1cabd26ab94cbb2f1d690a779a0d599930f7e7ab3d89783325610a597410b3f" exitCode=0 Dec 01 09:02:13 crc kubenswrapper[4813]: I1201 09:02:13.005339 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-1" event={"ID":"7790813b-eccc-442b-a400-7fb5863599a7","Type":"ContainerDied","Data":"d1cabd26ab94cbb2f1d690a779a0d599930f7e7ab3d89783325610a597410b3f"} Dec 01 09:02:13 crc kubenswrapper[4813]: I1201 09:02:13.016447 4813 generic.go:334] "Generic (PLEG): container finished" podID="43861577-d77d-45ed-95ea-25a5dffa1b46" containerID="16e8434757ae8743d0e044be613fba36879ad53137caa53097886a44032e9949" exitCode=0 Dec 01 09:02:13 crc kubenswrapper[4813]: I1201 09:02:13.016505 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-2" event={"ID":"43861577-d77d-45ed-95ea-25a5dffa1b46","Type":"ContainerDied","Data":"16e8434757ae8743d0e044be613fba36879ad53137caa53097886a44032e9949"} Dec 01 09:02:13 crc kubenswrapper[4813]: I1201 09:02:13.017229 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-2" event={"ID":"43861577-d77d-45ed-95ea-25a5dffa1b46","Type":"ContainerDied","Data":"5b61076ea0b84ec21636b478ed4ea8081d507abb0a8eb7925b449e58e195f906"} Dec 01 09:02:13 crc kubenswrapper[4813]: I1201 09:02:13.017132 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-2" Dec 01 09:02:13 crc kubenswrapper[4813]: I1201 09:02:13.017327 4813 scope.go:117] "RemoveContainer" containerID="16e8434757ae8743d0e044be613fba36879ad53137caa53097886a44032e9949" Dec 01 09:02:13 crc kubenswrapper[4813]: I1201 09:02:13.049267 4813 scope.go:117] "RemoveContainer" containerID="72494af8d33fe162c335935341321833cde5cf4fce403b9ba7eaf72165deb984" Dec 01 09:02:13 crc kubenswrapper[4813]: I1201 09:02:13.054860 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43861577-d77d-45ed-95ea-25a5dffa1b46-config-data" (OuterVolumeSpecName: "config-data") pod "43861577-d77d-45ed-95ea-25a5dffa1b46" (UID: "43861577-d77d-45ed-95ea-25a5dffa1b46"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:02:13 crc kubenswrapper[4813]: I1201 09:02:13.073756 4813 scope.go:117] "RemoveContainer" containerID="16e8434757ae8743d0e044be613fba36879ad53137caa53097886a44032e9949" Dec 01 09:02:13 crc kubenswrapper[4813]: E1201 09:02:13.074278 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"16e8434757ae8743d0e044be613fba36879ad53137caa53097886a44032e9949\": container with ID starting with 16e8434757ae8743d0e044be613fba36879ad53137caa53097886a44032e9949 not found: ID does not exist" containerID="16e8434757ae8743d0e044be613fba36879ad53137caa53097886a44032e9949" Dec 01 09:02:13 crc kubenswrapper[4813]: I1201 09:02:13.074333 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"16e8434757ae8743d0e044be613fba36879ad53137caa53097886a44032e9949"} err="failed to get container status \"16e8434757ae8743d0e044be613fba36879ad53137caa53097886a44032e9949\": rpc error: code = NotFound desc = could not find container \"16e8434757ae8743d0e044be613fba36879ad53137caa53097886a44032e9949\": container with ID starting with 16e8434757ae8743d0e044be613fba36879ad53137caa53097886a44032e9949 not found: ID does not exist" Dec 01 09:02:13 crc kubenswrapper[4813]: I1201 09:02:13.074364 4813 scope.go:117] "RemoveContainer" containerID="72494af8d33fe162c335935341321833cde5cf4fce403b9ba7eaf72165deb984" Dec 01 09:02:13 crc kubenswrapper[4813]: E1201 09:02:13.075237 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"72494af8d33fe162c335935341321833cde5cf4fce403b9ba7eaf72165deb984\": container with ID starting with 72494af8d33fe162c335935341321833cde5cf4fce403b9ba7eaf72165deb984 not found: ID does not exist" containerID="72494af8d33fe162c335935341321833cde5cf4fce403b9ba7eaf72165deb984" Dec 01 09:02:13 crc kubenswrapper[4813]: I1201 09:02:13.075266 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"72494af8d33fe162c335935341321833cde5cf4fce403b9ba7eaf72165deb984"} err="failed to get container status \"72494af8d33fe162c335935341321833cde5cf4fce403b9ba7eaf72165deb984\": rpc error: code = NotFound desc = could not find container \"72494af8d33fe162c335935341321833cde5cf4fce403b9ba7eaf72165deb984\": container with ID starting with 72494af8d33fe162c335935341321833cde5cf4fce403b9ba7eaf72165deb984 not found: ID does not exist" Dec 01 09:02:13 crc kubenswrapper[4813]: I1201 09:02:13.081515 4813 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage13-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage13-crc\") on node \"crc\" " Dec 01 09:02:13 crc kubenswrapper[4813]: I1201 09:02:13.081545 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dhtk6\" (UniqueName: \"kubernetes.io/projected/43861577-d77d-45ed-95ea-25a5dffa1b46-kube-api-access-dhtk6\") on node \"crc\" DevicePath \"\"" Dec 01 09:02:13 crc kubenswrapper[4813]: I1201 09:02:13.081557 4813 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/43861577-d77d-45ed-95ea-25a5dffa1b46-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 09:02:13 crc kubenswrapper[4813]: I1201 09:02:13.081567 4813 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/43861577-d77d-45ed-95ea-25a5dffa1b46-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 09:02:13 crc kubenswrapper[4813]: I1201 09:02:13.081583 4813 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" " Dec 01 09:02:13 crc kubenswrapper[4813]: I1201 09:02:13.093611 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-1" Dec 01 09:02:13 crc kubenswrapper[4813]: I1201 09:02:13.094804 4813 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage12-crc" (UniqueName: "kubernetes.io/local-volume/local-storage12-crc") on node "crc" Dec 01 09:02:13 crc kubenswrapper[4813]: I1201 09:02:13.098614 4813 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage13-crc" (UniqueName: "kubernetes.io/local-volume/local-storage13-crc") on node "crc" Dec 01 09:02:13 crc kubenswrapper[4813]: I1201 09:02:13.182832 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance-cache\" (UniqueName: \"kubernetes.io/local-volume/local-storage17-crc\") pod \"7790813b-eccc-442b-a400-7fb5863599a7\" (UID: \"7790813b-eccc-442b-a400-7fb5863599a7\") " Dec 01 09:02:13 crc kubenswrapper[4813]: I1201 09:02:13.182920 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/7790813b-eccc-442b-a400-7fb5863599a7-etc-nvme\") pod \"7790813b-eccc-442b-a400-7fb5863599a7\" (UID: \"7790813b-eccc-442b-a400-7fb5863599a7\") " Dec 01 09:02:13 crc kubenswrapper[4813]: I1201 09:02:13.182985 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7790813b-eccc-442b-a400-7fb5863599a7-logs\") pod \"7790813b-eccc-442b-a400-7fb5863599a7\" (UID: \"7790813b-eccc-442b-a400-7fb5863599a7\") " Dec 01 09:02:13 crc kubenswrapper[4813]: I1201 09:02:13.183015 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/7790813b-eccc-442b-a400-7fb5863599a7-lib-modules\") pod \"7790813b-eccc-442b-a400-7fb5863599a7\" (UID: \"7790813b-eccc-442b-a400-7fb5863599a7\") " Dec 01 09:02:13 crc kubenswrapper[4813]: I1201 09:02:13.183041 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"7790813b-eccc-442b-a400-7fb5863599a7\" (UID: \"7790813b-eccc-442b-a400-7fb5863599a7\") " Dec 01 09:02:13 crc kubenswrapper[4813]: I1201 09:02:13.183091 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7790813b-eccc-442b-a400-7fb5863599a7-config-data\") pod \"7790813b-eccc-442b-a400-7fb5863599a7\" (UID: \"7790813b-eccc-442b-a400-7fb5863599a7\") " Dec 01 09:02:13 crc kubenswrapper[4813]: I1201 09:02:13.183140 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7790813b-eccc-442b-a400-7fb5863599a7-scripts\") pod \"7790813b-eccc-442b-a400-7fb5863599a7\" (UID: \"7790813b-eccc-442b-a400-7fb5863599a7\") " Dec 01 09:02:13 crc kubenswrapper[4813]: I1201 09:02:13.183166 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/7790813b-eccc-442b-a400-7fb5863599a7-sys\") pod \"7790813b-eccc-442b-a400-7fb5863599a7\" (UID: \"7790813b-eccc-442b-a400-7fb5863599a7\") " Dec 01 09:02:13 crc kubenswrapper[4813]: I1201 09:02:13.183203 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7790813b-eccc-442b-a400-7fb5863599a7-httpd-run\") pod \"7790813b-eccc-442b-a400-7fb5863599a7\" (UID: \"7790813b-eccc-442b-a400-7fb5863599a7\") " Dec 01 09:02:13 crc kubenswrapper[4813]: I1201 09:02:13.183224 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/7790813b-eccc-442b-a400-7fb5863599a7-run\") pod \"7790813b-eccc-442b-a400-7fb5863599a7\" (UID: \"7790813b-eccc-442b-a400-7fb5863599a7\") " Dec 01 09:02:13 crc kubenswrapper[4813]: I1201 09:02:13.183246 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/7790813b-eccc-442b-a400-7fb5863599a7-etc-iscsi\") pod \"7790813b-eccc-442b-a400-7fb5863599a7\" (UID: \"7790813b-eccc-442b-a400-7fb5863599a7\") " Dec 01 09:02:13 crc kubenswrapper[4813]: I1201 09:02:13.183279 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tkp6h\" (UniqueName: \"kubernetes.io/projected/7790813b-eccc-442b-a400-7fb5863599a7-kube-api-access-tkp6h\") pod \"7790813b-eccc-442b-a400-7fb5863599a7\" (UID: \"7790813b-eccc-442b-a400-7fb5863599a7\") " Dec 01 09:02:13 crc kubenswrapper[4813]: I1201 09:02:13.183302 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/7790813b-eccc-442b-a400-7fb5863599a7-var-locks-brick\") pod \"7790813b-eccc-442b-a400-7fb5863599a7\" (UID: \"7790813b-eccc-442b-a400-7fb5863599a7\") " Dec 01 09:02:13 crc kubenswrapper[4813]: I1201 09:02:13.183320 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/7790813b-eccc-442b-a400-7fb5863599a7-dev\") pod \"7790813b-eccc-442b-a400-7fb5863599a7\" (UID: \"7790813b-eccc-442b-a400-7fb5863599a7\") " Dec 01 09:02:13 crc kubenswrapper[4813]: I1201 09:02:13.183580 4813 reconciler_common.go:293] "Volume detached for volume \"local-storage13-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage13-crc\") on node \"crc\" DevicePath \"\"" Dec 01 09:02:13 crc kubenswrapper[4813]: I1201 09:02:13.183596 4813 reconciler_common.go:293] "Volume detached for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" DevicePath \"\"" Dec 01 09:02:13 crc kubenswrapper[4813]: I1201 09:02:13.183748 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7790813b-eccc-442b-a400-7fb5863599a7-dev" (OuterVolumeSpecName: "dev") pod "7790813b-eccc-442b-a400-7fb5863599a7" (UID: "7790813b-eccc-442b-a400-7fb5863599a7"). InnerVolumeSpecName "dev". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 09:02:13 crc kubenswrapper[4813]: I1201 09:02:13.185300 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7790813b-eccc-442b-a400-7fb5863599a7-run" (OuterVolumeSpecName: "run") pod "7790813b-eccc-442b-a400-7fb5863599a7" (UID: "7790813b-eccc-442b-a400-7fb5863599a7"). InnerVolumeSpecName "run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 09:02:13 crc kubenswrapper[4813]: I1201 09:02:13.185383 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7790813b-eccc-442b-a400-7fb5863599a7-sys" (OuterVolumeSpecName: "sys") pod "7790813b-eccc-442b-a400-7fb5863599a7" (UID: "7790813b-eccc-442b-a400-7fb5863599a7"). InnerVolumeSpecName "sys". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 09:02:13 crc kubenswrapper[4813]: I1201 09:02:13.185389 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7790813b-eccc-442b-a400-7fb5863599a7-etc-iscsi" (OuterVolumeSpecName: "etc-iscsi") pod "7790813b-eccc-442b-a400-7fb5863599a7" (UID: "7790813b-eccc-442b-a400-7fb5863599a7"). InnerVolumeSpecName "etc-iscsi". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 09:02:13 crc kubenswrapper[4813]: I1201 09:02:13.185497 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7790813b-eccc-442b-a400-7fb5863599a7-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "7790813b-eccc-442b-a400-7fb5863599a7" (UID: "7790813b-eccc-442b-a400-7fb5863599a7"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 09:02:13 crc kubenswrapper[4813]: I1201 09:02:13.185527 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7790813b-eccc-442b-a400-7fb5863599a7-etc-nvme" (OuterVolumeSpecName: "etc-nvme") pod "7790813b-eccc-442b-a400-7fb5863599a7" (UID: "7790813b-eccc-442b-a400-7fb5863599a7"). InnerVolumeSpecName "etc-nvme". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 09:02:13 crc kubenswrapper[4813]: I1201 09:02:13.185742 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7790813b-eccc-442b-a400-7fb5863599a7-var-locks-brick" (OuterVolumeSpecName: "var-locks-brick") pod "7790813b-eccc-442b-a400-7fb5863599a7" (UID: "7790813b-eccc-442b-a400-7fb5863599a7"). InnerVolumeSpecName "var-locks-brick". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 09:02:13 crc kubenswrapper[4813]: I1201 09:02:13.186580 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage17-crc" (OuterVolumeSpecName: "glance-cache") pod "7790813b-eccc-442b-a400-7fb5863599a7" (UID: "7790813b-eccc-442b-a400-7fb5863599a7"). InnerVolumeSpecName "local-storage17-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 01 09:02:13 crc kubenswrapper[4813]: I1201 09:02:13.188307 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7790813b-eccc-442b-a400-7fb5863599a7-scripts" (OuterVolumeSpecName: "scripts") pod "7790813b-eccc-442b-a400-7fb5863599a7" (UID: "7790813b-eccc-442b-a400-7fb5863599a7"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:02:13 crc kubenswrapper[4813]: I1201 09:02:13.189151 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage08-crc" (OuterVolumeSpecName: "glance") pod "7790813b-eccc-442b-a400-7fb5863599a7" (UID: "7790813b-eccc-442b-a400-7fb5863599a7"). InnerVolumeSpecName "local-storage08-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 01 09:02:13 crc kubenswrapper[4813]: I1201 09:02:13.189365 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7790813b-eccc-442b-a400-7fb5863599a7-logs" (OuterVolumeSpecName: "logs") pod "7790813b-eccc-442b-a400-7fb5863599a7" (UID: "7790813b-eccc-442b-a400-7fb5863599a7"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 09:02:13 crc kubenswrapper[4813]: I1201 09:02:13.189500 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7790813b-eccc-442b-a400-7fb5863599a7-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "7790813b-eccc-442b-a400-7fb5863599a7" (UID: "7790813b-eccc-442b-a400-7fb5863599a7"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 09:02:13 crc kubenswrapper[4813]: I1201 09:02:13.189930 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7790813b-eccc-442b-a400-7fb5863599a7-kube-api-access-tkp6h" (OuterVolumeSpecName: "kube-api-access-tkp6h") pod "7790813b-eccc-442b-a400-7fb5863599a7" (UID: "7790813b-eccc-442b-a400-7fb5863599a7"). InnerVolumeSpecName "kube-api-access-tkp6h". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:02:13 crc kubenswrapper[4813]: I1201 09:02:13.234463 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7790813b-eccc-442b-a400-7fb5863599a7-config-data" (OuterVolumeSpecName: "config-data") pod "7790813b-eccc-442b-a400-7fb5863599a7" (UID: "7790813b-eccc-442b-a400-7fb5863599a7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:02:13 crc kubenswrapper[4813]: I1201 09:02:13.285302 4813 reconciler_common.go:293] "Volume detached for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/7790813b-eccc-442b-a400-7fb5863599a7-etc-nvme\") on node \"crc\" DevicePath \"\"" Dec 01 09:02:13 crc kubenswrapper[4813]: I1201 09:02:13.285344 4813 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7790813b-eccc-442b-a400-7fb5863599a7-logs\") on node \"crc\" DevicePath \"\"" Dec 01 09:02:13 crc kubenswrapper[4813]: I1201 09:02:13.285356 4813 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/7790813b-eccc-442b-a400-7fb5863599a7-lib-modules\") on node \"crc\" DevicePath \"\"" Dec 01 09:02:13 crc kubenswrapper[4813]: I1201 09:02:13.285415 4813 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" " Dec 01 09:02:13 crc kubenswrapper[4813]: I1201 09:02:13.285432 4813 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7790813b-eccc-442b-a400-7fb5863599a7-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 09:02:13 crc kubenswrapper[4813]: I1201 09:02:13.285440 4813 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7790813b-eccc-442b-a400-7fb5863599a7-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 09:02:13 crc kubenswrapper[4813]: I1201 09:02:13.285450 4813 reconciler_common.go:293] "Volume detached for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/7790813b-eccc-442b-a400-7fb5863599a7-sys\") on node \"crc\" DevicePath \"\"" Dec 01 09:02:13 crc kubenswrapper[4813]: I1201 09:02:13.285461 4813 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7790813b-eccc-442b-a400-7fb5863599a7-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 01 09:02:13 crc kubenswrapper[4813]: I1201 09:02:13.285471 4813 reconciler_common.go:293] "Volume detached for volume \"run\" (UniqueName: \"kubernetes.io/host-path/7790813b-eccc-442b-a400-7fb5863599a7-run\") on node \"crc\" DevicePath \"\"" Dec 01 09:02:13 crc kubenswrapper[4813]: I1201 09:02:13.285483 4813 reconciler_common.go:293] "Volume detached for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/7790813b-eccc-442b-a400-7fb5863599a7-etc-iscsi\") on node \"crc\" DevicePath \"\"" Dec 01 09:02:13 crc kubenswrapper[4813]: I1201 09:02:13.285494 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tkp6h\" (UniqueName: \"kubernetes.io/projected/7790813b-eccc-442b-a400-7fb5863599a7-kube-api-access-tkp6h\") on node \"crc\" DevicePath \"\"" Dec 01 09:02:13 crc kubenswrapper[4813]: I1201 09:02:13.285507 4813 reconciler_common.go:293] "Volume detached for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/7790813b-eccc-442b-a400-7fb5863599a7-dev\") on node \"crc\" DevicePath \"\"" Dec 01 09:02:13 crc kubenswrapper[4813]: I1201 09:02:13.285517 4813 reconciler_common.go:293] "Volume detached for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/7790813b-eccc-442b-a400-7fb5863599a7-var-locks-brick\") on node \"crc\" DevicePath \"\"" Dec 01 09:02:13 crc kubenswrapper[4813]: I1201 09:02:13.285548 4813 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage17-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage17-crc\") on node \"crc\" " Dec 01 09:02:13 crc kubenswrapper[4813]: I1201 09:02:13.298110 4813 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage08-crc" (UniqueName: "kubernetes.io/local-volume/local-storage08-crc") on node "crc" Dec 01 09:02:13 crc kubenswrapper[4813]: I1201 09:02:13.301098 4813 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage17-crc" (UniqueName: "kubernetes.io/local-volume/local-storage17-crc") on node "crc" Dec 01 09:02:13 crc kubenswrapper[4813]: I1201 09:02:13.367659 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-2"] Dec 01 09:02:13 crc kubenswrapper[4813]: I1201 09:02:13.379741 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-2"] Dec 01 09:02:13 crc kubenswrapper[4813]: I1201 09:02:13.387775 4813 reconciler_common.go:293] "Volume detached for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" DevicePath \"\"" Dec 01 09:02:13 crc kubenswrapper[4813]: I1201 09:02:13.387814 4813 reconciler_common.go:293] "Volume detached for volume \"local-storage17-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage17-crc\") on node \"crc\" DevicePath \"\"" Dec 01 09:02:13 crc kubenswrapper[4813]: I1201 09:02:13.987163 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-1" Dec 01 09:02:14 crc kubenswrapper[4813]: I1201 09:02:14.000527 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/e58d9e76-7271-47c3-a7aa-a5a802bfdb41-sys\") pod \"e58d9e76-7271-47c3-a7aa-a5a802bfdb41\" (UID: \"e58d9e76-7271-47c3-a7aa-a5a802bfdb41\") " Dec 01 09:02:14 crc kubenswrapper[4813]: I1201 09:02:14.000633 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/e58d9e76-7271-47c3-a7aa-a5a802bfdb41-etc-nvme\") pod \"e58d9e76-7271-47c3-a7aa-a5a802bfdb41\" (UID: \"e58d9e76-7271-47c3-a7aa-a5a802bfdb41\") " Dec 01 09:02:14 crc kubenswrapper[4813]: I1201 09:02:14.000645 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e58d9e76-7271-47c3-a7aa-a5a802bfdb41-sys" (OuterVolumeSpecName: "sys") pod "e58d9e76-7271-47c3-a7aa-a5a802bfdb41" (UID: "e58d9e76-7271-47c3-a7aa-a5a802bfdb41"). InnerVolumeSpecName "sys". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 09:02:14 crc kubenswrapper[4813]: I1201 09:02:14.000676 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance-cache\" (UniqueName: \"kubernetes.io/local-volume/local-storage20-crc\") pod \"e58d9e76-7271-47c3-a7aa-a5a802bfdb41\" (UID: \"e58d9e76-7271-47c3-a7aa-a5a802bfdb41\") " Dec 01 09:02:14 crc kubenswrapper[4813]: I1201 09:02:14.000694 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e58d9e76-7271-47c3-a7aa-a5a802bfdb41-etc-nvme" (OuterVolumeSpecName: "etc-nvme") pod "e58d9e76-7271-47c3-a7aa-a5a802bfdb41" (UID: "e58d9e76-7271-47c3-a7aa-a5a802bfdb41"). InnerVolumeSpecName "etc-nvme". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 09:02:14 crc kubenswrapper[4813]: I1201 09:02:14.000705 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/e58d9e76-7271-47c3-a7aa-a5a802bfdb41-var-locks-brick\") pod \"e58d9e76-7271-47c3-a7aa-a5a802bfdb41\" (UID: \"e58d9e76-7271-47c3-a7aa-a5a802bfdb41\") " Dec 01 09:02:14 crc kubenswrapper[4813]: I1201 09:02:14.000729 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/e58d9e76-7271-47c3-a7aa-a5a802bfdb41-run\") pod \"e58d9e76-7271-47c3-a7aa-a5a802bfdb41\" (UID: \"e58d9e76-7271-47c3-a7aa-a5a802bfdb41\") " Dec 01 09:02:14 crc kubenswrapper[4813]: I1201 09:02:14.000757 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e58d9e76-7271-47c3-a7aa-a5a802bfdb41-config-data\") pod \"e58d9e76-7271-47c3-a7aa-a5a802bfdb41\" (UID: \"e58d9e76-7271-47c3-a7aa-a5a802bfdb41\") " Dec 01 09:02:14 crc kubenswrapper[4813]: I1201 09:02:14.000798 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/e58d9e76-7271-47c3-a7aa-a5a802bfdb41-etc-iscsi\") pod \"e58d9e76-7271-47c3-a7aa-a5a802bfdb41\" (UID: \"e58d9e76-7271-47c3-a7aa-a5a802bfdb41\") " Dec 01 09:02:14 crc kubenswrapper[4813]: I1201 09:02:14.000822 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e58d9e76-7271-47c3-a7aa-a5a802bfdb41-logs\") pod \"e58d9e76-7271-47c3-a7aa-a5a802bfdb41\" (UID: \"e58d9e76-7271-47c3-a7aa-a5a802bfdb41\") " Dec 01 09:02:14 crc kubenswrapper[4813]: I1201 09:02:14.000822 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e58d9e76-7271-47c3-a7aa-a5a802bfdb41-var-locks-brick" (OuterVolumeSpecName: "var-locks-brick") pod "e58d9e76-7271-47c3-a7aa-a5a802bfdb41" (UID: "e58d9e76-7271-47c3-a7aa-a5a802bfdb41"). InnerVolumeSpecName "var-locks-brick". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 09:02:14 crc kubenswrapper[4813]: I1201 09:02:14.000861 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e58d9e76-7271-47c3-a7aa-a5a802bfdb41-run" (OuterVolumeSpecName: "run") pod "e58d9e76-7271-47c3-a7aa-a5a802bfdb41" (UID: "e58d9e76-7271-47c3-a7aa-a5a802bfdb41"). InnerVolumeSpecName "run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 09:02:14 crc kubenswrapper[4813]: I1201 09:02:14.000884 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e58d9e76-7271-47c3-a7aa-a5a802bfdb41-httpd-run\") pod \"e58d9e76-7271-47c3-a7aa-a5a802bfdb41\" (UID: \"e58d9e76-7271-47c3-a7aa-a5a802bfdb41\") " Dec 01 09:02:14 crc kubenswrapper[4813]: I1201 09:02:14.000942 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/e58d9e76-7271-47c3-a7aa-a5a802bfdb41-lib-modules\") pod \"e58d9e76-7271-47c3-a7aa-a5a802bfdb41\" (UID: \"e58d9e76-7271-47c3-a7aa-a5a802bfdb41\") " Dec 01 09:02:14 crc kubenswrapper[4813]: I1201 09:02:14.000996 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage14-crc\") pod \"e58d9e76-7271-47c3-a7aa-a5a802bfdb41\" (UID: \"e58d9e76-7271-47c3-a7aa-a5a802bfdb41\") " Dec 01 09:02:14 crc kubenswrapper[4813]: I1201 09:02:14.001077 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e58d9e76-7271-47c3-a7aa-a5a802bfdb41-scripts\") pod \"e58d9e76-7271-47c3-a7aa-a5a802bfdb41\" (UID: \"e58d9e76-7271-47c3-a7aa-a5a802bfdb41\") " Dec 01 09:02:14 crc kubenswrapper[4813]: I1201 09:02:14.001114 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-njccq\" (UniqueName: \"kubernetes.io/projected/e58d9e76-7271-47c3-a7aa-a5a802bfdb41-kube-api-access-njccq\") pod \"e58d9e76-7271-47c3-a7aa-a5a802bfdb41\" (UID: \"e58d9e76-7271-47c3-a7aa-a5a802bfdb41\") " Dec 01 09:02:14 crc kubenswrapper[4813]: I1201 09:02:14.001148 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/e58d9e76-7271-47c3-a7aa-a5a802bfdb41-dev\") pod \"e58d9e76-7271-47c3-a7aa-a5a802bfdb41\" (UID: \"e58d9e76-7271-47c3-a7aa-a5a802bfdb41\") " Dec 01 09:02:14 crc kubenswrapper[4813]: I1201 09:02:14.001515 4813 reconciler_common.go:293] "Volume detached for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/e58d9e76-7271-47c3-a7aa-a5a802bfdb41-sys\") on node \"crc\" DevicePath \"\"" Dec 01 09:02:14 crc kubenswrapper[4813]: I1201 09:02:14.001528 4813 reconciler_common.go:293] "Volume detached for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/e58d9e76-7271-47c3-a7aa-a5a802bfdb41-etc-nvme\") on node \"crc\" DevicePath \"\"" Dec 01 09:02:14 crc kubenswrapper[4813]: I1201 09:02:14.001537 4813 reconciler_common.go:293] "Volume detached for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/e58d9e76-7271-47c3-a7aa-a5a802bfdb41-var-locks-brick\") on node \"crc\" DevicePath \"\"" Dec 01 09:02:14 crc kubenswrapper[4813]: I1201 09:02:14.001548 4813 reconciler_common.go:293] "Volume detached for volume \"run\" (UniqueName: \"kubernetes.io/host-path/e58d9e76-7271-47c3-a7aa-a5a802bfdb41-run\") on node \"crc\" DevicePath \"\"" Dec 01 09:02:14 crc kubenswrapper[4813]: I1201 09:02:14.000885 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e58d9e76-7271-47c3-a7aa-a5a802bfdb41-etc-iscsi" (OuterVolumeSpecName: "etc-iscsi") pod "e58d9e76-7271-47c3-a7aa-a5a802bfdb41" (UID: "e58d9e76-7271-47c3-a7aa-a5a802bfdb41"). InnerVolumeSpecName "etc-iscsi". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 09:02:14 crc kubenswrapper[4813]: I1201 09:02:14.001177 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e58d9e76-7271-47c3-a7aa-a5a802bfdb41-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "e58d9e76-7271-47c3-a7aa-a5a802bfdb41" (UID: "e58d9e76-7271-47c3-a7aa-a5a802bfdb41"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 09:02:14 crc kubenswrapper[4813]: I1201 09:02:14.001231 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e58d9e76-7271-47c3-a7aa-a5a802bfdb41-logs" (OuterVolumeSpecName: "logs") pod "e58d9e76-7271-47c3-a7aa-a5a802bfdb41" (UID: "e58d9e76-7271-47c3-a7aa-a5a802bfdb41"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 09:02:14 crc kubenswrapper[4813]: I1201 09:02:14.001603 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e58d9e76-7271-47c3-a7aa-a5a802bfdb41-dev" (OuterVolumeSpecName: "dev") pod "e58d9e76-7271-47c3-a7aa-a5a802bfdb41" (UID: "e58d9e76-7271-47c3-a7aa-a5a802bfdb41"). InnerVolumeSpecName "dev". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 09:02:14 crc kubenswrapper[4813]: I1201 09:02:14.003145 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e58d9e76-7271-47c3-a7aa-a5a802bfdb41-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "e58d9e76-7271-47c3-a7aa-a5a802bfdb41" (UID: "e58d9e76-7271-47c3-a7aa-a5a802bfdb41"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 09:02:14 crc kubenswrapper[4813]: I1201 09:02:14.006787 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage20-crc" (OuterVolumeSpecName: "glance-cache") pod "e58d9e76-7271-47c3-a7aa-a5a802bfdb41" (UID: "e58d9e76-7271-47c3-a7aa-a5a802bfdb41"). InnerVolumeSpecName "local-storage20-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 01 09:02:14 crc kubenswrapper[4813]: I1201 09:02:14.007554 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage14-crc" (OuterVolumeSpecName: "glance") pod "e58d9e76-7271-47c3-a7aa-a5a802bfdb41" (UID: "e58d9e76-7271-47c3-a7aa-a5a802bfdb41"). InnerVolumeSpecName "local-storage14-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 01 09:02:14 crc kubenswrapper[4813]: I1201 09:02:14.007756 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e58d9e76-7271-47c3-a7aa-a5a802bfdb41-kube-api-access-njccq" (OuterVolumeSpecName: "kube-api-access-njccq") pod "e58d9e76-7271-47c3-a7aa-a5a802bfdb41" (UID: "e58d9e76-7271-47c3-a7aa-a5a802bfdb41"). InnerVolumeSpecName "kube-api-access-njccq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:02:14 crc kubenswrapper[4813]: I1201 09:02:14.008059 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e58d9e76-7271-47c3-a7aa-a5a802bfdb41-scripts" (OuterVolumeSpecName: "scripts") pod "e58d9e76-7271-47c3-a7aa-a5a802bfdb41" (UID: "e58d9e76-7271-47c3-a7aa-a5a802bfdb41"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:02:14 crc kubenswrapper[4813]: I1201 09:02:14.036704 4813 generic.go:334] "Generic (PLEG): container finished" podID="0bb179fc-fb0c-420a-a432-5a0d28884e23" containerID="a2ff5f4871bec4b409097a90d5929c75feb2f96c17f8cf1b75cbfa57f68a9374" exitCode=0 Dec 01 09:02:14 crc kubenswrapper[4813]: I1201 09:02:14.036807 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-2" event={"ID":"0bb179fc-fb0c-420a-a432-5a0d28884e23","Type":"ContainerDied","Data":"a2ff5f4871bec4b409097a90d5929c75feb2f96c17f8cf1b75cbfa57f68a9374"} Dec 01 09:02:14 crc kubenswrapper[4813]: I1201 09:02:14.036848 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-2" event={"ID":"0bb179fc-fb0c-420a-a432-5a0d28884e23","Type":"ContainerDied","Data":"4f734a074b74139ba6affc141d884d25fad6a85dd48c9b655f11b259d73c68a6"} Dec 01 09:02:14 crc kubenswrapper[4813]: I1201 09:02:14.036865 4813 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4f734a074b74139ba6affc141d884d25fad6a85dd48c9b655f11b259d73c68a6" Dec 01 09:02:14 crc kubenswrapper[4813]: I1201 09:02:14.047818 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-2" Dec 01 09:02:14 crc kubenswrapper[4813]: I1201 09:02:14.058207 4813 generic.go:334] "Generic (PLEG): container finished" podID="e58d9e76-7271-47c3-a7aa-a5a802bfdb41" containerID="6999bd053c70cefb09a4011183136ca0736cf8688cd73cedf3627e8fb0b87dd7" exitCode=0 Dec 01 09:02:14 crc kubenswrapper[4813]: I1201 09:02:14.058320 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-1" event={"ID":"e58d9e76-7271-47c3-a7aa-a5a802bfdb41","Type":"ContainerDied","Data":"6999bd053c70cefb09a4011183136ca0736cf8688cd73cedf3627e8fb0b87dd7"} Dec 01 09:02:14 crc kubenswrapper[4813]: I1201 09:02:14.058505 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-1" event={"ID":"e58d9e76-7271-47c3-a7aa-a5a802bfdb41","Type":"ContainerDied","Data":"a07606be2b91ee3510badb11590d95baadd2564242ac6b630cdc16395498dd27"} Dec 01 09:02:14 crc kubenswrapper[4813]: I1201 09:02:14.058573 4813 scope.go:117] "RemoveContainer" containerID="6999bd053c70cefb09a4011183136ca0736cf8688cd73cedf3627e8fb0b87dd7" Dec 01 09:02:14 crc kubenswrapper[4813]: I1201 09:02:14.058861 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-1" Dec 01 09:02:14 crc kubenswrapper[4813]: I1201 09:02:14.063354 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e58d9e76-7271-47c3-a7aa-a5a802bfdb41-config-data" (OuterVolumeSpecName: "config-data") pod "e58d9e76-7271-47c3-a7aa-a5a802bfdb41" (UID: "e58d9e76-7271-47c3-a7aa-a5a802bfdb41"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:02:14 crc kubenswrapper[4813]: I1201 09:02:14.075451 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-1" event={"ID":"7790813b-eccc-442b-a400-7fb5863599a7","Type":"ContainerDied","Data":"40dd0b7da1121b814b593d0717d87acc08858d5045aebf43e888488cafad2f25"} Dec 01 09:02:14 crc kubenswrapper[4813]: I1201 09:02:14.075563 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-1" Dec 01 09:02:14 crc kubenswrapper[4813]: I1201 09:02:14.096521 4813 scope.go:117] "RemoveContainer" containerID="61191cf315a8ec0efc4e03441826363ac77f4ce803b8cd8107d188390367161b" Dec 01 09:02:14 crc kubenswrapper[4813]: I1201 09:02:14.102801 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/0bb179fc-fb0c-420a-a432-5a0d28884e23-var-locks-brick\") pod \"0bb179fc-fb0c-420a-a432-5a0d28884e23\" (UID: \"0bb179fc-fb0c-420a-a432-5a0d28884e23\") " Dec 01 09:02:14 crc kubenswrapper[4813]: I1201 09:02:14.103115 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0bb179fc-fb0c-420a-a432-5a0d28884e23-var-locks-brick" (OuterVolumeSpecName: "var-locks-brick") pod "0bb179fc-fb0c-420a-a432-5a0d28884e23" (UID: "0bb179fc-fb0c-420a-a432-5a0d28884e23"). InnerVolumeSpecName "var-locks-brick". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 09:02:14 crc kubenswrapper[4813]: I1201 09:02:14.106639 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/0bb179fc-fb0c-420a-a432-5a0d28884e23-sys\") pod \"0bb179fc-fb0c-420a-a432-5a0d28884e23\" (UID: \"0bb179fc-fb0c-420a-a432-5a0d28884e23\") " Dec 01 09:02:14 crc kubenswrapper[4813]: I1201 09:02:14.109143 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/0bb179fc-fb0c-420a-a432-5a0d28884e23-etc-nvme\") pod \"0bb179fc-fb0c-420a-a432-5a0d28884e23\" (UID: \"0bb179fc-fb0c-420a-a432-5a0d28884e23\") " Dec 01 09:02:14 crc kubenswrapper[4813]: I1201 09:02:14.109191 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/0bb179fc-fb0c-420a-a432-5a0d28884e23-run\") pod \"0bb179fc-fb0c-420a-a432-5a0d28884e23\" (UID: \"0bb179fc-fb0c-420a-a432-5a0d28884e23\") " Dec 01 09:02:14 crc kubenswrapper[4813]: I1201 09:02:14.109238 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0bb179fc-fb0c-420a-a432-5a0d28884e23-config-data\") pod \"0bb179fc-fb0c-420a-a432-5a0d28884e23\" (UID: \"0bb179fc-fb0c-420a-a432-5a0d28884e23\") " Dec 01 09:02:14 crc kubenswrapper[4813]: I1201 09:02:14.109266 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cdzkz\" (UniqueName: \"kubernetes.io/projected/0bb179fc-fb0c-420a-a432-5a0d28884e23-kube-api-access-cdzkz\") pod \"0bb179fc-fb0c-420a-a432-5a0d28884e23\" (UID: \"0bb179fc-fb0c-420a-a432-5a0d28884e23\") " Dec 01 09:02:14 crc kubenswrapper[4813]: I1201 09:02:14.109316 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/0bb179fc-fb0c-420a-a432-5a0d28884e23-lib-modules\") pod \"0bb179fc-fb0c-420a-a432-5a0d28884e23\" (UID: \"0bb179fc-fb0c-420a-a432-5a0d28884e23\") " Dec 01 09:02:14 crc kubenswrapper[4813]: I1201 09:02:14.109378 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/0bb179fc-fb0c-420a-a432-5a0d28884e23-httpd-run\") pod \"0bb179fc-fb0c-420a-a432-5a0d28884e23\" (UID: \"0bb179fc-fb0c-420a-a432-5a0d28884e23\") " Dec 01 09:02:14 crc kubenswrapper[4813]: I1201 09:02:14.109399 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0bb179fc-fb0c-420a-a432-5a0d28884e23-scripts\") pod \"0bb179fc-fb0c-420a-a432-5a0d28884e23\" (UID: \"0bb179fc-fb0c-420a-a432-5a0d28884e23\") " Dec 01 09:02:14 crc kubenswrapper[4813]: I1201 09:02:14.109422 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0bb179fc-fb0c-420a-a432-5a0d28884e23-logs\") pod \"0bb179fc-fb0c-420a-a432-5a0d28884e23\" (UID: \"0bb179fc-fb0c-420a-a432-5a0d28884e23\") " Dec 01 09:02:14 crc kubenswrapper[4813]: I1201 09:02:14.109451 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/0bb179fc-fb0c-420a-a432-5a0d28884e23-etc-iscsi\") pod \"0bb179fc-fb0c-420a-a432-5a0d28884e23\" (UID: \"0bb179fc-fb0c-420a-a432-5a0d28884e23\") " Dec 01 09:02:14 crc kubenswrapper[4813]: I1201 09:02:14.109500 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"0bb179fc-fb0c-420a-a432-5a0d28884e23\" (UID: \"0bb179fc-fb0c-420a-a432-5a0d28884e23\") " Dec 01 09:02:14 crc kubenswrapper[4813]: I1201 09:02:14.109626 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/0bb179fc-fb0c-420a-a432-5a0d28884e23-dev\") pod \"0bb179fc-fb0c-420a-a432-5a0d28884e23\" (UID: \"0bb179fc-fb0c-420a-a432-5a0d28884e23\") " Dec 01 09:02:14 crc kubenswrapper[4813]: I1201 09:02:14.109754 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance-cache\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"0bb179fc-fb0c-420a-a432-5a0d28884e23\" (UID: \"0bb179fc-fb0c-420a-a432-5a0d28884e23\") " Dec 01 09:02:14 crc kubenswrapper[4813]: I1201 09:02:14.107184 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0bb179fc-fb0c-420a-a432-5a0d28884e23-sys" (OuterVolumeSpecName: "sys") pod "0bb179fc-fb0c-420a-a432-5a0d28884e23" (UID: "0bb179fc-fb0c-420a-a432-5a0d28884e23"). InnerVolumeSpecName "sys". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 09:02:14 crc kubenswrapper[4813]: I1201 09:02:14.110291 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0bb179fc-fb0c-420a-a432-5a0d28884e23-etc-nvme" (OuterVolumeSpecName: "etc-nvme") pod "0bb179fc-fb0c-420a-a432-5a0d28884e23" (UID: "0bb179fc-fb0c-420a-a432-5a0d28884e23"). InnerVolumeSpecName "etc-nvme". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 09:02:14 crc kubenswrapper[4813]: I1201 09:02:14.110350 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0bb179fc-fb0c-420a-a432-5a0d28884e23-run" (OuterVolumeSpecName: "run") pod "0bb179fc-fb0c-420a-a432-5a0d28884e23" (UID: "0bb179fc-fb0c-420a-a432-5a0d28884e23"). InnerVolumeSpecName "run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 09:02:14 crc kubenswrapper[4813]: I1201 09:02:14.110740 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0bb179fc-fb0c-420a-a432-5a0d28884e23-etc-iscsi" (OuterVolumeSpecName: "etc-iscsi") pod "0bb179fc-fb0c-420a-a432-5a0d28884e23" (UID: "0bb179fc-fb0c-420a-a432-5a0d28884e23"). InnerVolumeSpecName "etc-iscsi". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 09:02:14 crc kubenswrapper[4813]: I1201 09:02:14.110807 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0bb179fc-fb0c-420a-a432-5a0d28884e23-dev" (OuterVolumeSpecName: "dev") pod "0bb179fc-fb0c-420a-a432-5a0d28884e23" (UID: "0bb179fc-fb0c-420a-a432-5a0d28884e23"). InnerVolumeSpecName "dev". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 09:02:14 crc kubenswrapper[4813]: I1201 09:02:14.110810 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0bb179fc-fb0c-420a-a432-5a0d28884e23-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "0bb179fc-fb0c-420a-a432-5a0d28884e23" (UID: "0bb179fc-fb0c-420a-a432-5a0d28884e23"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 09:02:14 crc kubenswrapper[4813]: I1201 09:02:14.111164 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0bb179fc-fb0c-420a-a432-5a0d28884e23-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "0bb179fc-fb0c-420a-a432-5a0d28884e23" (UID: "0bb179fc-fb0c-420a-a432-5a0d28884e23"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 09:02:14 crc kubenswrapper[4813]: I1201 09:02:14.111551 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0bb179fc-fb0c-420a-a432-5a0d28884e23-logs" (OuterVolumeSpecName: "logs") pod "0bb179fc-fb0c-420a-a432-5a0d28884e23" (UID: "0bb179fc-fb0c-420a-a432-5a0d28884e23"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 09:02:14 crc kubenswrapper[4813]: I1201 09:02:14.114390 4813 reconciler_common.go:293] "Volume detached for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/0bb179fc-fb0c-420a-a432-5a0d28884e23-sys\") on node \"crc\" DevicePath \"\"" Dec 01 09:02:14 crc kubenswrapper[4813]: I1201 09:02:14.114424 4813 reconciler_common.go:293] "Volume detached for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/0bb179fc-fb0c-420a-a432-5a0d28884e23-etc-nvme\") on node \"crc\" DevicePath \"\"" Dec 01 09:02:14 crc kubenswrapper[4813]: I1201 09:02:14.114456 4813 reconciler_common.go:293] "Volume detached for volume \"run\" (UniqueName: \"kubernetes.io/host-path/0bb179fc-fb0c-420a-a432-5a0d28884e23-run\") on node \"crc\" DevicePath \"\"" Dec 01 09:02:14 crc kubenswrapper[4813]: I1201 09:02:14.114624 4813 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage20-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage20-crc\") on node \"crc\" " Dec 01 09:02:14 crc kubenswrapper[4813]: I1201 09:02:14.114651 4813 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e58d9e76-7271-47c3-a7aa-a5a802bfdb41-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 09:02:14 crc kubenswrapper[4813]: I1201 09:02:14.114665 4813 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/0bb179fc-fb0c-420a-a432-5a0d28884e23-lib-modules\") on node \"crc\" DevicePath \"\"" Dec 01 09:02:14 crc kubenswrapper[4813]: I1201 09:02:14.114678 4813 reconciler_common.go:293] "Volume detached for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/e58d9e76-7271-47c3-a7aa-a5a802bfdb41-etc-iscsi\") on node \"crc\" DevicePath \"\"" Dec 01 09:02:14 crc kubenswrapper[4813]: I1201 09:02:14.114700 4813 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e58d9e76-7271-47c3-a7aa-a5a802bfdb41-logs\") on node \"crc\" DevicePath \"\"" Dec 01 09:02:14 crc kubenswrapper[4813]: I1201 09:02:14.114742 4813 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e58d9e76-7271-47c3-a7aa-a5a802bfdb41-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 01 09:02:14 crc kubenswrapper[4813]: I1201 09:02:14.114756 4813 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/0bb179fc-fb0c-420a-a432-5a0d28884e23-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 01 09:02:14 crc kubenswrapper[4813]: I1201 09:02:14.114766 4813 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0bb179fc-fb0c-420a-a432-5a0d28884e23-logs\") on node \"crc\" DevicePath \"\"" Dec 01 09:02:14 crc kubenswrapper[4813]: I1201 09:02:14.114779 4813 reconciler_common.go:293] "Volume detached for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/0bb179fc-fb0c-420a-a432-5a0d28884e23-etc-iscsi\") on node \"crc\" DevicePath \"\"" Dec 01 09:02:14 crc kubenswrapper[4813]: I1201 09:02:14.114791 4813 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/e58d9e76-7271-47c3-a7aa-a5a802bfdb41-lib-modules\") on node \"crc\" DevicePath \"\"" Dec 01 09:02:14 crc kubenswrapper[4813]: I1201 09:02:14.114811 4813 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage14-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage14-crc\") on node \"crc\" " Dec 01 09:02:14 crc kubenswrapper[4813]: I1201 09:02:14.114823 4813 reconciler_common.go:293] "Volume detached for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/0bb179fc-fb0c-420a-a432-5a0d28884e23-dev\") on node \"crc\" DevicePath \"\"" Dec 01 09:02:14 crc kubenswrapper[4813]: I1201 09:02:14.114834 4813 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e58d9e76-7271-47c3-a7aa-a5a802bfdb41-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 09:02:14 crc kubenswrapper[4813]: I1201 09:02:14.114848 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-njccq\" (UniqueName: \"kubernetes.io/projected/e58d9e76-7271-47c3-a7aa-a5a802bfdb41-kube-api-access-njccq\") on node \"crc\" DevicePath \"\"" Dec 01 09:02:14 crc kubenswrapper[4813]: I1201 09:02:14.114859 4813 reconciler_common.go:293] "Volume detached for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/e58d9e76-7271-47c3-a7aa-a5a802bfdb41-dev\") on node \"crc\" DevicePath \"\"" Dec 01 09:02:14 crc kubenswrapper[4813]: I1201 09:02:14.114870 4813 reconciler_common.go:293] "Volume detached for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/0bb179fc-fb0c-420a-a432-5a0d28884e23-var-locks-brick\") on node \"crc\" DevicePath \"\"" Dec 01 09:02:14 crc kubenswrapper[4813]: I1201 09:02:14.117534 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-1"] Dec 01 09:02:14 crc kubenswrapper[4813]: I1201 09:02:14.126658 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage01-crc" (OuterVolumeSpecName: "glance-cache") pod "0bb179fc-fb0c-420a-a432-5a0d28884e23" (UID: "0bb179fc-fb0c-420a-a432-5a0d28884e23"). InnerVolumeSpecName "local-storage01-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 01 09:02:14 crc kubenswrapper[4813]: I1201 09:02:14.136559 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0bb179fc-fb0c-420a-a432-5a0d28884e23-kube-api-access-cdzkz" (OuterVolumeSpecName: "kube-api-access-cdzkz") pod "0bb179fc-fb0c-420a-a432-5a0d28884e23" (UID: "0bb179fc-fb0c-420a-a432-5a0d28884e23"). InnerVolumeSpecName "kube-api-access-cdzkz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:02:14 crc kubenswrapper[4813]: I1201 09:02:14.140267 4813 scope.go:117] "RemoveContainer" containerID="6999bd053c70cefb09a4011183136ca0736cf8688cd73cedf3627e8fb0b87dd7" Dec 01 09:02:14 crc kubenswrapper[4813]: I1201 09:02:14.140771 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0bb179fc-fb0c-420a-a432-5a0d28884e23-scripts" (OuterVolumeSpecName: "scripts") pod "0bb179fc-fb0c-420a-a432-5a0d28884e23" (UID: "0bb179fc-fb0c-420a-a432-5a0d28884e23"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:02:14 crc kubenswrapper[4813]: E1201 09:02:14.141094 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6999bd053c70cefb09a4011183136ca0736cf8688cd73cedf3627e8fb0b87dd7\": container with ID starting with 6999bd053c70cefb09a4011183136ca0736cf8688cd73cedf3627e8fb0b87dd7 not found: ID does not exist" containerID="6999bd053c70cefb09a4011183136ca0736cf8688cd73cedf3627e8fb0b87dd7" Dec 01 09:02:14 crc kubenswrapper[4813]: I1201 09:02:14.141141 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6999bd053c70cefb09a4011183136ca0736cf8688cd73cedf3627e8fb0b87dd7"} err="failed to get container status \"6999bd053c70cefb09a4011183136ca0736cf8688cd73cedf3627e8fb0b87dd7\": rpc error: code = NotFound desc = could not find container \"6999bd053c70cefb09a4011183136ca0736cf8688cd73cedf3627e8fb0b87dd7\": container with ID starting with 6999bd053c70cefb09a4011183136ca0736cf8688cd73cedf3627e8fb0b87dd7 not found: ID does not exist" Dec 01 09:02:14 crc kubenswrapper[4813]: I1201 09:02:14.141278 4813 scope.go:117] "RemoveContainer" containerID="61191cf315a8ec0efc4e03441826363ac77f4ce803b8cd8107d188390367161b" Dec 01 09:02:14 crc kubenswrapper[4813]: I1201 09:02:14.141279 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage05-crc" (OuterVolumeSpecName: "glance") pod "0bb179fc-fb0c-420a-a432-5a0d28884e23" (UID: "0bb179fc-fb0c-420a-a432-5a0d28884e23"). InnerVolumeSpecName "local-storage05-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 01 09:02:14 crc kubenswrapper[4813]: E1201 09:02:14.141738 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"61191cf315a8ec0efc4e03441826363ac77f4ce803b8cd8107d188390367161b\": container with ID starting with 61191cf315a8ec0efc4e03441826363ac77f4ce803b8cd8107d188390367161b not found: ID does not exist" containerID="61191cf315a8ec0efc4e03441826363ac77f4ce803b8cd8107d188390367161b" Dec 01 09:02:14 crc kubenswrapper[4813]: I1201 09:02:14.141776 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"61191cf315a8ec0efc4e03441826363ac77f4ce803b8cd8107d188390367161b"} err="failed to get container status \"61191cf315a8ec0efc4e03441826363ac77f4ce803b8cd8107d188390367161b\": rpc error: code = NotFound desc = could not find container \"61191cf315a8ec0efc4e03441826363ac77f4ce803b8cd8107d188390367161b\": container with ID starting with 61191cf315a8ec0efc4e03441826363ac77f4ce803b8cd8107d188390367161b not found: ID does not exist" Dec 01 09:02:14 crc kubenswrapper[4813]: I1201 09:02:14.141789 4813 scope.go:117] "RemoveContainer" containerID="d1cabd26ab94cbb2f1d690a779a0d599930f7e7ab3d89783325610a597410b3f" Dec 01 09:02:14 crc kubenswrapper[4813]: I1201 09:02:14.152866 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-1"] Dec 01 09:02:14 crc kubenswrapper[4813]: I1201 09:02:14.171315 4813 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage20-crc" (UniqueName: "kubernetes.io/local-volume/local-storage20-crc") on node "crc" Dec 01 09:02:14 crc kubenswrapper[4813]: I1201 09:02:14.173124 4813 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage14-crc" (UniqueName: "kubernetes.io/local-volume/local-storage14-crc") on node "crc" Dec 01 09:02:14 crc kubenswrapper[4813]: I1201 09:02:14.174214 4813 scope.go:117] "RemoveContainer" containerID="a1e8ceb6fe1d2778c4ba07d2f6328f0b7a52d25bf8d2de9445e17eb86a2d71ab" Dec 01 09:02:14 crc kubenswrapper[4813]: I1201 09:02:14.185320 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0bb179fc-fb0c-420a-a432-5a0d28884e23-config-data" (OuterVolumeSpecName: "config-data") pod "0bb179fc-fb0c-420a-a432-5a0d28884e23" (UID: "0bb179fc-fb0c-420a-a432-5a0d28884e23"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:02:14 crc kubenswrapper[4813]: I1201 09:02:14.217479 4813 reconciler_common.go:293] "Volume detached for volume \"local-storage20-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage20-crc\") on node \"crc\" DevicePath \"\"" Dec 01 09:02:14 crc kubenswrapper[4813]: I1201 09:02:14.217520 4813 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0bb179fc-fb0c-420a-a432-5a0d28884e23-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 09:02:14 crc kubenswrapper[4813]: I1201 09:02:14.217534 4813 reconciler_common.go:293] "Volume detached for volume \"local-storage14-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage14-crc\") on node \"crc\" DevicePath \"\"" Dec 01 09:02:14 crc kubenswrapper[4813]: I1201 09:02:14.217580 4813 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" " Dec 01 09:02:14 crc kubenswrapper[4813]: I1201 09:02:14.217599 4813 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" " Dec 01 09:02:14 crc kubenswrapper[4813]: I1201 09:02:14.217612 4813 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0bb179fc-fb0c-420a-a432-5a0d28884e23-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 09:02:14 crc kubenswrapper[4813]: I1201 09:02:14.217626 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cdzkz\" (UniqueName: \"kubernetes.io/projected/0bb179fc-fb0c-420a-a432-5a0d28884e23-kube-api-access-cdzkz\") on node \"crc\" DevicePath \"\"" Dec 01 09:02:14 crc kubenswrapper[4813]: I1201 09:02:14.232964 4813 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage01-crc" (UniqueName: "kubernetes.io/local-volume/local-storage01-crc") on node "crc" Dec 01 09:02:14 crc kubenswrapper[4813]: I1201 09:02:14.246044 4813 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage05-crc" (UniqueName: "kubernetes.io/local-volume/local-storage05-crc") on node "crc" Dec 01 09:02:14 crc kubenswrapper[4813]: I1201 09:02:14.319333 4813 reconciler_common.go:293] "Volume detached for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" DevicePath \"\"" Dec 01 09:02:14 crc kubenswrapper[4813]: I1201 09:02:14.319390 4813 reconciler_common.go:293] "Volume detached for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" DevicePath \"\"" Dec 01 09:02:14 crc kubenswrapper[4813]: I1201 09:02:14.400942 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43861577-d77d-45ed-95ea-25a5dffa1b46" path="/var/lib/kubelet/pods/43861577-d77d-45ed-95ea-25a5dffa1b46/volumes" Dec 01 09:02:14 crc kubenswrapper[4813]: I1201 09:02:14.401586 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7790813b-eccc-442b-a400-7fb5863599a7" path="/var/lib/kubelet/pods/7790813b-eccc-442b-a400-7fb5863599a7/volumes" Dec 01 09:02:14 crc kubenswrapper[4813]: I1201 09:02:14.402189 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-1"] Dec 01 09:02:14 crc kubenswrapper[4813]: I1201 09:02:14.403299 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-1"] Dec 01 09:02:15 crc kubenswrapper[4813]: I1201 09:02:15.086315 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-2" Dec 01 09:02:15 crc kubenswrapper[4813]: I1201 09:02:15.111896 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-2"] Dec 01 09:02:15 crc kubenswrapper[4813]: I1201 09:02:15.118581 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-2"] Dec 01 09:02:15 crc kubenswrapper[4813]: I1201 09:02:15.769601 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-0"] Dec 01 09:02:15 crc kubenswrapper[4813]: I1201 09:02:15.769893 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-external-api-0" podUID="7b8bf821-52ae-409b-8e17-fc8e1302ed10" containerName="glance-log" containerID="cri-o://56c363c42bb8f1be549639c72070e76cf11757d4ca387266e8486456ffae3383" gracePeriod=30 Dec 01 09:02:15 crc kubenswrapper[4813]: I1201 09:02:15.770081 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-external-api-0" podUID="7b8bf821-52ae-409b-8e17-fc8e1302ed10" containerName="glance-httpd" containerID="cri-o://a22c275b1abdf242357c033d2a976a1e0bf84f3b733c91633972190ba9de2c51" gracePeriod=30 Dec 01 09:02:16 crc kubenswrapper[4813]: I1201 09:02:16.096714 4813 generic.go:334] "Generic (PLEG): container finished" podID="7b8bf821-52ae-409b-8e17-fc8e1302ed10" containerID="56c363c42bb8f1be549639c72070e76cf11757d4ca387266e8486456ffae3383" exitCode=143 Dec 01 09:02:16 crc kubenswrapper[4813]: I1201 09:02:16.096846 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-0" event={"ID":"7b8bf821-52ae-409b-8e17-fc8e1302ed10","Type":"ContainerDied","Data":"56c363c42bb8f1be549639c72070e76cf11757d4ca387266e8486456ffae3383"} Dec 01 09:02:16 crc kubenswrapper[4813]: I1201 09:02:16.200540 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Dec 01 09:02:16 crc kubenswrapper[4813]: I1201 09:02:16.200875 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-internal-api-0" podUID="a3fd4804-b7ff-42d4-aa58-999cb59efbc7" containerName="glance-log" containerID="cri-o://6781e4dd9f6569955b120e287571b0574a25a313f8bbb4c49c326fc1bd841f7d" gracePeriod=30 Dec 01 09:02:16 crc kubenswrapper[4813]: I1201 09:02:16.200989 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-internal-api-0" podUID="a3fd4804-b7ff-42d4-aa58-999cb59efbc7" containerName="glance-httpd" containerID="cri-o://d89dec11e5c1011779b313ec517b6cceb51e641055e65c21caf627d7b47703b2" gracePeriod=30 Dec 01 09:02:16 crc kubenswrapper[4813]: I1201 09:02:16.404524 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0bb179fc-fb0c-420a-a432-5a0d28884e23" path="/var/lib/kubelet/pods/0bb179fc-fb0c-420a-a432-5a0d28884e23/volumes" Dec 01 09:02:16 crc kubenswrapper[4813]: I1201 09:02:16.405377 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e58d9e76-7271-47c3-a7aa-a5a802bfdb41" path="/var/lib/kubelet/pods/e58d9e76-7271-47c3-a7aa-a5a802bfdb41/volumes" Dec 01 09:02:17 crc kubenswrapper[4813]: I1201 09:02:17.109103 4813 generic.go:334] "Generic (PLEG): container finished" podID="a3fd4804-b7ff-42d4-aa58-999cb59efbc7" containerID="6781e4dd9f6569955b120e287571b0574a25a313f8bbb4c49c326fc1bd841f7d" exitCode=143 Dec 01 09:02:17 crc kubenswrapper[4813]: I1201 09:02:17.109412 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"a3fd4804-b7ff-42d4-aa58-999cb59efbc7","Type":"ContainerDied","Data":"6781e4dd9f6569955b120e287571b0574a25a313f8bbb4c49c326fc1bd841f7d"} Dec 01 09:02:19 crc kubenswrapper[4813]: I1201 09:02:19.130005 4813 generic.go:334] "Generic (PLEG): container finished" podID="7b8bf821-52ae-409b-8e17-fc8e1302ed10" containerID="a22c275b1abdf242357c033d2a976a1e0bf84f3b733c91633972190ba9de2c51" exitCode=0 Dec 01 09:02:19 crc kubenswrapper[4813]: I1201 09:02:19.130096 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-0" event={"ID":"7b8bf821-52ae-409b-8e17-fc8e1302ed10","Type":"ContainerDied","Data":"a22c275b1abdf242357c033d2a976a1e0bf84f3b733c91633972190ba9de2c51"} Dec 01 09:02:19 crc kubenswrapper[4813]: I1201 09:02:19.294519 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-0" Dec 01 09:02:19 crc kubenswrapper[4813]: I1201 09:02:19.397374 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/7b8bf821-52ae-409b-8e17-fc8e1302ed10-sys\") pod \"7b8bf821-52ae-409b-8e17-fc8e1302ed10\" (UID: \"7b8bf821-52ae-409b-8e17-fc8e1302ed10\") " Dec 01 09:02:19 crc kubenswrapper[4813]: I1201 09:02:19.397418 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/7b8bf821-52ae-409b-8e17-fc8e1302ed10-lib-modules\") pod \"7b8bf821-52ae-409b-8e17-fc8e1302ed10\" (UID: \"7b8bf821-52ae-409b-8e17-fc8e1302ed10\") " Dec 01 09:02:19 crc kubenswrapper[4813]: I1201 09:02:19.397475 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-km49c\" (UniqueName: \"kubernetes.io/projected/7b8bf821-52ae-409b-8e17-fc8e1302ed10-kube-api-access-km49c\") pod \"7b8bf821-52ae-409b-8e17-fc8e1302ed10\" (UID: \"7b8bf821-52ae-409b-8e17-fc8e1302ed10\") " Dec 01 09:02:19 crc kubenswrapper[4813]: I1201 09:02:19.397505 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/7b8bf821-52ae-409b-8e17-fc8e1302ed10-var-locks-brick\") pod \"7b8bf821-52ae-409b-8e17-fc8e1302ed10\" (UID: \"7b8bf821-52ae-409b-8e17-fc8e1302ed10\") " Dec 01 09:02:19 crc kubenswrapper[4813]: I1201 09:02:19.397540 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7b8bf821-52ae-409b-8e17-fc8e1302ed10-config-data\") pod \"7b8bf821-52ae-409b-8e17-fc8e1302ed10\" (UID: \"7b8bf821-52ae-409b-8e17-fc8e1302ed10\") " Dec 01 09:02:19 crc kubenswrapper[4813]: I1201 09:02:19.397579 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/7b8bf821-52ae-409b-8e17-fc8e1302ed10-dev\") pod \"7b8bf821-52ae-409b-8e17-fc8e1302ed10\" (UID: \"7b8bf821-52ae-409b-8e17-fc8e1302ed10\") " Dec 01 09:02:19 crc kubenswrapper[4813]: I1201 09:02:19.397612 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/7b8bf821-52ae-409b-8e17-fc8e1302ed10-etc-iscsi\") pod \"7b8bf821-52ae-409b-8e17-fc8e1302ed10\" (UID: \"7b8bf821-52ae-409b-8e17-fc8e1302ed10\") " Dec 01 09:02:19 crc kubenswrapper[4813]: I1201 09:02:19.397632 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/7b8bf821-52ae-409b-8e17-fc8e1302ed10-run\") pod \"7b8bf821-52ae-409b-8e17-fc8e1302ed10\" (UID: \"7b8bf821-52ae-409b-8e17-fc8e1302ed10\") " Dec 01 09:02:19 crc kubenswrapper[4813]: I1201 09:02:19.397673 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7b8bf821-52ae-409b-8e17-fc8e1302ed10-httpd-run\") pod \"7b8bf821-52ae-409b-8e17-fc8e1302ed10\" (UID: \"7b8bf821-52ae-409b-8e17-fc8e1302ed10\") " Dec 01 09:02:19 crc kubenswrapper[4813]: I1201 09:02:19.397728 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage16-crc\") pod \"7b8bf821-52ae-409b-8e17-fc8e1302ed10\" (UID: \"7b8bf821-52ae-409b-8e17-fc8e1302ed10\") " Dec 01 09:02:19 crc kubenswrapper[4813]: I1201 09:02:19.397749 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7b8bf821-52ae-409b-8e17-fc8e1302ed10-logs\") pod \"7b8bf821-52ae-409b-8e17-fc8e1302ed10\" (UID: \"7b8bf821-52ae-409b-8e17-fc8e1302ed10\") " Dec 01 09:02:19 crc kubenswrapper[4813]: I1201 09:02:19.397779 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance-cache\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"7b8bf821-52ae-409b-8e17-fc8e1302ed10\" (UID: \"7b8bf821-52ae-409b-8e17-fc8e1302ed10\") " Dec 01 09:02:19 crc kubenswrapper[4813]: I1201 09:02:19.397818 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7b8bf821-52ae-409b-8e17-fc8e1302ed10-scripts\") pod \"7b8bf821-52ae-409b-8e17-fc8e1302ed10\" (UID: \"7b8bf821-52ae-409b-8e17-fc8e1302ed10\") " Dec 01 09:02:19 crc kubenswrapper[4813]: I1201 09:02:19.397845 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/7b8bf821-52ae-409b-8e17-fc8e1302ed10-etc-nvme\") pod \"7b8bf821-52ae-409b-8e17-fc8e1302ed10\" (UID: \"7b8bf821-52ae-409b-8e17-fc8e1302ed10\") " Dec 01 09:02:19 crc kubenswrapper[4813]: I1201 09:02:19.398093 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7b8bf821-52ae-409b-8e17-fc8e1302ed10-etc-iscsi" (OuterVolumeSpecName: "etc-iscsi") pod "7b8bf821-52ae-409b-8e17-fc8e1302ed10" (UID: "7b8bf821-52ae-409b-8e17-fc8e1302ed10"). InnerVolumeSpecName "etc-iscsi". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 09:02:19 crc kubenswrapper[4813]: I1201 09:02:19.398143 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7b8bf821-52ae-409b-8e17-fc8e1302ed10-sys" (OuterVolumeSpecName: "sys") pod "7b8bf821-52ae-409b-8e17-fc8e1302ed10" (UID: "7b8bf821-52ae-409b-8e17-fc8e1302ed10"). InnerVolumeSpecName "sys". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 09:02:19 crc kubenswrapper[4813]: I1201 09:02:19.398152 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7b8bf821-52ae-409b-8e17-fc8e1302ed10-etc-nvme" (OuterVolumeSpecName: "etc-nvme") pod "7b8bf821-52ae-409b-8e17-fc8e1302ed10" (UID: "7b8bf821-52ae-409b-8e17-fc8e1302ed10"). InnerVolumeSpecName "etc-nvme". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 09:02:19 crc kubenswrapper[4813]: I1201 09:02:19.398178 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7b8bf821-52ae-409b-8e17-fc8e1302ed10-var-locks-brick" (OuterVolumeSpecName: "var-locks-brick") pod "7b8bf821-52ae-409b-8e17-fc8e1302ed10" (UID: "7b8bf821-52ae-409b-8e17-fc8e1302ed10"). InnerVolumeSpecName "var-locks-brick". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 09:02:19 crc kubenswrapper[4813]: I1201 09:02:19.398480 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7b8bf821-52ae-409b-8e17-fc8e1302ed10-logs" (OuterVolumeSpecName: "logs") pod "7b8bf821-52ae-409b-8e17-fc8e1302ed10" (UID: "7b8bf821-52ae-409b-8e17-fc8e1302ed10"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 09:02:19 crc kubenswrapper[4813]: I1201 09:02:19.398519 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7b8bf821-52ae-409b-8e17-fc8e1302ed10-run" (OuterVolumeSpecName: "run") pod "7b8bf821-52ae-409b-8e17-fc8e1302ed10" (UID: "7b8bf821-52ae-409b-8e17-fc8e1302ed10"). InnerVolumeSpecName "run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 09:02:19 crc kubenswrapper[4813]: I1201 09:02:19.398539 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7b8bf821-52ae-409b-8e17-fc8e1302ed10-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "7b8bf821-52ae-409b-8e17-fc8e1302ed10" (UID: "7b8bf821-52ae-409b-8e17-fc8e1302ed10"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 09:02:19 crc kubenswrapper[4813]: I1201 09:02:19.398713 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7b8bf821-52ae-409b-8e17-fc8e1302ed10-dev" (OuterVolumeSpecName: "dev") pod "7b8bf821-52ae-409b-8e17-fc8e1302ed10" (UID: "7b8bf821-52ae-409b-8e17-fc8e1302ed10"). InnerVolumeSpecName "dev". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 09:02:19 crc kubenswrapper[4813]: I1201 09:02:19.398804 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7b8bf821-52ae-409b-8e17-fc8e1302ed10-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "7b8bf821-52ae-409b-8e17-fc8e1302ed10" (UID: "7b8bf821-52ae-409b-8e17-fc8e1302ed10"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 09:02:19 crc kubenswrapper[4813]: I1201 09:02:19.418233 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7b8bf821-52ae-409b-8e17-fc8e1302ed10-scripts" (OuterVolumeSpecName: "scripts") pod "7b8bf821-52ae-409b-8e17-fc8e1302ed10" (UID: "7b8bf821-52ae-409b-8e17-fc8e1302ed10"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:02:19 crc kubenswrapper[4813]: I1201 09:02:19.418279 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7b8bf821-52ae-409b-8e17-fc8e1302ed10-kube-api-access-km49c" (OuterVolumeSpecName: "kube-api-access-km49c") pod "7b8bf821-52ae-409b-8e17-fc8e1302ed10" (UID: "7b8bf821-52ae-409b-8e17-fc8e1302ed10"). InnerVolumeSpecName "kube-api-access-km49c". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:02:19 crc kubenswrapper[4813]: I1201 09:02:19.418543 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage10-crc" (OuterVolumeSpecName: "glance-cache") pod "7b8bf821-52ae-409b-8e17-fc8e1302ed10" (UID: "7b8bf821-52ae-409b-8e17-fc8e1302ed10"). InnerVolumeSpecName "local-storage10-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 01 09:02:19 crc kubenswrapper[4813]: I1201 09:02:19.418749 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage16-crc" (OuterVolumeSpecName: "glance") pod "7b8bf821-52ae-409b-8e17-fc8e1302ed10" (UID: "7b8bf821-52ae-409b-8e17-fc8e1302ed10"). InnerVolumeSpecName "local-storage16-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 01 09:02:19 crc kubenswrapper[4813]: I1201 09:02:19.456158 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7b8bf821-52ae-409b-8e17-fc8e1302ed10-config-data" (OuterVolumeSpecName: "config-data") pod "7b8bf821-52ae-409b-8e17-fc8e1302ed10" (UID: "7b8bf821-52ae-409b-8e17-fc8e1302ed10"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:02:19 crc kubenswrapper[4813]: I1201 09:02:19.499466 4813 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7b8bf821-52ae-409b-8e17-fc8e1302ed10-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 09:02:19 crc kubenswrapper[4813]: I1201 09:02:19.499495 4813 reconciler_common.go:293] "Volume detached for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/7b8bf821-52ae-409b-8e17-fc8e1302ed10-etc-nvme\") on node \"crc\" DevicePath \"\"" Dec 01 09:02:19 crc kubenswrapper[4813]: I1201 09:02:19.499505 4813 reconciler_common.go:293] "Volume detached for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/7b8bf821-52ae-409b-8e17-fc8e1302ed10-sys\") on node \"crc\" DevicePath \"\"" Dec 01 09:02:19 crc kubenswrapper[4813]: I1201 09:02:19.499514 4813 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/7b8bf821-52ae-409b-8e17-fc8e1302ed10-lib-modules\") on node \"crc\" DevicePath \"\"" Dec 01 09:02:19 crc kubenswrapper[4813]: I1201 09:02:19.499522 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-km49c\" (UniqueName: \"kubernetes.io/projected/7b8bf821-52ae-409b-8e17-fc8e1302ed10-kube-api-access-km49c\") on node \"crc\" DevicePath \"\"" Dec 01 09:02:19 crc kubenswrapper[4813]: I1201 09:02:19.499685 4813 reconciler_common.go:293] "Volume detached for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/7b8bf821-52ae-409b-8e17-fc8e1302ed10-var-locks-brick\") on node \"crc\" DevicePath \"\"" Dec 01 09:02:19 crc kubenswrapper[4813]: I1201 09:02:19.499695 4813 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7b8bf821-52ae-409b-8e17-fc8e1302ed10-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 09:02:19 crc kubenswrapper[4813]: I1201 09:02:19.499703 4813 reconciler_common.go:293] "Volume detached for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/7b8bf821-52ae-409b-8e17-fc8e1302ed10-dev\") on node \"crc\" DevicePath \"\"" Dec 01 09:02:19 crc kubenswrapper[4813]: I1201 09:02:19.499710 4813 reconciler_common.go:293] "Volume detached for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/7b8bf821-52ae-409b-8e17-fc8e1302ed10-etc-iscsi\") on node \"crc\" DevicePath \"\"" Dec 01 09:02:19 crc kubenswrapper[4813]: I1201 09:02:19.499719 4813 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7b8bf821-52ae-409b-8e17-fc8e1302ed10-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 01 09:02:19 crc kubenswrapper[4813]: I1201 09:02:19.499727 4813 reconciler_common.go:293] "Volume detached for volume \"run\" (UniqueName: \"kubernetes.io/host-path/7b8bf821-52ae-409b-8e17-fc8e1302ed10-run\") on node \"crc\" DevicePath \"\"" Dec 01 09:02:19 crc kubenswrapper[4813]: I1201 09:02:19.499751 4813 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage16-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage16-crc\") on node \"crc\" " Dec 01 09:02:19 crc kubenswrapper[4813]: I1201 09:02:19.499759 4813 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7b8bf821-52ae-409b-8e17-fc8e1302ed10-logs\") on node \"crc\" DevicePath \"\"" Dec 01 09:02:19 crc kubenswrapper[4813]: I1201 09:02:19.499793 4813 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" " Dec 01 09:02:19 crc kubenswrapper[4813]: I1201 09:02:19.516083 4813 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage16-crc" (UniqueName: "kubernetes.io/local-volume/local-storage16-crc") on node "crc" Dec 01 09:02:19 crc kubenswrapper[4813]: I1201 09:02:19.520634 4813 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage10-crc" (UniqueName: "kubernetes.io/local-volume/local-storage10-crc") on node "crc" Dec 01 09:02:19 crc kubenswrapper[4813]: I1201 09:02:19.575538 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:02:19 crc kubenswrapper[4813]: I1201 09:02:19.600383 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/a3fd4804-b7ff-42d4-aa58-999cb59efbc7-run\") pod \"a3fd4804-b7ff-42d4-aa58-999cb59efbc7\" (UID: \"a3fd4804-b7ff-42d4-aa58-999cb59efbc7\") " Dec 01 09:02:19 crc kubenswrapper[4813]: I1201 09:02:19.600443 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/a3fd4804-b7ff-42d4-aa58-999cb59efbc7-lib-modules\") pod \"a3fd4804-b7ff-42d4-aa58-999cb59efbc7\" (UID: \"a3fd4804-b7ff-42d4-aa58-999cb59efbc7\") " Dec 01 09:02:19 crc kubenswrapper[4813]: I1201 09:02:19.600470 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zlg5m\" (UniqueName: \"kubernetes.io/projected/a3fd4804-b7ff-42d4-aa58-999cb59efbc7-kube-api-access-zlg5m\") pod \"a3fd4804-b7ff-42d4-aa58-999cb59efbc7\" (UID: \"a3fd4804-b7ff-42d4-aa58-999cb59efbc7\") " Dec 01 09:02:19 crc kubenswrapper[4813]: I1201 09:02:19.600527 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a3fd4804-b7ff-42d4-aa58-999cb59efbc7-logs\") pod \"a3fd4804-b7ff-42d4-aa58-999cb59efbc7\" (UID: \"a3fd4804-b7ff-42d4-aa58-999cb59efbc7\") " Dec 01 09:02:19 crc kubenswrapper[4813]: I1201 09:02:19.600571 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a3fd4804-b7ff-42d4-aa58-999cb59efbc7-config-data\") pod \"a3fd4804-b7ff-42d4-aa58-999cb59efbc7\" (UID: \"a3fd4804-b7ff-42d4-aa58-999cb59efbc7\") " Dec 01 09:02:19 crc kubenswrapper[4813]: I1201 09:02:19.600609 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a3fd4804-b7ff-42d4-aa58-999cb59efbc7-scripts\") pod \"a3fd4804-b7ff-42d4-aa58-999cb59efbc7\" (UID: \"a3fd4804-b7ff-42d4-aa58-999cb59efbc7\") " Dec 01 09:02:19 crc kubenswrapper[4813]: I1201 09:02:19.600630 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/a3fd4804-b7ff-42d4-aa58-999cb59efbc7-sys\") pod \"a3fd4804-b7ff-42d4-aa58-999cb59efbc7\" (UID: \"a3fd4804-b7ff-42d4-aa58-999cb59efbc7\") " Dec 01 09:02:19 crc kubenswrapper[4813]: I1201 09:02:19.600660 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance-cache\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"a3fd4804-b7ff-42d4-aa58-999cb59efbc7\" (UID: \"a3fd4804-b7ff-42d4-aa58-999cb59efbc7\") " Dec 01 09:02:19 crc kubenswrapper[4813]: I1201 09:02:19.600711 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"a3fd4804-b7ff-42d4-aa58-999cb59efbc7\" (UID: \"a3fd4804-b7ff-42d4-aa58-999cb59efbc7\") " Dec 01 09:02:19 crc kubenswrapper[4813]: I1201 09:02:19.600731 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/a3fd4804-b7ff-42d4-aa58-999cb59efbc7-dev\") pod \"a3fd4804-b7ff-42d4-aa58-999cb59efbc7\" (UID: \"a3fd4804-b7ff-42d4-aa58-999cb59efbc7\") " Dec 01 09:02:19 crc kubenswrapper[4813]: I1201 09:02:19.600800 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/a3fd4804-b7ff-42d4-aa58-999cb59efbc7-httpd-run\") pod \"a3fd4804-b7ff-42d4-aa58-999cb59efbc7\" (UID: \"a3fd4804-b7ff-42d4-aa58-999cb59efbc7\") " Dec 01 09:02:19 crc kubenswrapper[4813]: I1201 09:02:19.600862 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/a3fd4804-b7ff-42d4-aa58-999cb59efbc7-var-locks-brick\") pod \"a3fd4804-b7ff-42d4-aa58-999cb59efbc7\" (UID: \"a3fd4804-b7ff-42d4-aa58-999cb59efbc7\") " Dec 01 09:02:19 crc kubenswrapper[4813]: I1201 09:02:19.600949 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/a3fd4804-b7ff-42d4-aa58-999cb59efbc7-etc-iscsi\") pod \"a3fd4804-b7ff-42d4-aa58-999cb59efbc7\" (UID: \"a3fd4804-b7ff-42d4-aa58-999cb59efbc7\") " Dec 01 09:02:19 crc kubenswrapper[4813]: I1201 09:02:19.601019 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/a3fd4804-b7ff-42d4-aa58-999cb59efbc7-etc-nvme\") pod \"a3fd4804-b7ff-42d4-aa58-999cb59efbc7\" (UID: \"a3fd4804-b7ff-42d4-aa58-999cb59efbc7\") " Dec 01 09:02:19 crc kubenswrapper[4813]: I1201 09:02:19.602789 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a3fd4804-b7ff-42d4-aa58-999cb59efbc7-sys" (OuterVolumeSpecName: "sys") pod "a3fd4804-b7ff-42d4-aa58-999cb59efbc7" (UID: "a3fd4804-b7ff-42d4-aa58-999cb59efbc7"). InnerVolumeSpecName "sys". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 09:02:19 crc kubenswrapper[4813]: I1201 09:02:19.602958 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a3fd4804-b7ff-42d4-aa58-999cb59efbc7-run" (OuterVolumeSpecName: "run") pod "a3fd4804-b7ff-42d4-aa58-999cb59efbc7" (UID: "a3fd4804-b7ff-42d4-aa58-999cb59efbc7"). InnerVolumeSpecName "run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 09:02:19 crc kubenswrapper[4813]: I1201 09:02:19.603062 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a3fd4804-b7ff-42d4-aa58-999cb59efbc7-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "a3fd4804-b7ff-42d4-aa58-999cb59efbc7" (UID: "a3fd4804-b7ff-42d4-aa58-999cb59efbc7"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 09:02:19 crc kubenswrapper[4813]: I1201 09:02:19.604161 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a3fd4804-b7ff-42d4-aa58-999cb59efbc7-var-locks-brick" (OuterVolumeSpecName: "var-locks-brick") pod "a3fd4804-b7ff-42d4-aa58-999cb59efbc7" (UID: "a3fd4804-b7ff-42d4-aa58-999cb59efbc7"). InnerVolumeSpecName "var-locks-brick". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 09:02:19 crc kubenswrapper[4813]: I1201 09:02:19.604430 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a3fd4804-b7ff-42d4-aa58-999cb59efbc7-etc-nvme" (OuterVolumeSpecName: "etc-nvme") pod "a3fd4804-b7ff-42d4-aa58-999cb59efbc7" (UID: "a3fd4804-b7ff-42d4-aa58-999cb59efbc7"). InnerVolumeSpecName "etc-nvme". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 09:02:19 crc kubenswrapper[4813]: I1201 09:02:19.604457 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a3fd4804-b7ff-42d4-aa58-999cb59efbc7-etc-iscsi" (OuterVolumeSpecName: "etc-iscsi") pod "a3fd4804-b7ff-42d4-aa58-999cb59efbc7" (UID: "a3fd4804-b7ff-42d4-aa58-999cb59efbc7"). InnerVolumeSpecName "etc-iscsi". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 09:02:19 crc kubenswrapper[4813]: I1201 09:02:19.604756 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a3fd4804-b7ff-42d4-aa58-999cb59efbc7-dev" (OuterVolumeSpecName: "dev") pod "a3fd4804-b7ff-42d4-aa58-999cb59efbc7" (UID: "a3fd4804-b7ff-42d4-aa58-999cb59efbc7"). InnerVolumeSpecName "dev". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 09:02:19 crc kubenswrapper[4813]: I1201 09:02:19.605311 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a3fd4804-b7ff-42d4-aa58-999cb59efbc7-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "a3fd4804-b7ff-42d4-aa58-999cb59efbc7" (UID: "a3fd4804-b7ff-42d4-aa58-999cb59efbc7"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 09:02:19 crc kubenswrapper[4813]: I1201 09:02:19.605380 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a3fd4804-b7ff-42d4-aa58-999cb59efbc7-logs" (OuterVolumeSpecName: "logs") pod "a3fd4804-b7ff-42d4-aa58-999cb59efbc7" (UID: "a3fd4804-b7ff-42d4-aa58-999cb59efbc7"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 09:02:19 crc kubenswrapper[4813]: I1201 09:02:19.605592 4813 reconciler_common.go:293] "Volume detached for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/a3fd4804-b7ff-42d4-aa58-999cb59efbc7-dev\") on node \"crc\" DevicePath \"\"" Dec 01 09:02:19 crc kubenswrapper[4813]: I1201 09:02:19.605620 4813 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/a3fd4804-b7ff-42d4-aa58-999cb59efbc7-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 01 09:02:19 crc kubenswrapper[4813]: I1201 09:02:19.605666 4813 reconciler_common.go:293] "Volume detached for volume \"local-storage16-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage16-crc\") on node \"crc\" DevicePath \"\"" Dec 01 09:02:19 crc kubenswrapper[4813]: I1201 09:02:19.605679 4813 reconciler_common.go:293] "Volume detached for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/a3fd4804-b7ff-42d4-aa58-999cb59efbc7-var-locks-brick\") on node \"crc\" DevicePath \"\"" Dec 01 09:02:19 crc kubenswrapper[4813]: I1201 09:02:19.605693 4813 reconciler_common.go:293] "Volume detached for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" DevicePath \"\"" Dec 01 09:02:19 crc kubenswrapper[4813]: I1201 09:02:19.605704 4813 reconciler_common.go:293] "Volume detached for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/a3fd4804-b7ff-42d4-aa58-999cb59efbc7-etc-iscsi\") on node \"crc\" DevicePath \"\"" Dec 01 09:02:19 crc kubenswrapper[4813]: I1201 09:02:19.605715 4813 reconciler_common.go:293] "Volume detached for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/a3fd4804-b7ff-42d4-aa58-999cb59efbc7-etc-nvme\") on node \"crc\" DevicePath \"\"" Dec 01 09:02:19 crc kubenswrapper[4813]: I1201 09:02:19.605725 4813 reconciler_common.go:293] "Volume detached for volume \"run\" (UniqueName: \"kubernetes.io/host-path/a3fd4804-b7ff-42d4-aa58-999cb59efbc7-run\") on node \"crc\" DevicePath \"\"" Dec 01 09:02:19 crc kubenswrapper[4813]: I1201 09:02:19.605736 4813 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/a3fd4804-b7ff-42d4-aa58-999cb59efbc7-lib-modules\") on node \"crc\" DevicePath \"\"" Dec 01 09:02:19 crc kubenswrapper[4813]: I1201 09:02:19.605746 4813 reconciler_common.go:293] "Volume detached for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/a3fd4804-b7ff-42d4-aa58-999cb59efbc7-sys\") on node \"crc\" DevicePath \"\"" Dec 01 09:02:19 crc kubenswrapper[4813]: I1201 09:02:19.607595 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage04-crc" (OuterVolumeSpecName: "glance") pod "a3fd4804-b7ff-42d4-aa58-999cb59efbc7" (UID: "a3fd4804-b7ff-42d4-aa58-999cb59efbc7"). InnerVolumeSpecName "local-storage04-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 01 09:02:19 crc kubenswrapper[4813]: I1201 09:02:19.608015 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a3fd4804-b7ff-42d4-aa58-999cb59efbc7-kube-api-access-zlg5m" (OuterVolumeSpecName: "kube-api-access-zlg5m") pod "a3fd4804-b7ff-42d4-aa58-999cb59efbc7" (UID: "a3fd4804-b7ff-42d4-aa58-999cb59efbc7"). InnerVolumeSpecName "kube-api-access-zlg5m". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:02:19 crc kubenswrapper[4813]: I1201 09:02:19.608919 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a3fd4804-b7ff-42d4-aa58-999cb59efbc7-scripts" (OuterVolumeSpecName: "scripts") pod "a3fd4804-b7ff-42d4-aa58-999cb59efbc7" (UID: "a3fd4804-b7ff-42d4-aa58-999cb59efbc7"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:02:19 crc kubenswrapper[4813]: I1201 09:02:19.609292 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage03-crc" (OuterVolumeSpecName: "glance-cache") pod "a3fd4804-b7ff-42d4-aa58-999cb59efbc7" (UID: "a3fd4804-b7ff-42d4-aa58-999cb59efbc7"). InnerVolumeSpecName "local-storage03-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 01 09:02:19 crc kubenswrapper[4813]: I1201 09:02:19.652112 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a3fd4804-b7ff-42d4-aa58-999cb59efbc7-config-data" (OuterVolumeSpecName: "config-data") pod "a3fd4804-b7ff-42d4-aa58-999cb59efbc7" (UID: "a3fd4804-b7ff-42d4-aa58-999cb59efbc7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:02:19 crc kubenswrapper[4813]: I1201 09:02:19.706699 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zlg5m\" (UniqueName: \"kubernetes.io/projected/a3fd4804-b7ff-42d4-aa58-999cb59efbc7-kube-api-access-zlg5m\") on node \"crc\" DevicePath \"\"" Dec 01 09:02:19 crc kubenswrapper[4813]: I1201 09:02:19.706747 4813 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a3fd4804-b7ff-42d4-aa58-999cb59efbc7-logs\") on node \"crc\" DevicePath \"\"" Dec 01 09:02:19 crc kubenswrapper[4813]: I1201 09:02:19.706764 4813 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a3fd4804-b7ff-42d4-aa58-999cb59efbc7-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 09:02:19 crc kubenswrapper[4813]: I1201 09:02:19.706778 4813 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a3fd4804-b7ff-42d4-aa58-999cb59efbc7-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 09:02:19 crc kubenswrapper[4813]: I1201 09:02:19.706821 4813 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" " Dec 01 09:02:19 crc kubenswrapper[4813]: I1201 09:02:19.706840 4813 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" " Dec 01 09:02:19 crc kubenswrapper[4813]: I1201 09:02:19.721225 4813 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage04-crc" (UniqueName: "kubernetes.io/local-volume/local-storage04-crc") on node "crc" Dec 01 09:02:19 crc kubenswrapper[4813]: I1201 09:02:19.723106 4813 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage03-crc" (UniqueName: "kubernetes.io/local-volume/local-storage03-crc") on node "crc" Dec 01 09:02:19 crc kubenswrapper[4813]: I1201 09:02:19.807609 4813 reconciler_common.go:293] "Volume detached for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" DevicePath \"\"" Dec 01 09:02:19 crc kubenswrapper[4813]: I1201 09:02:19.807944 4813 reconciler_common.go:293] "Volume detached for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" DevicePath \"\"" Dec 01 09:02:20 crc kubenswrapper[4813]: I1201 09:02:20.161457 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-0" event={"ID":"7b8bf821-52ae-409b-8e17-fc8e1302ed10","Type":"ContainerDied","Data":"ee0e6ad4fc73cf0cefcbbef4c9024d0c4e23c891a852b9d53bae9d0ab1476118"} Dec 01 09:02:20 crc kubenswrapper[4813]: I1201 09:02:20.161520 4813 scope.go:117] "RemoveContainer" containerID="a22c275b1abdf242357c033d2a976a1e0bf84f3b733c91633972190ba9de2c51" Dec 01 09:02:20 crc kubenswrapper[4813]: I1201 09:02:20.161531 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-0" Dec 01 09:02:20 crc kubenswrapper[4813]: I1201 09:02:20.168326 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:02:20 crc kubenswrapper[4813]: I1201 09:02:20.168346 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"a3fd4804-b7ff-42d4-aa58-999cb59efbc7","Type":"ContainerDied","Data":"d89dec11e5c1011779b313ec517b6cceb51e641055e65c21caf627d7b47703b2"} Dec 01 09:02:20 crc kubenswrapper[4813]: I1201 09:02:20.168215 4813 generic.go:334] "Generic (PLEG): container finished" podID="a3fd4804-b7ff-42d4-aa58-999cb59efbc7" containerID="d89dec11e5c1011779b313ec517b6cceb51e641055e65c21caf627d7b47703b2" exitCode=0 Dec 01 09:02:20 crc kubenswrapper[4813]: I1201 09:02:20.179711 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"a3fd4804-b7ff-42d4-aa58-999cb59efbc7","Type":"ContainerDied","Data":"dbd0a66eff1c44932b1223799123a9e56a244df412f20f7aa8661fe3c1c12d26"} Dec 01 09:02:20 crc kubenswrapper[4813]: I1201 09:02:20.205310 4813 scope.go:117] "RemoveContainer" containerID="56c363c42bb8f1be549639c72070e76cf11757d4ca387266e8486456ffae3383" Dec 01 09:02:20 crc kubenswrapper[4813]: I1201 09:02:20.219758 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-0"] Dec 01 09:02:20 crc kubenswrapper[4813]: I1201 09:02:20.224744 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-0"] Dec 01 09:02:20 crc kubenswrapper[4813]: I1201 09:02:20.227099 4813 scope.go:117] "RemoveContainer" containerID="d89dec11e5c1011779b313ec517b6cceb51e641055e65c21caf627d7b47703b2" Dec 01 09:02:20 crc kubenswrapper[4813]: I1201 09:02:20.231706 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Dec 01 09:02:20 crc kubenswrapper[4813]: I1201 09:02:20.237042 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Dec 01 09:02:20 crc kubenswrapper[4813]: I1201 09:02:20.251335 4813 scope.go:117] "RemoveContainer" containerID="6781e4dd9f6569955b120e287571b0574a25a313f8bbb4c49c326fc1bd841f7d" Dec 01 09:02:20 crc kubenswrapper[4813]: I1201 09:02:20.270553 4813 scope.go:117] "RemoveContainer" containerID="d89dec11e5c1011779b313ec517b6cceb51e641055e65c21caf627d7b47703b2" Dec 01 09:02:20 crc kubenswrapper[4813]: E1201 09:02:20.271411 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d89dec11e5c1011779b313ec517b6cceb51e641055e65c21caf627d7b47703b2\": container with ID starting with d89dec11e5c1011779b313ec517b6cceb51e641055e65c21caf627d7b47703b2 not found: ID does not exist" containerID="d89dec11e5c1011779b313ec517b6cceb51e641055e65c21caf627d7b47703b2" Dec 01 09:02:20 crc kubenswrapper[4813]: I1201 09:02:20.271490 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d89dec11e5c1011779b313ec517b6cceb51e641055e65c21caf627d7b47703b2"} err="failed to get container status \"d89dec11e5c1011779b313ec517b6cceb51e641055e65c21caf627d7b47703b2\": rpc error: code = NotFound desc = could not find container \"d89dec11e5c1011779b313ec517b6cceb51e641055e65c21caf627d7b47703b2\": container with ID starting with d89dec11e5c1011779b313ec517b6cceb51e641055e65c21caf627d7b47703b2 not found: ID does not exist" Dec 01 09:02:20 crc kubenswrapper[4813]: I1201 09:02:20.271529 4813 scope.go:117] "RemoveContainer" containerID="6781e4dd9f6569955b120e287571b0574a25a313f8bbb4c49c326fc1bd841f7d" Dec 01 09:02:20 crc kubenswrapper[4813]: E1201 09:02:20.272250 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6781e4dd9f6569955b120e287571b0574a25a313f8bbb4c49c326fc1bd841f7d\": container with ID starting with 6781e4dd9f6569955b120e287571b0574a25a313f8bbb4c49c326fc1bd841f7d not found: ID does not exist" containerID="6781e4dd9f6569955b120e287571b0574a25a313f8bbb4c49c326fc1bd841f7d" Dec 01 09:02:20 crc kubenswrapper[4813]: I1201 09:02:20.272296 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6781e4dd9f6569955b120e287571b0574a25a313f8bbb4c49c326fc1bd841f7d"} err="failed to get container status \"6781e4dd9f6569955b120e287571b0574a25a313f8bbb4c49c326fc1bd841f7d\": rpc error: code = NotFound desc = could not find container \"6781e4dd9f6569955b120e287571b0574a25a313f8bbb4c49c326fc1bd841f7d\": container with ID starting with 6781e4dd9f6569955b120e287571b0574a25a313f8bbb4c49c326fc1bd841f7d not found: ID does not exist" Dec 01 09:02:20 crc kubenswrapper[4813]: I1201 09:02:20.402105 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7b8bf821-52ae-409b-8e17-fc8e1302ed10" path="/var/lib/kubelet/pods/7b8bf821-52ae-409b-8e17-fc8e1302ed10/volumes" Dec 01 09:02:20 crc kubenswrapper[4813]: I1201 09:02:20.402709 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a3fd4804-b7ff-42d4-aa58-999cb59efbc7" path="/var/lib/kubelet/pods/a3fd4804-b7ff-42d4-aa58-999cb59efbc7/volumes" Dec 01 09:02:21 crc kubenswrapper[4813]: I1201 09:02:21.554884 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-db-sync-9lgcq"] Dec 01 09:02:21 crc kubenswrapper[4813]: I1201 09:02:21.563449 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-db-sync-9lgcq"] Dec 01 09:02:21 crc kubenswrapper[4813]: I1201 09:02:21.657870 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glanceedab-account-delete-fqq74"] Dec 01 09:02:21 crc kubenswrapper[4813]: E1201 09:02:21.658318 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7790813b-eccc-442b-a400-7fb5863599a7" containerName="glance-httpd" Dec 01 09:02:21 crc kubenswrapper[4813]: I1201 09:02:21.658353 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="7790813b-eccc-442b-a400-7fb5863599a7" containerName="glance-httpd" Dec 01 09:02:21 crc kubenswrapper[4813]: E1201 09:02:21.658366 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7b8bf821-52ae-409b-8e17-fc8e1302ed10" containerName="glance-httpd" Dec 01 09:02:21 crc kubenswrapper[4813]: I1201 09:02:21.658374 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="7b8bf821-52ae-409b-8e17-fc8e1302ed10" containerName="glance-httpd" Dec 01 09:02:21 crc kubenswrapper[4813]: E1201 09:02:21.658391 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7790813b-eccc-442b-a400-7fb5863599a7" containerName="glance-log" Dec 01 09:02:21 crc kubenswrapper[4813]: I1201 09:02:21.658401 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="7790813b-eccc-442b-a400-7fb5863599a7" containerName="glance-log" Dec 01 09:02:21 crc kubenswrapper[4813]: E1201 09:02:21.658416 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a3fd4804-b7ff-42d4-aa58-999cb59efbc7" containerName="glance-httpd" Dec 01 09:02:21 crc kubenswrapper[4813]: I1201 09:02:21.658424 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="a3fd4804-b7ff-42d4-aa58-999cb59efbc7" containerName="glance-httpd" Dec 01 09:02:21 crc kubenswrapper[4813]: E1201 09:02:21.658436 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e58d9e76-7271-47c3-a7aa-a5a802bfdb41" containerName="glance-log" Dec 01 09:02:21 crc kubenswrapper[4813]: I1201 09:02:21.658443 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="e58d9e76-7271-47c3-a7aa-a5a802bfdb41" containerName="glance-log" Dec 01 09:02:21 crc kubenswrapper[4813]: E1201 09:02:21.658456 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="43861577-d77d-45ed-95ea-25a5dffa1b46" containerName="glance-httpd" Dec 01 09:02:21 crc kubenswrapper[4813]: I1201 09:02:21.658464 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="43861577-d77d-45ed-95ea-25a5dffa1b46" containerName="glance-httpd" Dec 01 09:02:21 crc kubenswrapper[4813]: E1201 09:02:21.658479 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7b8bf821-52ae-409b-8e17-fc8e1302ed10" containerName="glance-log" Dec 01 09:02:21 crc kubenswrapper[4813]: I1201 09:02:21.658486 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="7b8bf821-52ae-409b-8e17-fc8e1302ed10" containerName="glance-log" Dec 01 09:02:21 crc kubenswrapper[4813]: E1201 09:02:21.658496 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0bb179fc-fb0c-420a-a432-5a0d28884e23" containerName="glance-log" Dec 01 09:02:21 crc kubenswrapper[4813]: I1201 09:02:21.658504 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="0bb179fc-fb0c-420a-a432-5a0d28884e23" containerName="glance-log" Dec 01 09:02:21 crc kubenswrapper[4813]: E1201 09:02:21.658522 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="43861577-d77d-45ed-95ea-25a5dffa1b46" containerName="glance-log" Dec 01 09:02:21 crc kubenswrapper[4813]: I1201 09:02:21.658530 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="43861577-d77d-45ed-95ea-25a5dffa1b46" containerName="glance-log" Dec 01 09:02:21 crc kubenswrapper[4813]: E1201 09:02:21.658543 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e58d9e76-7271-47c3-a7aa-a5a802bfdb41" containerName="glance-httpd" Dec 01 09:02:21 crc kubenswrapper[4813]: I1201 09:02:21.658550 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="e58d9e76-7271-47c3-a7aa-a5a802bfdb41" containerName="glance-httpd" Dec 01 09:02:21 crc kubenswrapper[4813]: E1201 09:02:21.658567 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a3fd4804-b7ff-42d4-aa58-999cb59efbc7" containerName="glance-log" Dec 01 09:02:21 crc kubenswrapper[4813]: I1201 09:02:21.658575 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="a3fd4804-b7ff-42d4-aa58-999cb59efbc7" containerName="glance-log" Dec 01 09:02:21 crc kubenswrapper[4813]: E1201 09:02:21.658590 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0bb179fc-fb0c-420a-a432-5a0d28884e23" containerName="glance-httpd" Dec 01 09:02:21 crc kubenswrapper[4813]: I1201 09:02:21.658598 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="0bb179fc-fb0c-420a-a432-5a0d28884e23" containerName="glance-httpd" Dec 01 09:02:21 crc kubenswrapper[4813]: I1201 09:02:21.658787 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="0bb179fc-fb0c-420a-a432-5a0d28884e23" containerName="glance-log" Dec 01 09:02:21 crc kubenswrapper[4813]: I1201 09:02:21.658814 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="43861577-d77d-45ed-95ea-25a5dffa1b46" containerName="glance-httpd" Dec 01 09:02:21 crc kubenswrapper[4813]: I1201 09:02:21.658827 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="7790813b-eccc-442b-a400-7fb5863599a7" containerName="glance-httpd" Dec 01 09:02:21 crc kubenswrapper[4813]: I1201 09:02:21.658840 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="7b8bf821-52ae-409b-8e17-fc8e1302ed10" containerName="glance-httpd" Dec 01 09:02:21 crc kubenswrapper[4813]: I1201 09:02:21.658853 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="e58d9e76-7271-47c3-a7aa-a5a802bfdb41" containerName="glance-httpd" Dec 01 09:02:21 crc kubenswrapper[4813]: I1201 09:02:21.658863 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="a3fd4804-b7ff-42d4-aa58-999cb59efbc7" containerName="glance-httpd" Dec 01 09:02:21 crc kubenswrapper[4813]: I1201 09:02:21.658875 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="43861577-d77d-45ed-95ea-25a5dffa1b46" containerName="glance-log" Dec 01 09:02:21 crc kubenswrapper[4813]: I1201 09:02:21.658889 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="7790813b-eccc-442b-a400-7fb5863599a7" containerName="glance-log" Dec 01 09:02:21 crc kubenswrapper[4813]: I1201 09:02:21.658899 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="a3fd4804-b7ff-42d4-aa58-999cb59efbc7" containerName="glance-log" Dec 01 09:02:21 crc kubenswrapper[4813]: I1201 09:02:21.658911 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="e58d9e76-7271-47c3-a7aa-a5a802bfdb41" containerName="glance-log" Dec 01 09:02:21 crc kubenswrapper[4813]: I1201 09:02:21.658921 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="0bb179fc-fb0c-420a-a432-5a0d28884e23" containerName="glance-httpd" Dec 01 09:02:21 crc kubenswrapper[4813]: I1201 09:02:21.658931 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="7b8bf821-52ae-409b-8e17-fc8e1302ed10" containerName="glance-log" Dec 01 09:02:21 crc kubenswrapper[4813]: I1201 09:02:21.659633 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glanceedab-account-delete-fqq74" Dec 01 09:02:21 crc kubenswrapper[4813]: I1201 09:02:21.666570 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glanceedab-account-delete-fqq74"] Dec 01 09:02:21 crc kubenswrapper[4813]: I1201 09:02:21.736478 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2128b3a5-1af3-48c4-bca2-10fa25cfb9f3-operator-scripts\") pod \"glanceedab-account-delete-fqq74\" (UID: \"2128b3a5-1af3-48c4-bca2-10fa25cfb9f3\") " pod="glance-kuttl-tests/glanceedab-account-delete-fqq74" Dec 01 09:02:21 crc kubenswrapper[4813]: I1201 09:02:21.736809 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vgb54\" (UniqueName: \"kubernetes.io/projected/2128b3a5-1af3-48c4-bca2-10fa25cfb9f3-kube-api-access-vgb54\") pod \"glanceedab-account-delete-fqq74\" (UID: \"2128b3a5-1af3-48c4-bca2-10fa25cfb9f3\") " pod="glance-kuttl-tests/glanceedab-account-delete-fqq74" Dec 01 09:02:21 crc kubenswrapper[4813]: I1201 09:02:21.846917 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vgb54\" (UniqueName: \"kubernetes.io/projected/2128b3a5-1af3-48c4-bca2-10fa25cfb9f3-kube-api-access-vgb54\") pod \"glanceedab-account-delete-fqq74\" (UID: \"2128b3a5-1af3-48c4-bca2-10fa25cfb9f3\") " pod="glance-kuttl-tests/glanceedab-account-delete-fqq74" Dec 01 09:02:21 crc kubenswrapper[4813]: I1201 09:02:21.846984 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2128b3a5-1af3-48c4-bca2-10fa25cfb9f3-operator-scripts\") pod \"glanceedab-account-delete-fqq74\" (UID: \"2128b3a5-1af3-48c4-bca2-10fa25cfb9f3\") " pod="glance-kuttl-tests/glanceedab-account-delete-fqq74" Dec 01 09:02:21 crc kubenswrapper[4813]: I1201 09:02:21.847957 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2128b3a5-1af3-48c4-bca2-10fa25cfb9f3-operator-scripts\") pod \"glanceedab-account-delete-fqq74\" (UID: \"2128b3a5-1af3-48c4-bca2-10fa25cfb9f3\") " pod="glance-kuttl-tests/glanceedab-account-delete-fqq74" Dec 01 09:02:21 crc kubenswrapper[4813]: I1201 09:02:21.866224 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vgb54\" (UniqueName: \"kubernetes.io/projected/2128b3a5-1af3-48c4-bca2-10fa25cfb9f3-kube-api-access-vgb54\") pod \"glanceedab-account-delete-fqq74\" (UID: \"2128b3a5-1af3-48c4-bca2-10fa25cfb9f3\") " pod="glance-kuttl-tests/glanceedab-account-delete-fqq74" Dec 01 09:02:21 crc kubenswrapper[4813]: I1201 09:02:21.975266 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glanceedab-account-delete-fqq74" Dec 01 09:02:22 crc kubenswrapper[4813]: I1201 09:02:22.290899 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glanceedab-account-delete-fqq74"] Dec 01 09:02:22 crc kubenswrapper[4813]: I1201 09:02:22.420234 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd1b1c1f-46a6-4c70-a583-45da831d2cb1" path="/var/lib/kubelet/pods/cd1b1c1f-46a6-4c70-a583-45da831d2cb1/volumes" Dec 01 09:02:23 crc kubenswrapper[4813]: I1201 09:02:23.226041 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glanceedab-account-delete-fqq74" event={"ID":"2128b3a5-1af3-48c4-bca2-10fa25cfb9f3","Type":"ContainerStarted","Data":"b1f1e4333cca2a8d91aa2d3416981ba00a00c8bd2ea576ab629d745b8760ca3f"} Dec 01 09:02:23 crc kubenswrapper[4813]: I1201 09:02:23.226098 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glanceedab-account-delete-fqq74" event={"ID":"2128b3a5-1af3-48c4-bca2-10fa25cfb9f3","Type":"ContainerStarted","Data":"30af5316fb10ca2b3309d7424feecbd859067bb4202c1c3509c0776b7f8fa9c9"} Dec 01 09:02:23 crc kubenswrapper[4813]: I1201 09:02:23.247188 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glanceedab-account-delete-fqq74" podStartSLOduration=2.247136713 podStartE2EDuration="2.247136713s" podCreationTimestamp="2025-12-01 09:02:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 09:02:23.245450675 +0000 UTC m=+1327.868372291" watchObservedRunningTime="2025-12-01 09:02:23.247136713 +0000 UTC m=+1327.870058309" Dec 01 09:02:24 crc kubenswrapper[4813]: I1201 09:02:24.238557 4813 generic.go:334] "Generic (PLEG): container finished" podID="2128b3a5-1af3-48c4-bca2-10fa25cfb9f3" containerID="b1f1e4333cca2a8d91aa2d3416981ba00a00c8bd2ea576ab629d745b8760ca3f" exitCode=0 Dec 01 09:02:24 crc kubenswrapper[4813]: I1201 09:02:24.238626 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glanceedab-account-delete-fqq74" event={"ID":"2128b3a5-1af3-48c4-bca2-10fa25cfb9f3","Type":"ContainerDied","Data":"b1f1e4333cca2a8d91aa2d3416981ba00a00c8bd2ea576ab629d745b8760ca3f"} Dec 01 09:02:25 crc kubenswrapper[4813]: I1201 09:02:25.610798 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glanceedab-account-delete-fqq74" Dec 01 09:02:25 crc kubenswrapper[4813]: I1201 09:02:25.734249 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vgb54\" (UniqueName: \"kubernetes.io/projected/2128b3a5-1af3-48c4-bca2-10fa25cfb9f3-kube-api-access-vgb54\") pod \"2128b3a5-1af3-48c4-bca2-10fa25cfb9f3\" (UID: \"2128b3a5-1af3-48c4-bca2-10fa25cfb9f3\") " Dec 01 09:02:25 crc kubenswrapper[4813]: I1201 09:02:25.735333 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2128b3a5-1af3-48c4-bca2-10fa25cfb9f3-operator-scripts\") pod \"2128b3a5-1af3-48c4-bca2-10fa25cfb9f3\" (UID: \"2128b3a5-1af3-48c4-bca2-10fa25cfb9f3\") " Dec 01 09:02:25 crc kubenswrapper[4813]: I1201 09:02:25.735748 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2128b3a5-1af3-48c4-bca2-10fa25cfb9f3-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "2128b3a5-1af3-48c4-bca2-10fa25cfb9f3" (UID: "2128b3a5-1af3-48c4-bca2-10fa25cfb9f3"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 09:02:25 crc kubenswrapper[4813]: I1201 09:02:25.736334 4813 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2128b3a5-1af3-48c4-bca2-10fa25cfb9f3-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 09:02:25 crc kubenswrapper[4813]: I1201 09:02:25.742325 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2128b3a5-1af3-48c4-bca2-10fa25cfb9f3-kube-api-access-vgb54" (OuterVolumeSpecName: "kube-api-access-vgb54") pod "2128b3a5-1af3-48c4-bca2-10fa25cfb9f3" (UID: "2128b3a5-1af3-48c4-bca2-10fa25cfb9f3"). InnerVolumeSpecName "kube-api-access-vgb54". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:02:25 crc kubenswrapper[4813]: I1201 09:02:25.837989 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vgb54\" (UniqueName: \"kubernetes.io/projected/2128b3a5-1af3-48c4-bca2-10fa25cfb9f3-kube-api-access-vgb54\") on node \"crc\" DevicePath \"\"" Dec 01 09:02:26 crc kubenswrapper[4813]: I1201 09:02:26.261730 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glanceedab-account-delete-fqq74" event={"ID":"2128b3a5-1af3-48c4-bca2-10fa25cfb9f3","Type":"ContainerDied","Data":"30af5316fb10ca2b3309d7424feecbd859067bb4202c1c3509c0776b7f8fa9c9"} Dec 01 09:02:26 crc kubenswrapper[4813]: I1201 09:02:26.261771 4813 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="30af5316fb10ca2b3309d7424feecbd859067bb4202c1c3509c0776b7f8fa9c9" Dec 01 09:02:26 crc kubenswrapper[4813]: I1201 09:02:26.261795 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glanceedab-account-delete-fqq74" Dec 01 09:02:26 crc kubenswrapper[4813]: I1201 09:02:26.683147 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-db-create-p4lj2"] Dec 01 09:02:26 crc kubenswrapper[4813]: I1201 09:02:26.691004 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-db-create-p4lj2"] Dec 01 09:02:26 crc kubenswrapper[4813]: I1201 09:02:26.705014 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-edab-account-create-update-gcfkg"] Dec 01 09:02:26 crc kubenswrapper[4813]: I1201 09:02:26.711812 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glanceedab-account-delete-fqq74"] Dec 01 09:02:26 crc kubenswrapper[4813]: I1201 09:02:26.718401 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-edab-account-create-update-gcfkg"] Dec 01 09:02:26 crc kubenswrapper[4813]: I1201 09:02:26.724480 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glanceedab-account-delete-fqq74"] Dec 01 09:02:27 crc kubenswrapper[4813]: I1201 09:02:27.192282 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-c6e3-account-create-update-h46sk"] Dec 01 09:02:27 crc kubenswrapper[4813]: E1201 09:02:27.192930 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2128b3a5-1af3-48c4-bca2-10fa25cfb9f3" containerName="mariadb-account-delete" Dec 01 09:02:27 crc kubenswrapper[4813]: I1201 09:02:27.193037 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="2128b3a5-1af3-48c4-bca2-10fa25cfb9f3" containerName="mariadb-account-delete" Dec 01 09:02:27 crc kubenswrapper[4813]: I1201 09:02:27.193268 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="2128b3a5-1af3-48c4-bca2-10fa25cfb9f3" containerName="mariadb-account-delete" Dec 01 09:02:27 crc kubenswrapper[4813]: I1201 09:02:27.193910 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-c6e3-account-create-update-h46sk" Dec 01 09:02:27 crc kubenswrapper[4813]: I1201 09:02:27.196193 4813 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-db-secret" Dec 01 09:02:27 crc kubenswrapper[4813]: I1201 09:02:27.201084 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-db-create-p5mf4"] Dec 01 09:02:27 crc kubenswrapper[4813]: I1201 09:02:27.202280 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-p5mf4" Dec 01 09:02:27 crc kubenswrapper[4813]: I1201 09:02:27.218350 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-create-p5mf4"] Dec 01 09:02:27 crc kubenswrapper[4813]: I1201 09:02:27.227130 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-c6e3-account-create-update-h46sk"] Dec 01 09:02:27 crc kubenswrapper[4813]: I1201 09:02:27.260594 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d3dfa26e-2774-4129-9bfc-5cf40aa3e942-operator-scripts\") pod \"glance-c6e3-account-create-update-h46sk\" (UID: \"d3dfa26e-2774-4129-9bfc-5cf40aa3e942\") " pod="glance-kuttl-tests/glance-c6e3-account-create-update-h46sk" Dec 01 09:02:27 crc kubenswrapper[4813]: I1201 09:02:27.260672 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x9vvj\" (UniqueName: \"kubernetes.io/projected/d3dfa26e-2774-4129-9bfc-5cf40aa3e942-kube-api-access-x9vvj\") pod \"glance-c6e3-account-create-update-h46sk\" (UID: \"d3dfa26e-2774-4129-9bfc-5cf40aa3e942\") " pod="glance-kuttl-tests/glance-c6e3-account-create-update-h46sk" Dec 01 09:02:27 crc kubenswrapper[4813]: I1201 09:02:27.260844 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/11bd67f4-fce9-479e-858b-44e6b1829d81-operator-scripts\") pod \"glance-db-create-p5mf4\" (UID: \"11bd67f4-fce9-479e-858b-44e6b1829d81\") " pod="glance-kuttl-tests/glance-db-create-p5mf4" Dec 01 09:02:27 crc kubenswrapper[4813]: I1201 09:02:27.260889 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mqr8l\" (UniqueName: \"kubernetes.io/projected/11bd67f4-fce9-479e-858b-44e6b1829d81-kube-api-access-mqr8l\") pod \"glance-db-create-p5mf4\" (UID: \"11bd67f4-fce9-479e-858b-44e6b1829d81\") " pod="glance-kuttl-tests/glance-db-create-p5mf4" Dec 01 09:02:27 crc kubenswrapper[4813]: I1201 09:02:27.362252 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/11bd67f4-fce9-479e-858b-44e6b1829d81-operator-scripts\") pod \"glance-db-create-p5mf4\" (UID: \"11bd67f4-fce9-479e-858b-44e6b1829d81\") " pod="glance-kuttl-tests/glance-db-create-p5mf4" Dec 01 09:02:27 crc kubenswrapper[4813]: I1201 09:02:27.362312 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mqr8l\" (UniqueName: \"kubernetes.io/projected/11bd67f4-fce9-479e-858b-44e6b1829d81-kube-api-access-mqr8l\") pod \"glance-db-create-p5mf4\" (UID: \"11bd67f4-fce9-479e-858b-44e6b1829d81\") " pod="glance-kuttl-tests/glance-db-create-p5mf4" Dec 01 09:02:27 crc kubenswrapper[4813]: I1201 09:02:27.362359 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d3dfa26e-2774-4129-9bfc-5cf40aa3e942-operator-scripts\") pod \"glance-c6e3-account-create-update-h46sk\" (UID: \"d3dfa26e-2774-4129-9bfc-5cf40aa3e942\") " pod="glance-kuttl-tests/glance-c6e3-account-create-update-h46sk" Dec 01 09:02:27 crc kubenswrapper[4813]: I1201 09:02:27.362398 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x9vvj\" (UniqueName: \"kubernetes.io/projected/d3dfa26e-2774-4129-9bfc-5cf40aa3e942-kube-api-access-x9vvj\") pod \"glance-c6e3-account-create-update-h46sk\" (UID: \"d3dfa26e-2774-4129-9bfc-5cf40aa3e942\") " pod="glance-kuttl-tests/glance-c6e3-account-create-update-h46sk" Dec 01 09:02:27 crc kubenswrapper[4813]: I1201 09:02:27.363116 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/11bd67f4-fce9-479e-858b-44e6b1829d81-operator-scripts\") pod \"glance-db-create-p5mf4\" (UID: \"11bd67f4-fce9-479e-858b-44e6b1829d81\") " pod="glance-kuttl-tests/glance-db-create-p5mf4" Dec 01 09:02:27 crc kubenswrapper[4813]: I1201 09:02:27.363314 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d3dfa26e-2774-4129-9bfc-5cf40aa3e942-operator-scripts\") pod \"glance-c6e3-account-create-update-h46sk\" (UID: \"d3dfa26e-2774-4129-9bfc-5cf40aa3e942\") " pod="glance-kuttl-tests/glance-c6e3-account-create-update-h46sk" Dec 01 09:02:27 crc kubenswrapper[4813]: I1201 09:02:27.381504 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mqr8l\" (UniqueName: \"kubernetes.io/projected/11bd67f4-fce9-479e-858b-44e6b1829d81-kube-api-access-mqr8l\") pod \"glance-db-create-p5mf4\" (UID: \"11bd67f4-fce9-479e-858b-44e6b1829d81\") " pod="glance-kuttl-tests/glance-db-create-p5mf4" Dec 01 09:02:27 crc kubenswrapper[4813]: I1201 09:02:27.382493 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x9vvj\" (UniqueName: \"kubernetes.io/projected/d3dfa26e-2774-4129-9bfc-5cf40aa3e942-kube-api-access-x9vvj\") pod \"glance-c6e3-account-create-update-h46sk\" (UID: \"d3dfa26e-2774-4129-9bfc-5cf40aa3e942\") " pod="glance-kuttl-tests/glance-c6e3-account-create-update-h46sk" Dec 01 09:02:27 crc kubenswrapper[4813]: I1201 09:02:27.513429 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-c6e3-account-create-update-h46sk" Dec 01 09:02:27 crc kubenswrapper[4813]: I1201 09:02:27.524939 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-p5mf4" Dec 01 09:02:28 crc kubenswrapper[4813]: I1201 09:02:28.433484 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2128b3a5-1af3-48c4-bca2-10fa25cfb9f3" path="/var/lib/kubelet/pods/2128b3a5-1af3-48c4-bca2-10fa25cfb9f3/volumes" Dec 01 09:02:28 crc kubenswrapper[4813]: I1201 09:02:28.438125 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5dc86ea3-5e32-4c91-9c14-da28209fd375" path="/var/lib/kubelet/pods/5dc86ea3-5e32-4c91-9c14-da28209fd375/volumes" Dec 01 09:02:28 crc kubenswrapper[4813]: I1201 09:02:28.439729 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="85bca524-2f65-4994-8471-8c45ac8cdd22" path="/var/lib/kubelet/pods/85bca524-2f65-4994-8471-8c45ac8cdd22/volumes" Dec 01 09:02:28 crc kubenswrapper[4813]: I1201 09:02:28.440871 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-create-p5mf4"] Dec 01 09:02:28 crc kubenswrapper[4813]: I1201 09:02:28.486956 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-c6e3-account-create-update-h46sk"] Dec 01 09:02:28 crc kubenswrapper[4813]: W1201 09:02:28.499485 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd3dfa26e_2774_4129_9bfc_5cf40aa3e942.slice/crio-f680eaa6f7ff9cb4ba5e6ea00eb6213f16fc703dd6cc282b7d58bdd19da4ae06 WatchSource:0}: Error finding container f680eaa6f7ff9cb4ba5e6ea00eb6213f16fc703dd6cc282b7d58bdd19da4ae06: Status 404 returned error can't find the container with id f680eaa6f7ff9cb4ba5e6ea00eb6213f16fc703dd6cc282b7d58bdd19da4ae06 Dec 01 09:02:29 crc kubenswrapper[4813]: I1201 09:02:29.289731 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-create-p5mf4" event={"ID":"11bd67f4-fce9-479e-858b-44e6b1829d81","Type":"ContainerStarted","Data":"b4e5ecf79bdec70cdef568e7ab234ac0de6101613bf4e190c96d980bed839811"} Dec 01 09:02:29 crc kubenswrapper[4813]: I1201 09:02:29.290110 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-create-p5mf4" event={"ID":"11bd67f4-fce9-479e-858b-44e6b1829d81","Type":"ContainerStarted","Data":"cc99195c4f7f71b9836b14f0f672c3d0e6206959faa6675fe011fc7191c6e879"} Dec 01 09:02:29 crc kubenswrapper[4813]: I1201 09:02:29.292499 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-c6e3-account-create-update-h46sk" event={"ID":"d3dfa26e-2774-4129-9bfc-5cf40aa3e942","Type":"ContainerStarted","Data":"8250ba203c774f9ed961312dac84a6a807f187353d50fe2580fe5ee85b4bd6f1"} Dec 01 09:02:29 crc kubenswrapper[4813]: I1201 09:02:29.292546 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-c6e3-account-create-update-h46sk" event={"ID":"d3dfa26e-2774-4129-9bfc-5cf40aa3e942","Type":"ContainerStarted","Data":"f680eaa6f7ff9cb4ba5e6ea00eb6213f16fc703dd6cc282b7d58bdd19da4ae06"} Dec 01 09:02:29 crc kubenswrapper[4813]: I1201 09:02:29.327083 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-db-create-p5mf4" podStartSLOduration=2.327050759 podStartE2EDuration="2.327050759s" podCreationTimestamp="2025-12-01 09:02:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 09:02:29.312663552 +0000 UTC m=+1333.935585148" watchObservedRunningTime="2025-12-01 09:02:29.327050759 +0000 UTC m=+1333.949972385" Dec 01 09:02:30 crc kubenswrapper[4813]: I1201 09:02:30.305950 4813 generic.go:334] "Generic (PLEG): container finished" podID="11bd67f4-fce9-479e-858b-44e6b1829d81" containerID="b4e5ecf79bdec70cdef568e7ab234ac0de6101613bf4e190c96d980bed839811" exitCode=0 Dec 01 09:02:30 crc kubenswrapper[4813]: I1201 09:02:30.306133 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-create-p5mf4" event={"ID":"11bd67f4-fce9-479e-858b-44e6b1829d81","Type":"ContainerDied","Data":"b4e5ecf79bdec70cdef568e7ab234ac0de6101613bf4e190c96d980bed839811"} Dec 01 09:02:30 crc kubenswrapper[4813]: I1201 09:02:30.327270 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-c6e3-account-create-update-h46sk" podStartSLOduration=3.327228461 podStartE2EDuration="3.327228461s" podCreationTimestamp="2025-12-01 09:02:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 09:02:29.335830717 +0000 UTC m=+1333.958752343" watchObservedRunningTime="2025-12-01 09:02:30.327228461 +0000 UTC m=+1334.950150067" Dec 01 09:02:31 crc kubenswrapper[4813]: I1201 09:02:31.611752 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-p5mf4" Dec 01 09:02:31 crc kubenswrapper[4813]: I1201 09:02:31.747114 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mqr8l\" (UniqueName: \"kubernetes.io/projected/11bd67f4-fce9-479e-858b-44e6b1829d81-kube-api-access-mqr8l\") pod \"11bd67f4-fce9-479e-858b-44e6b1829d81\" (UID: \"11bd67f4-fce9-479e-858b-44e6b1829d81\") " Dec 01 09:02:31 crc kubenswrapper[4813]: I1201 09:02:31.747227 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/11bd67f4-fce9-479e-858b-44e6b1829d81-operator-scripts\") pod \"11bd67f4-fce9-479e-858b-44e6b1829d81\" (UID: \"11bd67f4-fce9-479e-858b-44e6b1829d81\") " Dec 01 09:02:31 crc kubenswrapper[4813]: I1201 09:02:31.747944 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/11bd67f4-fce9-479e-858b-44e6b1829d81-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "11bd67f4-fce9-479e-858b-44e6b1829d81" (UID: "11bd67f4-fce9-479e-858b-44e6b1829d81"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 09:02:31 crc kubenswrapper[4813]: I1201 09:02:31.755733 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/11bd67f4-fce9-479e-858b-44e6b1829d81-kube-api-access-mqr8l" (OuterVolumeSpecName: "kube-api-access-mqr8l") pod "11bd67f4-fce9-479e-858b-44e6b1829d81" (UID: "11bd67f4-fce9-479e-858b-44e6b1829d81"). InnerVolumeSpecName "kube-api-access-mqr8l". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:02:31 crc kubenswrapper[4813]: I1201 09:02:31.850164 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mqr8l\" (UniqueName: \"kubernetes.io/projected/11bd67f4-fce9-479e-858b-44e6b1829d81-kube-api-access-mqr8l\") on node \"crc\" DevicePath \"\"" Dec 01 09:02:31 crc kubenswrapper[4813]: I1201 09:02:31.850208 4813 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/11bd67f4-fce9-479e-858b-44e6b1829d81-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 09:02:32 crc kubenswrapper[4813]: I1201 09:02:32.333700 4813 generic.go:334] "Generic (PLEG): container finished" podID="d3dfa26e-2774-4129-9bfc-5cf40aa3e942" containerID="8250ba203c774f9ed961312dac84a6a807f187353d50fe2580fe5ee85b4bd6f1" exitCode=0 Dec 01 09:02:32 crc kubenswrapper[4813]: I1201 09:02:32.333790 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-c6e3-account-create-update-h46sk" event={"ID":"d3dfa26e-2774-4129-9bfc-5cf40aa3e942","Type":"ContainerDied","Data":"8250ba203c774f9ed961312dac84a6a807f187353d50fe2580fe5ee85b4bd6f1"} Dec 01 09:02:32 crc kubenswrapper[4813]: I1201 09:02:32.335678 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-create-p5mf4" event={"ID":"11bd67f4-fce9-479e-858b-44e6b1829d81","Type":"ContainerDied","Data":"cc99195c4f7f71b9836b14f0f672c3d0e6206959faa6675fe011fc7191c6e879"} Dec 01 09:02:32 crc kubenswrapper[4813]: I1201 09:02:32.335723 4813 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cc99195c4f7f71b9836b14f0f672c3d0e6206959faa6675fe011fc7191c6e879" Dec 01 09:02:32 crc kubenswrapper[4813]: I1201 09:02:32.335773 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-p5mf4" Dec 01 09:02:33 crc kubenswrapper[4813]: I1201 09:02:33.706641 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-c6e3-account-create-update-h46sk" Dec 01 09:02:33 crc kubenswrapper[4813]: I1201 09:02:33.881822 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x9vvj\" (UniqueName: \"kubernetes.io/projected/d3dfa26e-2774-4129-9bfc-5cf40aa3e942-kube-api-access-x9vvj\") pod \"d3dfa26e-2774-4129-9bfc-5cf40aa3e942\" (UID: \"d3dfa26e-2774-4129-9bfc-5cf40aa3e942\") " Dec 01 09:02:33 crc kubenswrapper[4813]: I1201 09:02:33.882208 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d3dfa26e-2774-4129-9bfc-5cf40aa3e942-operator-scripts\") pod \"d3dfa26e-2774-4129-9bfc-5cf40aa3e942\" (UID: \"d3dfa26e-2774-4129-9bfc-5cf40aa3e942\") " Dec 01 09:02:33 crc kubenswrapper[4813]: I1201 09:02:33.882762 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d3dfa26e-2774-4129-9bfc-5cf40aa3e942-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "d3dfa26e-2774-4129-9bfc-5cf40aa3e942" (UID: "d3dfa26e-2774-4129-9bfc-5cf40aa3e942"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 09:02:33 crc kubenswrapper[4813]: I1201 09:02:33.888972 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d3dfa26e-2774-4129-9bfc-5cf40aa3e942-kube-api-access-x9vvj" (OuterVolumeSpecName: "kube-api-access-x9vvj") pod "d3dfa26e-2774-4129-9bfc-5cf40aa3e942" (UID: "d3dfa26e-2774-4129-9bfc-5cf40aa3e942"). InnerVolumeSpecName "kube-api-access-x9vvj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:02:33 crc kubenswrapper[4813]: I1201 09:02:33.983623 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x9vvj\" (UniqueName: \"kubernetes.io/projected/d3dfa26e-2774-4129-9bfc-5cf40aa3e942-kube-api-access-x9vvj\") on node \"crc\" DevicePath \"\"" Dec 01 09:02:33 crc kubenswrapper[4813]: I1201 09:02:33.983679 4813 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d3dfa26e-2774-4129-9bfc-5cf40aa3e942-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 09:02:34 crc kubenswrapper[4813]: I1201 09:02:34.364701 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-c6e3-account-create-update-h46sk" event={"ID":"d3dfa26e-2774-4129-9bfc-5cf40aa3e942","Type":"ContainerDied","Data":"f680eaa6f7ff9cb4ba5e6ea00eb6213f16fc703dd6cc282b7d58bdd19da4ae06"} Dec 01 09:02:34 crc kubenswrapper[4813]: I1201 09:02:34.364744 4813 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f680eaa6f7ff9cb4ba5e6ea00eb6213f16fc703dd6cc282b7d58bdd19da4ae06" Dec 01 09:02:34 crc kubenswrapper[4813]: I1201 09:02:34.364858 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-c6e3-account-create-update-h46sk" Dec 01 09:02:37 crc kubenswrapper[4813]: I1201 09:02:37.427640 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-db-sync-rdsjx"] Dec 01 09:02:37 crc kubenswrapper[4813]: E1201 09:02:37.429239 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="11bd67f4-fce9-479e-858b-44e6b1829d81" containerName="mariadb-database-create" Dec 01 09:02:37 crc kubenswrapper[4813]: I1201 09:02:37.429377 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="11bd67f4-fce9-479e-858b-44e6b1829d81" containerName="mariadb-database-create" Dec 01 09:02:37 crc kubenswrapper[4813]: E1201 09:02:37.429467 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d3dfa26e-2774-4129-9bfc-5cf40aa3e942" containerName="mariadb-account-create-update" Dec 01 09:02:37 crc kubenswrapper[4813]: I1201 09:02:37.429534 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="d3dfa26e-2774-4129-9bfc-5cf40aa3e942" containerName="mariadb-account-create-update" Dec 01 09:02:37 crc kubenswrapper[4813]: I1201 09:02:37.429860 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="d3dfa26e-2774-4129-9bfc-5cf40aa3e942" containerName="mariadb-account-create-update" Dec 01 09:02:37 crc kubenswrapper[4813]: I1201 09:02:37.429955 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="11bd67f4-fce9-479e-858b-44e6b1829d81" containerName="mariadb-database-create" Dec 01 09:02:37 crc kubenswrapper[4813]: I1201 09:02:37.430928 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-rdsjx" Dec 01 09:02:37 crc kubenswrapper[4813]: I1201 09:02:37.434238 4813 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-glance-dockercfg-l9t5b" Dec 01 09:02:37 crc kubenswrapper[4813]: I1201 09:02:37.434289 4813 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-config-data" Dec 01 09:02:37 crc kubenswrapper[4813]: I1201 09:02:37.436582 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-sync-rdsjx"] Dec 01 09:02:37 crc kubenswrapper[4813]: I1201 09:02:37.456939 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/9ffc7257-c24e-4234-a314-e9a61f9e436f-db-sync-config-data\") pod \"glance-db-sync-rdsjx\" (UID: \"9ffc7257-c24e-4234-a314-e9a61f9e436f\") " pod="glance-kuttl-tests/glance-db-sync-rdsjx" Dec 01 09:02:37 crc kubenswrapper[4813]: I1201 09:02:37.457003 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9ffc7257-c24e-4234-a314-e9a61f9e436f-config-data\") pod \"glance-db-sync-rdsjx\" (UID: \"9ffc7257-c24e-4234-a314-e9a61f9e436f\") " pod="glance-kuttl-tests/glance-db-sync-rdsjx" Dec 01 09:02:37 crc kubenswrapper[4813]: I1201 09:02:37.457038 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wgh56\" (UniqueName: \"kubernetes.io/projected/9ffc7257-c24e-4234-a314-e9a61f9e436f-kube-api-access-wgh56\") pod \"glance-db-sync-rdsjx\" (UID: \"9ffc7257-c24e-4234-a314-e9a61f9e436f\") " pod="glance-kuttl-tests/glance-db-sync-rdsjx" Dec 01 09:02:37 crc kubenswrapper[4813]: I1201 09:02:37.558406 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/9ffc7257-c24e-4234-a314-e9a61f9e436f-db-sync-config-data\") pod \"glance-db-sync-rdsjx\" (UID: \"9ffc7257-c24e-4234-a314-e9a61f9e436f\") " pod="glance-kuttl-tests/glance-db-sync-rdsjx" Dec 01 09:02:37 crc kubenswrapper[4813]: I1201 09:02:37.558480 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9ffc7257-c24e-4234-a314-e9a61f9e436f-config-data\") pod \"glance-db-sync-rdsjx\" (UID: \"9ffc7257-c24e-4234-a314-e9a61f9e436f\") " pod="glance-kuttl-tests/glance-db-sync-rdsjx" Dec 01 09:02:37 crc kubenswrapper[4813]: I1201 09:02:37.558543 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wgh56\" (UniqueName: \"kubernetes.io/projected/9ffc7257-c24e-4234-a314-e9a61f9e436f-kube-api-access-wgh56\") pod \"glance-db-sync-rdsjx\" (UID: \"9ffc7257-c24e-4234-a314-e9a61f9e436f\") " pod="glance-kuttl-tests/glance-db-sync-rdsjx" Dec 01 09:02:37 crc kubenswrapper[4813]: I1201 09:02:37.563394 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/9ffc7257-c24e-4234-a314-e9a61f9e436f-db-sync-config-data\") pod \"glance-db-sync-rdsjx\" (UID: \"9ffc7257-c24e-4234-a314-e9a61f9e436f\") " pod="glance-kuttl-tests/glance-db-sync-rdsjx" Dec 01 09:02:37 crc kubenswrapper[4813]: I1201 09:02:37.563593 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9ffc7257-c24e-4234-a314-e9a61f9e436f-config-data\") pod \"glance-db-sync-rdsjx\" (UID: \"9ffc7257-c24e-4234-a314-e9a61f9e436f\") " pod="glance-kuttl-tests/glance-db-sync-rdsjx" Dec 01 09:02:37 crc kubenswrapper[4813]: I1201 09:02:37.591887 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wgh56\" (UniqueName: \"kubernetes.io/projected/9ffc7257-c24e-4234-a314-e9a61f9e436f-kube-api-access-wgh56\") pod \"glance-db-sync-rdsjx\" (UID: \"9ffc7257-c24e-4234-a314-e9a61f9e436f\") " pod="glance-kuttl-tests/glance-db-sync-rdsjx" Dec 01 09:02:37 crc kubenswrapper[4813]: I1201 09:02:37.764523 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-rdsjx" Dec 01 09:02:38 crc kubenswrapper[4813]: I1201 09:02:38.026262 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-sync-rdsjx"] Dec 01 09:02:38 crc kubenswrapper[4813]: I1201 09:02:38.406304 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-rdsjx" event={"ID":"9ffc7257-c24e-4234-a314-e9a61f9e436f","Type":"ContainerStarted","Data":"8abc35702c6ddfa143baaff38b2f9485f9407c64b17b505722fc8df9e3dbd409"} Dec 01 09:02:39 crc kubenswrapper[4813]: I1201 09:02:39.418614 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-rdsjx" event={"ID":"9ffc7257-c24e-4234-a314-e9a61f9e436f","Type":"ContainerStarted","Data":"b13bd94f99ec83c1cf6a08e3272145e5d16fcbaa439ac316c8433d38d61ccf67"} Dec 01 09:02:39 crc kubenswrapper[4813]: I1201 09:02:39.440405 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-db-sync-rdsjx" podStartSLOduration=2.440363769 podStartE2EDuration="2.440363769s" podCreationTimestamp="2025-12-01 09:02:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 09:02:39.436308535 +0000 UTC m=+1344.059230131" watchObservedRunningTime="2025-12-01 09:02:39.440363769 +0000 UTC m=+1344.063285355" Dec 01 09:02:42 crc kubenswrapper[4813]: I1201 09:02:42.451434 4813 generic.go:334] "Generic (PLEG): container finished" podID="9ffc7257-c24e-4234-a314-e9a61f9e436f" containerID="b13bd94f99ec83c1cf6a08e3272145e5d16fcbaa439ac316c8433d38d61ccf67" exitCode=0 Dec 01 09:02:42 crc kubenswrapper[4813]: I1201 09:02:42.451517 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-rdsjx" event={"ID":"9ffc7257-c24e-4234-a314-e9a61f9e436f","Type":"ContainerDied","Data":"b13bd94f99ec83c1cf6a08e3272145e5d16fcbaa439ac316c8433d38d61ccf67"} Dec 01 09:02:43 crc kubenswrapper[4813]: I1201 09:02:43.846254 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-rdsjx" Dec 01 09:02:44 crc kubenswrapper[4813]: I1201 09:02:44.001441 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wgh56\" (UniqueName: \"kubernetes.io/projected/9ffc7257-c24e-4234-a314-e9a61f9e436f-kube-api-access-wgh56\") pod \"9ffc7257-c24e-4234-a314-e9a61f9e436f\" (UID: \"9ffc7257-c24e-4234-a314-e9a61f9e436f\") " Dec 01 09:02:44 crc kubenswrapper[4813]: I1201 09:02:44.001587 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9ffc7257-c24e-4234-a314-e9a61f9e436f-config-data\") pod \"9ffc7257-c24e-4234-a314-e9a61f9e436f\" (UID: \"9ffc7257-c24e-4234-a314-e9a61f9e436f\") " Dec 01 09:02:44 crc kubenswrapper[4813]: I1201 09:02:44.001657 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/9ffc7257-c24e-4234-a314-e9a61f9e436f-db-sync-config-data\") pod \"9ffc7257-c24e-4234-a314-e9a61f9e436f\" (UID: \"9ffc7257-c24e-4234-a314-e9a61f9e436f\") " Dec 01 09:02:44 crc kubenswrapper[4813]: I1201 09:02:44.007602 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9ffc7257-c24e-4234-a314-e9a61f9e436f-kube-api-access-wgh56" (OuterVolumeSpecName: "kube-api-access-wgh56") pod "9ffc7257-c24e-4234-a314-e9a61f9e436f" (UID: "9ffc7257-c24e-4234-a314-e9a61f9e436f"). InnerVolumeSpecName "kube-api-access-wgh56". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:02:44 crc kubenswrapper[4813]: I1201 09:02:44.008165 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9ffc7257-c24e-4234-a314-e9a61f9e436f-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "9ffc7257-c24e-4234-a314-e9a61f9e436f" (UID: "9ffc7257-c24e-4234-a314-e9a61f9e436f"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:02:44 crc kubenswrapper[4813]: I1201 09:02:44.054774 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9ffc7257-c24e-4234-a314-e9a61f9e436f-config-data" (OuterVolumeSpecName: "config-data") pod "9ffc7257-c24e-4234-a314-e9a61f9e436f" (UID: "9ffc7257-c24e-4234-a314-e9a61f9e436f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:02:44 crc kubenswrapper[4813]: I1201 09:02:44.102897 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wgh56\" (UniqueName: \"kubernetes.io/projected/9ffc7257-c24e-4234-a314-e9a61f9e436f-kube-api-access-wgh56\") on node \"crc\" DevicePath \"\"" Dec 01 09:02:44 crc kubenswrapper[4813]: I1201 09:02:44.102927 4813 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9ffc7257-c24e-4234-a314-e9a61f9e436f-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 09:02:44 crc kubenswrapper[4813]: I1201 09:02:44.102937 4813 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/9ffc7257-c24e-4234-a314-e9a61f9e436f-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 09:02:44 crc kubenswrapper[4813]: I1201 09:02:44.472904 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-rdsjx" event={"ID":"9ffc7257-c24e-4234-a314-e9a61f9e436f","Type":"ContainerDied","Data":"8abc35702c6ddfa143baaff38b2f9485f9407c64b17b505722fc8df9e3dbd409"} Dec 01 09:02:44 crc kubenswrapper[4813]: I1201 09:02:44.472976 4813 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8abc35702c6ddfa143baaff38b2f9485f9407c64b17b505722fc8df9e3dbd409" Dec 01 09:02:44 crc kubenswrapper[4813]: I1201 09:02:44.473511 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-rdsjx" Dec 01 09:02:45 crc kubenswrapper[4813]: I1201 09:02:45.817757 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Dec 01 09:02:45 crc kubenswrapper[4813]: E1201 09:02:45.818399 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9ffc7257-c24e-4234-a314-e9a61f9e436f" containerName="glance-db-sync" Dec 01 09:02:45 crc kubenswrapper[4813]: I1201 09:02:45.818419 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="9ffc7257-c24e-4234-a314-e9a61f9e436f" containerName="glance-db-sync" Dec 01 09:02:45 crc kubenswrapper[4813]: I1201 09:02:45.818615 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="9ffc7257-c24e-4234-a314-e9a61f9e436f" containerName="glance-db-sync" Dec 01 09:02:45 crc kubenswrapper[4813]: I1201 09:02:45.819390 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Dec 01 09:02:45 crc kubenswrapper[4813]: I1201 09:02:45.821871 4813 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-scripts" Dec 01 09:02:45 crc kubenswrapper[4813]: I1201 09:02:45.821879 4813 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-default-single-config-data" Dec 01 09:02:45 crc kubenswrapper[4813]: I1201 09:02:45.821932 4813 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-glance-dockercfg-l9t5b" Dec 01 09:02:45 crc kubenswrapper[4813]: I1201 09:02:45.834059 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Dec 01 09:02:45 crc kubenswrapper[4813]: I1201 09:02:45.932532 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/c16ea1be-8dc4-42c6-b4b6-e8ae21599dd3-run\") pod \"glance-default-single-0\" (UID: \"c16ea1be-8dc4-42c6-b4b6-e8ae21599dd3\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 01 09:02:45 crc kubenswrapper[4813]: I1201 09:02:45.932574 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c16ea1be-8dc4-42c6-b4b6-e8ae21599dd3-scripts\") pod \"glance-default-single-0\" (UID: \"c16ea1be-8dc4-42c6-b4b6-e8ae21599dd3\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 01 09:02:45 crc kubenswrapper[4813]: I1201 09:02:45.932601 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage16-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage16-crc\") pod \"glance-default-single-0\" (UID: \"c16ea1be-8dc4-42c6-b4b6-e8ae21599dd3\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 01 09:02:45 crc kubenswrapper[4813]: I1201 09:02:45.932628 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/c16ea1be-8dc4-42c6-b4b6-e8ae21599dd3-dev\") pod \"glance-default-single-0\" (UID: \"c16ea1be-8dc4-42c6-b4b6-e8ae21599dd3\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 01 09:02:45 crc kubenswrapper[4813]: I1201 09:02:45.932661 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9znwm\" (UniqueName: \"kubernetes.io/projected/c16ea1be-8dc4-42c6-b4b6-e8ae21599dd3-kube-api-access-9znwm\") pod \"glance-default-single-0\" (UID: \"c16ea1be-8dc4-42c6-b4b6-e8ae21599dd3\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 01 09:02:45 crc kubenswrapper[4813]: I1201 09:02:45.932765 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/c16ea1be-8dc4-42c6-b4b6-e8ae21599dd3-lib-modules\") pod \"glance-default-single-0\" (UID: \"c16ea1be-8dc4-42c6-b4b6-e8ae21599dd3\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 01 09:02:45 crc kubenswrapper[4813]: I1201 09:02:45.932800 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/c16ea1be-8dc4-42c6-b4b6-e8ae21599dd3-var-locks-brick\") pod \"glance-default-single-0\" (UID: \"c16ea1be-8dc4-42c6-b4b6-e8ae21599dd3\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 01 09:02:45 crc kubenswrapper[4813]: I1201 09:02:45.932832 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/c16ea1be-8dc4-42c6-b4b6-e8ae21599dd3-sys\") pod \"glance-default-single-0\" (UID: \"c16ea1be-8dc4-42c6-b4b6-e8ae21599dd3\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 01 09:02:45 crc kubenswrapper[4813]: I1201 09:02:45.932876 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/c16ea1be-8dc4-42c6-b4b6-e8ae21599dd3-etc-nvme\") pod \"glance-default-single-0\" (UID: \"c16ea1be-8dc4-42c6-b4b6-e8ae21599dd3\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 01 09:02:45 crc kubenswrapper[4813]: I1201 09:02:45.932909 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c16ea1be-8dc4-42c6-b4b6-e8ae21599dd3-config-data\") pod \"glance-default-single-0\" (UID: \"c16ea1be-8dc4-42c6-b4b6-e8ae21599dd3\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 01 09:02:45 crc kubenswrapper[4813]: I1201 09:02:45.932927 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/c16ea1be-8dc4-42c6-b4b6-e8ae21599dd3-httpd-run\") pod \"glance-default-single-0\" (UID: \"c16ea1be-8dc4-42c6-b4b6-e8ae21599dd3\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 01 09:02:45 crc kubenswrapper[4813]: I1201 09:02:45.933066 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-single-0\" (UID: \"c16ea1be-8dc4-42c6-b4b6-e8ae21599dd3\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 01 09:02:45 crc kubenswrapper[4813]: I1201 09:02:45.933139 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/c16ea1be-8dc4-42c6-b4b6-e8ae21599dd3-etc-iscsi\") pod \"glance-default-single-0\" (UID: \"c16ea1be-8dc4-42c6-b4b6-e8ae21599dd3\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 01 09:02:45 crc kubenswrapper[4813]: I1201 09:02:45.933223 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c16ea1be-8dc4-42c6-b4b6-e8ae21599dd3-logs\") pod \"glance-default-single-0\" (UID: \"c16ea1be-8dc4-42c6-b4b6-e8ae21599dd3\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 01 09:02:46 crc kubenswrapper[4813]: I1201 09:02:46.034754 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c16ea1be-8dc4-42c6-b4b6-e8ae21599dd3-logs\") pod \"glance-default-single-0\" (UID: \"c16ea1be-8dc4-42c6-b4b6-e8ae21599dd3\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 01 09:02:46 crc kubenswrapper[4813]: I1201 09:02:46.034867 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/c16ea1be-8dc4-42c6-b4b6-e8ae21599dd3-run\") pod \"glance-default-single-0\" (UID: \"c16ea1be-8dc4-42c6-b4b6-e8ae21599dd3\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 01 09:02:46 crc kubenswrapper[4813]: I1201 09:02:46.034903 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c16ea1be-8dc4-42c6-b4b6-e8ae21599dd3-scripts\") pod \"glance-default-single-0\" (UID: \"c16ea1be-8dc4-42c6-b4b6-e8ae21599dd3\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 01 09:02:46 crc kubenswrapper[4813]: I1201 09:02:46.034947 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage16-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage16-crc\") pod \"glance-default-single-0\" (UID: \"c16ea1be-8dc4-42c6-b4b6-e8ae21599dd3\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 01 09:02:46 crc kubenswrapper[4813]: I1201 09:02:46.035034 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/c16ea1be-8dc4-42c6-b4b6-e8ae21599dd3-dev\") pod \"glance-default-single-0\" (UID: \"c16ea1be-8dc4-42c6-b4b6-e8ae21599dd3\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 01 09:02:46 crc kubenswrapper[4813]: I1201 09:02:46.035047 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/c16ea1be-8dc4-42c6-b4b6-e8ae21599dd3-run\") pod \"glance-default-single-0\" (UID: \"c16ea1be-8dc4-42c6-b4b6-e8ae21599dd3\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 01 09:02:46 crc kubenswrapper[4813]: I1201 09:02:46.035100 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9znwm\" (UniqueName: \"kubernetes.io/projected/c16ea1be-8dc4-42c6-b4b6-e8ae21599dd3-kube-api-access-9znwm\") pod \"glance-default-single-0\" (UID: \"c16ea1be-8dc4-42c6-b4b6-e8ae21599dd3\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 01 09:02:46 crc kubenswrapper[4813]: I1201 09:02:46.035158 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/c16ea1be-8dc4-42c6-b4b6-e8ae21599dd3-dev\") pod \"glance-default-single-0\" (UID: \"c16ea1be-8dc4-42c6-b4b6-e8ae21599dd3\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 01 09:02:46 crc kubenswrapper[4813]: I1201 09:02:46.035165 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/c16ea1be-8dc4-42c6-b4b6-e8ae21599dd3-lib-modules\") pod \"glance-default-single-0\" (UID: \"c16ea1be-8dc4-42c6-b4b6-e8ae21599dd3\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 01 09:02:46 crc kubenswrapper[4813]: I1201 09:02:46.035224 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/c16ea1be-8dc4-42c6-b4b6-e8ae21599dd3-var-locks-brick\") pod \"glance-default-single-0\" (UID: \"c16ea1be-8dc4-42c6-b4b6-e8ae21599dd3\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 01 09:02:46 crc kubenswrapper[4813]: I1201 09:02:46.035261 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/c16ea1be-8dc4-42c6-b4b6-e8ae21599dd3-sys\") pod \"glance-default-single-0\" (UID: \"c16ea1be-8dc4-42c6-b4b6-e8ae21599dd3\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 01 09:02:46 crc kubenswrapper[4813]: I1201 09:02:46.035300 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/c16ea1be-8dc4-42c6-b4b6-e8ae21599dd3-etc-nvme\") pod \"glance-default-single-0\" (UID: \"c16ea1be-8dc4-42c6-b4b6-e8ae21599dd3\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 01 09:02:46 crc kubenswrapper[4813]: I1201 09:02:46.035345 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c16ea1be-8dc4-42c6-b4b6-e8ae21599dd3-config-data\") pod \"glance-default-single-0\" (UID: \"c16ea1be-8dc4-42c6-b4b6-e8ae21599dd3\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 01 09:02:46 crc kubenswrapper[4813]: I1201 09:02:46.035384 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/c16ea1be-8dc4-42c6-b4b6-e8ae21599dd3-httpd-run\") pod \"glance-default-single-0\" (UID: \"c16ea1be-8dc4-42c6-b4b6-e8ae21599dd3\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 01 09:02:46 crc kubenswrapper[4813]: I1201 09:02:46.035397 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/c16ea1be-8dc4-42c6-b4b6-e8ae21599dd3-var-locks-brick\") pod \"glance-default-single-0\" (UID: \"c16ea1be-8dc4-42c6-b4b6-e8ae21599dd3\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 01 09:02:46 crc kubenswrapper[4813]: I1201 09:02:46.035431 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/c16ea1be-8dc4-42c6-b4b6-e8ae21599dd3-sys\") pod \"glance-default-single-0\" (UID: \"c16ea1be-8dc4-42c6-b4b6-e8ae21599dd3\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 01 09:02:46 crc kubenswrapper[4813]: I1201 09:02:46.035460 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/c16ea1be-8dc4-42c6-b4b6-e8ae21599dd3-etc-nvme\") pod \"glance-default-single-0\" (UID: \"c16ea1be-8dc4-42c6-b4b6-e8ae21599dd3\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 01 09:02:46 crc kubenswrapper[4813]: I1201 09:02:46.035458 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-single-0\" (UID: \"c16ea1be-8dc4-42c6-b4b6-e8ae21599dd3\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 01 09:02:46 crc kubenswrapper[4813]: I1201 09:02:46.035567 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/c16ea1be-8dc4-42c6-b4b6-e8ae21599dd3-etc-iscsi\") pod \"glance-default-single-0\" (UID: \"c16ea1be-8dc4-42c6-b4b6-e8ae21599dd3\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 01 09:02:46 crc kubenswrapper[4813]: I1201 09:02:46.035450 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/c16ea1be-8dc4-42c6-b4b6-e8ae21599dd3-lib-modules\") pod \"glance-default-single-0\" (UID: \"c16ea1be-8dc4-42c6-b4b6-e8ae21599dd3\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 01 09:02:46 crc kubenswrapper[4813]: I1201 09:02:46.035630 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/c16ea1be-8dc4-42c6-b4b6-e8ae21599dd3-etc-iscsi\") pod \"glance-default-single-0\" (UID: \"c16ea1be-8dc4-42c6-b4b6-e8ae21599dd3\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 01 09:02:46 crc kubenswrapper[4813]: I1201 09:02:46.035748 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c16ea1be-8dc4-42c6-b4b6-e8ae21599dd3-logs\") pod \"glance-default-single-0\" (UID: \"c16ea1be-8dc4-42c6-b4b6-e8ae21599dd3\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 01 09:02:46 crc kubenswrapper[4813]: I1201 09:02:46.035882 4813 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-single-0\" (UID: \"c16ea1be-8dc4-42c6-b4b6-e8ae21599dd3\") device mount path \"/mnt/openstack/pv10\"" pod="glance-kuttl-tests/glance-default-single-0" Dec 01 09:02:46 crc kubenswrapper[4813]: I1201 09:02:46.036143 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/c16ea1be-8dc4-42c6-b4b6-e8ae21599dd3-httpd-run\") pod \"glance-default-single-0\" (UID: \"c16ea1be-8dc4-42c6-b4b6-e8ae21599dd3\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 01 09:02:46 crc kubenswrapper[4813]: I1201 09:02:46.036212 4813 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage16-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage16-crc\") pod \"glance-default-single-0\" (UID: \"c16ea1be-8dc4-42c6-b4b6-e8ae21599dd3\") device mount path \"/mnt/openstack/pv16\"" pod="glance-kuttl-tests/glance-default-single-0" Dec 01 09:02:46 crc kubenswrapper[4813]: I1201 09:02:46.041787 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c16ea1be-8dc4-42c6-b4b6-e8ae21599dd3-scripts\") pod \"glance-default-single-0\" (UID: \"c16ea1be-8dc4-42c6-b4b6-e8ae21599dd3\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 01 09:02:46 crc kubenswrapper[4813]: I1201 09:02:46.043306 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c16ea1be-8dc4-42c6-b4b6-e8ae21599dd3-config-data\") pod \"glance-default-single-0\" (UID: \"c16ea1be-8dc4-42c6-b4b6-e8ae21599dd3\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 01 09:02:46 crc kubenswrapper[4813]: I1201 09:02:46.058429 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage16-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage16-crc\") pod \"glance-default-single-0\" (UID: \"c16ea1be-8dc4-42c6-b4b6-e8ae21599dd3\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 01 09:02:46 crc kubenswrapper[4813]: I1201 09:02:46.061461 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-single-0\" (UID: \"c16ea1be-8dc4-42c6-b4b6-e8ae21599dd3\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 01 09:02:46 crc kubenswrapper[4813]: I1201 09:02:46.066496 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9znwm\" (UniqueName: \"kubernetes.io/projected/c16ea1be-8dc4-42c6-b4b6-e8ae21599dd3-kube-api-access-9znwm\") pod \"glance-default-single-0\" (UID: \"c16ea1be-8dc4-42c6-b4b6-e8ae21599dd3\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 01 09:02:46 crc kubenswrapper[4813]: I1201 09:02:46.134438 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Dec 01 09:02:46 crc kubenswrapper[4813]: I1201 09:02:46.348561 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Dec 01 09:02:46 crc kubenswrapper[4813]: I1201 09:02:46.614134 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Dec 01 09:02:47 crc kubenswrapper[4813]: I1201 09:02:47.496385 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"c16ea1be-8dc4-42c6-b4b6-e8ae21599dd3","Type":"ContainerStarted","Data":"7872a60003eba784309e708e78c34dfbeca82f81efdd26bc20330a83ad3e3671"} Dec 01 09:02:47 crc kubenswrapper[4813]: I1201 09:02:47.496891 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-single-0" podUID="c16ea1be-8dc4-42c6-b4b6-e8ae21599dd3" containerName="glance-httpd" containerID="cri-o://7872a60003eba784309e708e78c34dfbeca82f81efdd26bc20330a83ad3e3671" gracePeriod=30 Dec 01 09:02:47 crc kubenswrapper[4813]: I1201 09:02:47.496901 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"c16ea1be-8dc4-42c6-b4b6-e8ae21599dd3","Type":"ContainerStarted","Data":"d880b020eacbb12975c9e06f834822a028126d54b98d675ea85e1e5c2a643d33"} Dec 01 09:02:47 crc kubenswrapper[4813]: I1201 09:02:47.496956 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"c16ea1be-8dc4-42c6-b4b6-e8ae21599dd3","Type":"ContainerStarted","Data":"ac8daff08f176d5d1049750befcd355863589e0bba040b0e00832d3e27f697b2"} Dec 01 09:02:47 crc kubenswrapper[4813]: I1201 09:02:47.496717 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-single-0" podUID="c16ea1be-8dc4-42c6-b4b6-e8ae21599dd3" containerName="glance-log" containerID="cri-o://d880b020eacbb12975c9e06f834822a028126d54b98d675ea85e1e5c2a643d33" gracePeriod=30 Dec 01 09:02:47 crc kubenswrapper[4813]: I1201 09:02:47.527925 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-default-single-0" podStartSLOduration=3.527905256 podStartE2EDuration="3.527905256s" podCreationTimestamp="2025-12-01 09:02:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 09:02:47.524943842 +0000 UTC m=+1352.147865428" watchObservedRunningTime="2025-12-01 09:02:47.527905256 +0000 UTC m=+1352.150826842" Dec 01 09:02:48 crc kubenswrapper[4813]: I1201 09:02:48.169242 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Dec 01 09:02:48 crc kubenswrapper[4813]: I1201 09:02:48.308111 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/c16ea1be-8dc4-42c6-b4b6-e8ae21599dd3-httpd-run\") pod \"c16ea1be-8dc4-42c6-b4b6-e8ae21599dd3\" (UID: \"c16ea1be-8dc4-42c6-b4b6-e8ae21599dd3\") " Dec 01 09:02:48 crc kubenswrapper[4813]: I1201 09:02:48.308184 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/c16ea1be-8dc4-42c6-b4b6-e8ae21599dd3-etc-nvme\") pod \"c16ea1be-8dc4-42c6-b4b6-e8ae21599dd3\" (UID: \"c16ea1be-8dc4-42c6-b4b6-e8ae21599dd3\") " Dec 01 09:02:48 crc kubenswrapper[4813]: I1201 09:02:48.308218 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9znwm\" (UniqueName: \"kubernetes.io/projected/c16ea1be-8dc4-42c6-b4b6-e8ae21599dd3-kube-api-access-9znwm\") pod \"c16ea1be-8dc4-42c6-b4b6-e8ae21599dd3\" (UID: \"c16ea1be-8dc4-42c6-b4b6-e8ae21599dd3\") " Dec 01 09:02:48 crc kubenswrapper[4813]: I1201 09:02:48.308269 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/c16ea1be-8dc4-42c6-b4b6-e8ae21599dd3-run\") pod \"c16ea1be-8dc4-42c6-b4b6-e8ae21599dd3\" (UID: \"c16ea1be-8dc4-42c6-b4b6-e8ae21599dd3\") " Dec 01 09:02:48 crc kubenswrapper[4813]: I1201 09:02:48.308319 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c16ea1be-8dc4-42c6-b4b6-e8ae21599dd3-logs\") pod \"c16ea1be-8dc4-42c6-b4b6-e8ae21599dd3\" (UID: \"c16ea1be-8dc4-42c6-b4b6-e8ae21599dd3\") " Dec 01 09:02:48 crc kubenswrapper[4813]: I1201 09:02:48.308345 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/c16ea1be-8dc4-42c6-b4b6-e8ae21599dd3-var-locks-brick\") pod \"c16ea1be-8dc4-42c6-b4b6-e8ae21599dd3\" (UID: \"c16ea1be-8dc4-42c6-b4b6-e8ae21599dd3\") " Dec 01 09:02:48 crc kubenswrapper[4813]: I1201 09:02:48.308378 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage16-crc\") pod \"c16ea1be-8dc4-42c6-b4b6-e8ae21599dd3\" (UID: \"c16ea1be-8dc4-42c6-b4b6-e8ae21599dd3\") " Dec 01 09:02:48 crc kubenswrapper[4813]: I1201 09:02:48.308401 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c16ea1be-8dc4-42c6-b4b6-e8ae21599dd3-run" (OuterVolumeSpecName: "run") pod "c16ea1be-8dc4-42c6-b4b6-e8ae21599dd3" (UID: "c16ea1be-8dc4-42c6-b4b6-e8ae21599dd3"). InnerVolumeSpecName "run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 09:02:48 crc kubenswrapper[4813]: I1201 09:02:48.308427 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c16ea1be-8dc4-42c6-b4b6-e8ae21599dd3-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "c16ea1be-8dc4-42c6-b4b6-e8ae21599dd3" (UID: "c16ea1be-8dc4-42c6-b4b6-e8ae21599dd3"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 09:02:48 crc kubenswrapper[4813]: I1201 09:02:48.308453 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c16ea1be-8dc4-42c6-b4b6-e8ae21599dd3-scripts\") pod \"c16ea1be-8dc4-42c6-b4b6-e8ae21599dd3\" (UID: \"c16ea1be-8dc4-42c6-b4b6-e8ae21599dd3\") " Dec 01 09:02:48 crc kubenswrapper[4813]: I1201 09:02:48.308476 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance-cache\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"c16ea1be-8dc4-42c6-b4b6-e8ae21599dd3\" (UID: \"c16ea1be-8dc4-42c6-b4b6-e8ae21599dd3\") " Dec 01 09:02:48 crc kubenswrapper[4813]: I1201 09:02:48.308476 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c16ea1be-8dc4-42c6-b4b6-e8ae21599dd3-var-locks-brick" (OuterVolumeSpecName: "var-locks-brick") pod "c16ea1be-8dc4-42c6-b4b6-e8ae21599dd3" (UID: "c16ea1be-8dc4-42c6-b4b6-e8ae21599dd3"). InnerVolumeSpecName "var-locks-brick". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 09:02:48 crc kubenswrapper[4813]: I1201 09:02:48.308540 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/c16ea1be-8dc4-42c6-b4b6-e8ae21599dd3-lib-modules\") pod \"c16ea1be-8dc4-42c6-b4b6-e8ae21599dd3\" (UID: \"c16ea1be-8dc4-42c6-b4b6-e8ae21599dd3\") " Dec 01 09:02:48 crc kubenswrapper[4813]: I1201 09:02:48.308575 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/c16ea1be-8dc4-42c6-b4b6-e8ae21599dd3-dev\") pod \"c16ea1be-8dc4-42c6-b4b6-e8ae21599dd3\" (UID: \"c16ea1be-8dc4-42c6-b4b6-e8ae21599dd3\") " Dec 01 09:02:48 crc kubenswrapper[4813]: I1201 09:02:48.308627 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/c16ea1be-8dc4-42c6-b4b6-e8ae21599dd3-sys\") pod \"c16ea1be-8dc4-42c6-b4b6-e8ae21599dd3\" (UID: \"c16ea1be-8dc4-42c6-b4b6-e8ae21599dd3\") " Dec 01 09:02:48 crc kubenswrapper[4813]: I1201 09:02:48.308663 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c16ea1be-8dc4-42c6-b4b6-e8ae21599dd3-config-data\") pod \"c16ea1be-8dc4-42c6-b4b6-e8ae21599dd3\" (UID: \"c16ea1be-8dc4-42c6-b4b6-e8ae21599dd3\") " Dec 01 09:02:48 crc kubenswrapper[4813]: I1201 09:02:48.308683 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/c16ea1be-8dc4-42c6-b4b6-e8ae21599dd3-etc-iscsi\") pod \"c16ea1be-8dc4-42c6-b4b6-e8ae21599dd3\" (UID: \"c16ea1be-8dc4-42c6-b4b6-e8ae21599dd3\") " Dec 01 09:02:48 crc kubenswrapper[4813]: I1201 09:02:48.308822 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c16ea1be-8dc4-42c6-b4b6-e8ae21599dd3-logs" (OuterVolumeSpecName: "logs") pod "c16ea1be-8dc4-42c6-b4b6-e8ae21599dd3" (UID: "c16ea1be-8dc4-42c6-b4b6-e8ae21599dd3"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 09:02:48 crc kubenswrapper[4813]: I1201 09:02:48.308870 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c16ea1be-8dc4-42c6-b4b6-e8ae21599dd3-etc-iscsi" (OuterVolumeSpecName: "etc-iscsi") pod "c16ea1be-8dc4-42c6-b4b6-e8ae21599dd3" (UID: "c16ea1be-8dc4-42c6-b4b6-e8ae21599dd3"). InnerVolumeSpecName "etc-iscsi". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 09:02:48 crc kubenswrapper[4813]: I1201 09:02:48.308859 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c16ea1be-8dc4-42c6-b4b6-e8ae21599dd3-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "c16ea1be-8dc4-42c6-b4b6-e8ae21599dd3" (UID: "c16ea1be-8dc4-42c6-b4b6-e8ae21599dd3"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 09:02:48 crc kubenswrapper[4813]: I1201 09:02:48.308906 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c16ea1be-8dc4-42c6-b4b6-e8ae21599dd3-dev" (OuterVolumeSpecName: "dev") pod "c16ea1be-8dc4-42c6-b4b6-e8ae21599dd3" (UID: "c16ea1be-8dc4-42c6-b4b6-e8ae21599dd3"). InnerVolumeSpecName "dev". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 09:02:48 crc kubenswrapper[4813]: I1201 09:02:48.308923 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c16ea1be-8dc4-42c6-b4b6-e8ae21599dd3-sys" (OuterVolumeSpecName: "sys") pod "c16ea1be-8dc4-42c6-b4b6-e8ae21599dd3" (UID: "c16ea1be-8dc4-42c6-b4b6-e8ae21599dd3"). InnerVolumeSpecName "sys". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 09:02:48 crc kubenswrapper[4813]: I1201 09:02:48.308478 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c16ea1be-8dc4-42c6-b4b6-e8ae21599dd3-etc-nvme" (OuterVolumeSpecName: "etc-nvme") pod "c16ea1be-8dc4-42c6-b4b6-e8ae21599dd3" (UID: "c16ea1be-8dc4-42c6-b4b6-e8ae21599dd3"). InnerVolumeSpecName "etc-nvme". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 09:02:48 crc kubenswrapper[4813]: I1201 09:02:48.313498 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage16-crc" (OuterVolumeSpecName: "glance") pod "c16ea1be-8dc4-42c6-b4b6-e8ae21599dd3" (UID: "c16ea1be-8dc4-42c6-b4b6-e8ae21599dd3"). InnerVolumeSpecName "local-storage16-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 01 09:02:48 crc kubenswrapper[4813]: I1201 09:02:48.318688 4813 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/c16ea1be-8dc4-42c6-b4b6-e8ae21599dd3-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 01 09:02:48 crc kubenswrapper[4813]: I1201 09:02:48.318714 4813 reconciler_common.go:293] "Volume detached for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/c16ea1be-8dc4-42c6-b4b6-e8ae21599dd3-etc-nvme\") on node \"crc\" DevicePath \"\"" Dec 01 09:02:48 crc kubenswrapper[4813]: I1201 09:02:48.318725 4813 reconciler_common.go:293] "Volume detached for volume \"run\" (UniqueName: \"kubernetes.io/host-path/c16ea1be-8dc4-42c6-b4b6-e8ae21599dd3-run\") on node \"crc\" DevicePath \"\"" Dec 01 09:02:48 crc kubenswrapper[4813]: I1201 09:02:48.318733 4813 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c16ea1be-8dc4-42c6-b4b6-e8ae21599dd3-logs\") on node \"crc\" DevicePath \"\"" Dec 01 09:02:48 crc kubenswrapper[4813]: I1201 09:02:48.318741 4813 reconciler_common.go:293] "Volume detached for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/c16ea1be-8dc4-42c6-b4b6-e8ae21599dd3-var-locks-brick\") on node \"crc\" DevicePath \"\"" Dec 01 09:02:48 crc kubenswrapper[4813]: I1201 09:02:48.318796 4813 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage16-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage16-crc\") on node \"crc\" " Dec 01 09:02:48 crc kubenswrapper[4813]: I1201 09:02:48.318808 4813 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/c16ea1be-8dc4-42c6-b4b6-e8ae21599dd3-lib-modules\") on node \"crc\" DevicePath \"\"" Dec 01 09:02:48 crc kubenswrapper[4813]: I1201 09:02:48.318816 4813 reconciler_common.go:293] "Volume detached for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/c16ea1be-8dc4-42c6-b4b6-e8ae21599dd3-dev\") on node \"crc\" DevicePath \"\"" Dec 01 09:02:48 crc kubenswrapper[4813]: I1201 09:02:48.318824 4813 reconciler_common.go:293] "Volume detached for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/c16ea1be-8dc4-42c6-b4b6-e8ae21599dd3-sys\") on node \"crc\" DevicePath \"\"" Dec 01 09:02:48 crc kubenswrapper[4813]: I1201 09:02:48.318834 4813 reconciler_common.go:293] "Volume detached for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/c16ea1be-8dc4-42c6-b4b6-e8ae21599dd3-etc-iscsi\") on node \"crc\" DevicePath \"\"" Dec 01 09:02:48 crc kubenswrapper[4813]: I1201 09:02:48.322704 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c16ea1be-8dc4-42c6-b4b6-e8ae21599dd3-scripts" (OuterVolumeSpecName: "scripts") pod "c16ea1be-8dc4-42c6-b4b6-e8ae21599dd3" (UID: "c16ea1be-8dc4-42c6-b4b6-e8ae21599dd3"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:02:48 crc kubenswrapper[4813]: I1201 09:02:48.323591 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage10-crc" (OuterVolumeSpecName: "glance-cache") pod "c16ea1be-8dc4-42c6-b4b6-e8ae21599dd3" (UID: "c16ea1be-8dc4-42c6-b4b6-e8ae21599dd3"). InnerVolumeSpecName "local-storage10-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 01 09:02:48 crc kubenswrapper[4813]: I1201 09:02:48.324194 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c16ea1be-8dc4-42c6-b4b6-e8ae21599dd3-kube-api-access-9znwm" (OuterVolumeSpecName: "kube-api-access-9znwm") pod "c16ea1be-8dc4-42c6-b4b6-e8ae21599dd3" (UID: "c16ea1be-8dc4-42c6-b4b6-e8ae21599dd3"). InnerVolumeSpecName "kube-api-access-9znwm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:02:48 crc kubenswrapper[4813]: I1201 09:02:48.333450 4813 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage16-crc" (UniqueName: "kubernetes.io/local-volume/local-storage16-crc") on node "crc" Dec 01 09:02:48 crc kubenswrapper[4813]: I1201 09:02:48.354339 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c16ea1be-8dc4-42c6-b4b6-e8ae21599dd3-config-data" (OuterVolumeSpecName: "config-data") pod "c16ea1be-8dc4-42c6-b4b6-e8ae21599dd3" (UID: "c16ea1be-8dc4-42c6-b4b6-e8ae21599dd3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:02:48 crc kubenswrapper[4813]: I1201 09:02:48.419747 4813 reconciler_common.go:293] "Volume detached for volume \"local-storage16-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage16-crc\") on node \"crc\" DevicePath \"\"" Dec 01 09:02:48 crc kubenswrapper[4813]: I1201 09:02:48.420052 4813 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" " Dec 01 09:02:48 crc kubenswrapper[4813]: I1201 09:02:48.420070 4813 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c16ea1be-8dc4-42c6-b4b6-e8ae21599dd3-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 09:02:48 crc kubenswrapper[4813]: I1201 09:02:48.420083 4813 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c16ea1be-8dc4-42c6-b4b6-e8ae21599dd3-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 09:02:48 crc kubenswrapper[4813]: I1201 09:02:48.420095 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9znwm\" (UniqueName: \"kubernetes.io/projected/c16ea1be-8dc4-42c6-b4b6-e8ae21599dd3-kube-api-access-9znwm\") on node \"crc\" DevicePath \"\"" Dec 01 09:02:48 crc kubenswrapper[4813]: W1201 09:02:48.423232 4813 watcher.go:93] Error while processing event ("/sys/fs/cgroup/user.slice/user-0.slice/session-c49.scope": 0x40000100 == IN_CREATE|IN_ISDIR): readdirent /sys/fs/cgroup/user.slice/user-0.slice/session-c49.scope: no such file or directory Dec 01 09:02:48 crc kubenswrapper[4813]: I1201 09:02:48.437014 4813 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage10-crc" (UniqueName: "kubernetes.io/local-volume/local-storage10-crc") on node "crc" Dec 01 09:02:48 crc kubenswrapper[4813]: I1201 09:02:48.510180 4813 generic.go:334] "Generic (PLEG): container finished" podID="c16ea1be-8dc4-42c6-b4b6-e8ae21599dd3" containerID="7872a60003eba784309e708e78c34dfbeca82f81efdd26bc20330a83ad3e3671" exitCode=143 Dec 01 09:02:48 crc kubenswrapper[4813]: I1201 09:02:48.510218 4813 generic.go:334] "Generic (PLEG): container finished" podID="c16ea1be-8dc4-42c6-b4b6-e8ae21599dd3" containerID="d880b020eacbb12975c9e06f834822a028126d54b98d675ea85e1e5c2a643d33" exitCode=143 Dec 01 09:02:48 crc kubenswrapper[4813]: I1201 09:02:48.510244 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"c16ea1be-8dc4-42c6-b4b6-e8ae21599dd3","Type":"ContainerDied","Data":"7872a60003eba784309e708e78c34dfbeca82f81efdd26bc20330a83ad3e3671"} Dec 01 09:02:48 crc kubenswrapper[4813]: I1201 09:02:48.510288 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"c16ea1be-8dc4-42c6-b4b6-e8ae21599dd3","Type":"ContainerDied","Data":"d880b020eacbb12975c9e06f834822a028126d54b98d675ea85e1e5c2a643d33"} Dec 01 09:02:48 crc kubenswrapper[4813]: I1201 09:02:48.510303 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"c16ea1be-8dc4-42c6-b4b6-e8ae21599dd3","Type":"ContainerDied","Data":"ac8daff08f176d5d1049750befcd355863589e0bba040b0e00832d3e27f697b2"} Dec 01 09:02:48 crc kubenswrapper[4813]: I1201 09:02:48.510348 4813 scope.go:117] "RemoveContainer" containerID="7872a60003eba784309e708e78c34dfbeca82f81efdd26bc20330a83ad3e3671" Dec 01 09:02:48 crc kubenswrapper[4813]: I1201 09:02:48.510491 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Dec 01 09:02:48 crc kubenswrapper[4813]: I1201 09:02:48.521355 4813 reconciler_common.go:293] "Volume detached for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" DevicePath \"\"" Dec 01 09:02:48 crc kubenswrapper[4813]: I1201 09:02:48.537519 4813 scope.go:117] "RemoveContainer" containerID="d880b020eacbb12975c9e06f834822a028126d54b98d675ea85e1e5c2a643d33" Dec 01 09:02:48 crc kubenswrapper[4813]: I1201 09:02:48.546036 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Dec 01 09:02:48 crc kubenswrapper[4813]: I1201 09:02:48.560704 4813 scope.go:117] "RemoveContainer" containerID="7872a60003eba784309e708e78c34dfbeca82f81efdd26bc20330a83ad3e3671" Dec 01 09:02:48 crc kubenswrapper[4813]: E1201 09:02:48.561515 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7872a60003eba784309e708e78c34dfbeca82f81efdd26bc20330a83ad3e3671\": container with ID starting with 7872a60003eba784309e708e78c34dfbeca82f81efdd26bc20330a83ad3e3671 not found: ID does not exist" containerID="7872a60003eba784309e708e78c34dfbeca82f81efdd26bc20330a83ad3e3671" Dec 01 09:02:48 crc kubenswrapper[4813]: I1201 09:02:48.561563 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7872a60003eba784309e708e78c34dfbeca82f81efdd26bc20330a83ad3e3671"} err="failed to get container status \"7872a60003eba784309e708e78c34dfbeca82f81efdd26bc20330a83ad3e3671\": rpc error: code = NotFound desc = could not find container \"7872a60003eba784309e708e78c34dfbeca82f81efdd26bc20330a83ad3e3671\": container with ID starting with 7872a60003eba784309e708e78c34dfbeca82f81efdd26bc20330a83ad3e3671 not found: ID does not exist" Dec 01 09:02:48 crc kubenswrapper[4813]: I1201 09:02:48.561606 4813 scope.go:117] "RemoveContainer" containerID="d880b020eacbb12975c9e06f834822a028126d54b98d675ea85e1e5c2a643d33" Dec 01 09:02:48 crc kubenswrapper[4813]: E1201 09:02:48.562159 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d880b020eacbb12975c9e06f834822a028126d54b98d675ea85e1e5c2a643d33\": container with ID starting with d880b020eacbb12975c9e06f834822a028126d54b98d675ea85e1e5c2a643d33 not found: ID does not exist" containerID="d880b020eacbb12975c9e06f834822a028126d54b98d675ea85e1e5c2a643d33" Dec 01 09:02:48 crc kubenswrapper[4813]: I1201 09:02:48.562197 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d880b020eacbb12975c9e06f834822a028126d54b98d675ea85e1e5c2a643d33"} err="failed to get container status \"d880b020eacbb12975c9e06f834822a028126d54b98d675ea85e1e5c2a643d33\": rpc error: code = NotFound desc = could not find container \"d880b020eacbb12975c9e06f834822a028126d54b98d675ea85e1e5c2a643d33\": container with ID starting with d880b020eacbb12975c9e06f834822a028126d54b98d675ea85e1e5c2a643d33 not found: ID does not exist" Dec 01 09:02:48 crc kubenswrapper[4813]: I1201 09:02:48.562227 4813 scope.go:117] "RemoveContainer" containerID="7872a60003eba784309e708e78c34dfbeca82f81efdd26bc20330a83ad3e3671" Dec 01 09:02:48 crc kubenswrapper[4813]: I1201 09:02:48.562529 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7872a60003eba784309e708e78c34dfbeca82f81efdd26bc20330a83ad3e3671"} err="failed to get container status \"7872a60003eba784309e708e78c34dfbeca82f81efdd26bc20330a83ad3e3671\": rpc error: code = NotFound desc = could not find container \"7872a60003eba784309e708e78c34dfbeca82f81efdd26bc20330a83ad3e3671\": container with ID starting with 7872a60003eba784309e708e78c34dfbeca82f81efdd26bc20330a83ad3e3671 not found: ID does not exist" Dec 01 09:02:48 crc kubenswrapper[4813]: I1201 09:02:48.562559 4813 scope.go:117] "RemoveContainer" containerID="d880b020eacbb12975c9e06f834822a028126d54b98d675ea85e1e5c2a643d33" Dec 01 09:02:48 crc kubenswrapper[4813]: I1201 09:02:48.562835 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d880b020eacbb12975c9e06f834822a028126d54b98d675ea85e1e5c2a643d33"} err="failed to get container status \"d880b020eacbb12975c9e06f834822a028126d54b98d675ea85e1e5c2a643d33\": rpc error: code = NotFound desc = could not find container \"d880b020eacbb12975c9e06f834822a028126d54b98d675ea85e1e5c2a643d33\": container with ID starting with d880b020eacbb12975c9e06f834822a028126d54b98d675ea85e1e5c2a643d33 not found: ID does not exist" Dec 01 09:02:48 crc kubenswrapper[4813]: I1201 09:02:48.565143 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Dec 01 09:02:48 crc kubenswrapper[4813]: I1201 09:02:48.576051 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Dec 01 09:02:48 crc kubenswrapper[4813]: E1201 09:02:48.576493 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c16ea1be-8dc4-42c6-b4b6-e8ae21599dd3" containerName="glance-httpd" Dec 01 09:02:48 crc kubenswrapper[4813]: I1201 09:02:48.576512 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="c16ea1be-8dc4-42c6-b4b6-e8ae21599dd3" containerName="glance-httpd" Dec 01 09:02:48 crc kubenswrapper[4813]: E1201 09:02:48.576532 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c16ea1be-8dc4-42c6-b4b6-e8ae21599dd3" containerName="glance-log" Dec 01 09:02:48 crc kubenswrapper[4813]: I1201 09:02:48.576538 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="c16ea1be-8dc4-42c6-b4b6-e8ae21599dd3" containerName="glance-log" Dec 01 09:02:48 crc kubenswrapper[4813]: I1201 09:02:48.576942 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="c16ea1be-8dc4-42c6-b4b6-e8ae21599dd3" containerName="glance-httpd" Dec 01 09:02:48 crc kubenswrapper[4813]: I1201 09:02:48.577010 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="c16ea1be-8dc4-42c6-b4b6-e8ae21599dd3" containerName="glance-log" Dec 01 09:02:48 crc kubenswrapper[4813]: I1201 09:02:48.578896 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Dec 01 09:02:48 crc kubenswrapper[4813]: I1201 09:02:48.582694 4813 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-glance-dockercfg-l9t5b" Dec 01 09:02:48 crc kubenswrapper[4813]: I1201 09:02:48.583092 4813 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-default-single-config-data" Dec 01 09:02:48 crc kubenswrapper[4813]: I1201 09:02:48.583926 4813 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-scripts" Dec 01 09:02:48 crc kubenswrapper[4813]: I1201 09:02:48.589572 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Dec 01 09:02:48 crc kubenswrapper[4813]: I1201 09:02:48.724473 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/9a0e144a-6298-4092-86fc-ba45b986d6ec-etc-iscsi\") pod \"glance-default-single-0\" (UID: \"9a0e144a-6298-4092-86fc-ba45b986d6ec\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 01 09:02:48 crc kubenswrapper[4813]: I1201 09:02:48.724512 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/9a0e144a-6298-4092-86fc-ba45b986d6ec-lib-modules\") pod \"glance-default-single-0\" (UID: \"9a0e144a-6298-4092-86fc-ba45b986d6ec\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 01 09:02:48 crc kubenswrapper[4813]: I1201 09:02:48.724566 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/9a0e144a-6298-4092-86fc-ba45b986d6ec-run\") pod \"glance-default-single-0\" (UID: \"9a0e144a-6298-4092-86fc-ba45b986d6ec\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 01 09:02:48 crc kubenswrapper[4813]: I1201 09:02:48.724822 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/9a0e144a-6298-4092-86fc-ba45b986d6ec-sys\") pod \"glance-default-single-0\" (UID: \"9a0e144a-6298-4092-86fc-ba45b986d6ec\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 01 09:02:48 crc kubenswrapper[4813]: I1201 09:02:48.724898 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/9a0e144a-6298-4092-86fc-ba45b986d6ec-dev\") pod \"glance-default-single-0\" (UID: \"9a0e144a-6298-4092-86fc-ba45b986d6ec\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 01 09:02:48 crc kubenswrapper[4813]: I1201 09:02:48.724991 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/9a0e144a-6298-4092-86fc-ba45b986d6ec-httpd-run\") pod \"glance-default-single-0\" (UID: \"9a0e144a-6298-4092-86fc-ba45b986d6ec\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 01 09:02:48 crc kubenswrapper[4813]: I1201 09:02:48.725083 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9a0e144a-6298-4092-86fc-ba45b986d6ec-config-data\") pod \"glance-default-single-0\" (UID: \"9a0e144a-6298-4092-86fc-ba45b986d6ec\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 01 09:02:48 crc kubenswrapper[4813]: I1201 09:02:48.725141 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-single-0\" (UID: \"9a0e144a-6298-4092-86fc-ba45b986d6ec\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 01 09:02:48 crc kubenswrapper[4813]: I1201 09:02:48.725196 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/9a0e144a-6298-4092-86fc-ba45b986d6ec-var-locks-brick\") pod \"glance-default-single-0\" (UID: \"9a0e144a-6298-4092-86fc-ba45b986d6ec\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 01 09:02:48 crc kubenswrapper[4813]: I1201 09:02:48.725238 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tj5p6\" (UniqueName: \"kubernetes.io/projected/9a0e144a-6298-4092-86fc-ba45b986d6ec-kube-api-access-tj5p6\") pod \"glance-default-single-0\" (UID: \"9a0e144a-6298-4092-86fc-ba45b986d6ec\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 01 09:02:48 crc kubenswrapper[4813]: I1201 09:02:48.725330 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/9a0e144a-6298-4092-86fc-ba45b986d6ec-etc-nvme\") pod \"glance-default-single-0\" (UID: \"9a0e144a-6298-4092-86fc-ba45b986d6ec\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 01 09:02:48 crc kubenswrapper[4813]: I1201 09:02:48.725443 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9a0e144a-6298-4092-86fc-ba45b986d6ec-scripts\") pod \"glance-default-single-0\" (UID: \"9a0e144a-6298-4092-86fc-ba45b986d6ec\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 01 09:02:48 crc kubenswrapper[4813]: I1201 09:02:48.725475 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9a0e144a-6298-4092-86fc-ba45b986d6ec-logs\") pod \"glance-default-single-0\" (UID: \"9a0e144a-6298-4092-86fc-ba45b986d6ec\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 01 09:02:48 crc kubenswrapper[4813]: I1201 09:02:48.725546 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage16-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage16-crc\") pod \"glance-default-single-0\" (UID: \"9a0e144a-6298-4092-86fc-ba45b986d6ec\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 01 09:02:48 crc kubenswrapper[4813]: I1201 09:02:48.827587 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/9a0e144a-6298-4092-86fc-ba45b986d6ec-etc-iscsi\") pod \"glance-default-single-0\" (UID: \"9a0e144a-6298-4092-86fc-ba45b986d6ec\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 01 09:02:48 crc kubenswrapper[4813]: I1201 09:02:48.827491 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/9a0e144a-6298-4092-86fc-ba45b986d6ec-etc-iscsi\") pod \"glance-default-single-0\" (UID: \"9a0e144a-6298-4092-86fc-ba45b986d6ec\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 01 09:02:48 crc kubenswrapper[4813]: I1201 09:02:48.828374 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/9a0e144a-6298-4092-86fc-ba45b986d6ec-lib-modules\") pod \"glance-default-single-0\" (UID: \"9a0e144a-6298-4092-86fc-ba45b986d6ec\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 01 09:02:48 crc kubenswrapper[4813]: I1201 09:02:48.828534 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/9a0e144a-6298-4092-86fc-ba45b986d6ec-lib-modules\") pod \"glance-default-single-0\" (UID: \"9a0e144a-6298-4092-86fc-ba45b986d6ec\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 01 09:02:48 crc kubenswrapper[4813]: I1201 09:02:48.828691 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/9a0e144a-6298-4092-86fc-ba45b986d6ec-run\") pod \"glance-default-single-0\" (UID: \"9a0e144a-6298-4092-86fc-ba45b986d6ec\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 01 09:02:48 crc kubenswrapper[4813]: I1201 09:02:48.828816 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/9a0e144a-6298-4092-86fc-ba45b986d6ec-run\") pod \"glance-default-single-0\" (UID: \"9a0e144a-6298-4092-86fc-ba45b986d6ec\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 01 09:02:48 crc kubenswrapper[4813]: I1201 09:02:48.829079 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/9a0e144a-6298-4092-86fc-ba45b986d6ec-sys\") pod \"glance-default-single-0\" (UID: \"9a0e144a-6298-4092-86fc-ba45b986d6ec\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 01 09:02:48 crc kubenswrapper[4813]: I1201 09:02:48.829282 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/9a0e144a-6298-4092-86fc-ba45b986d6ec-dev\") pod \"glance-default-single-0\" (UID: \"9a0e144a-6298-4092-86fc-ba45b986d6ec\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 01 09:02:48 crc kubenswrapper[4813]: I1201 09:02:48.829478 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/9a0e144a-6298-4092-86fc-ba45b986d6ec-httpd-run\") pod \"glance-default-single-0\" (UID: \"9a0e144a-6298-4092-86fc-ba45b986d6ec\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 01 09:02:48 crc kubenswrapper[4813]: I1201 09:02:48.829649 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9a0e144a-6298-4092-86fc-ba45b986d6ec-config-data\") pod \"glance-default-single-0\" (UID: \"9a0e144a-6298-4092-86fc-ba45b986d6ec\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 01 09:02:48 crc kubenswrapper[4813]: I1201 09:02:48.829190 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/9a0e144a-6298-4092-86fc-ba45b986d6ec-sys\") pod \"glance-default-single-0\" (UID: \"9a0e144a-6298-4092-86fc-ba45b986d6ec\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 01 09:02:48 crc kubenswrapper[4813]: I1201 09:02:48.829980 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/9a0e144a-6298-4092-86fc-ba45b986d6ec-httpd-run\") pod \"glance-default-single-0\" (UID: \"9a0e144a-6298-4092-86fc-ba45b986d6ec\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 01 09:02:48 crc kubenswrapper[4813]: I1201 09:02:48.829836 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-single-0\" (UID: \"9a0e144a-6298-4092-86fc-ba45b986d6ec\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 01 09:02:48 crc kubenswrapper[4813]: I1201 09:02:48.829364 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/9a0e144a-6298-4092-86fc-ba45b986d6ec-dev\") pod \"glance-default-single-0\" (UID: \"9a0e144a-6298-4092-86fc-ba45b986d6ec\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 01 09:02:48 crc kubenswrapper[4813]: I1201 09:02:48.830083 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/9a0e144a-6298-4092-86fc-ba45b986d6ec-var-locks-brick\") pod \"glance-default-single-0\" (UID: \"9a0e144a-6298-4092-86fc-ba45b986d6ec\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 01 09:02:48 crc kubenswrapper[4813]: I1201 09:02:48.830126 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tj5p6\" (UniqueName: \"kubernetes.io/projected/9a0e144a-6298-4092-86fc-ba45b986d6ec-kube-api-access-tj5p6\") pod \"glance-default-single-0\" (UID: \"9a0e144a-6298-4092-86fc-ba45b986d6ec\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 01 09:02:48 crc kubenswrapper[4813]: I1201 09:02:48.830211 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/9a0e144a-6298-4092-86fc-ba45b986d6ec-etc-nvme\") pod \"glance-default-single-0\" (UID: \"9a0e144a-6298-4092-86fc-ba45b986d6ec\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 01 09:02:48 crc kubenswrapper[4813]: I1201 09:02:48.830242 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/9a0e144a-6298-4092-86fc-ba45b986d6ec-var-locks-brick\") pod \"glance-default-single-0\" (UID: \"9a0e144a-6298-4092-86fc-ba45b986d6ec\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 01 09:02:48 crc kubenswrapper[4813]: I1201 09:02:48.830339 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/9a0e144a-6298-4092-86fc-ba45b986d6ec-etc-nvme\") pod \"glance-default-single-0\" (UID: \"9a0e144a-6298-4092-86fc-ba45b986d6ec\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 01 09:02:48 crc kubenswrapper[4813]: I1201 09:02:48.830347 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9a0e144a-6298-4092-86fc-ba45b986d6ec-scripts\") pod \"glance-default-single-0\" (UID: \"9a0e144a-6298-4092-86fc-ba45b986d6ec\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 01 09:02:48 crc kubenswrapper[4813]: I1201 09:02:48.830439 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9a0e144a-6298-4092-86fc-ba45b986d6ec-logs\") pod \"glance-default-single-0\" (UID: \"9a0e144a-6298-4092-86fc-ba45b986d6ec\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 01 09:02:48 crc kubenswrapper[4813]: I1201 09:02:48.830519 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage16-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage16-crc\") pod \"glance-default-single-0\" (UID: \"9a0e144a-6298-4092-86fc-ba45b986d6ec\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 01 09:02:48 crc kubenswrapper[4813]: I1201 09:02:48.830761 4813 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage16-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage16-crc\") pod \"glance-default-single-0\" (UID: \"9a0e144a-6298-4092-86fc-ba45b986d6ec\") device mount path \"/mnt/openstack/pv16\"" pod="glance-kuttl-tests/glance-default-single-0" Dec 01 09:02:48 crc kubenswrapper[4813]: I1201 09:02:48.831014 4813 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-single-0\" (UID: \"9a0e144a-6298-4092-86fc-ba45b986d6ec\") device mount path \"/mnt/openstack/pv10\"" pod="glance-kuttl-tests/glance-default-single-0" Dec 01 09:02:48 crc kubenswrapper[4813]: I1201 09:02:48.831370 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9a0e144a-6298-4092-86fc-ba45b986d6ec-logs\") pod \"glance-default-single-0\" (UID: \"9a0e144a-6298-4092-86fc-ba45b986d6ec\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 01 09:02:48 crc kubenswrapper[4813]: I1201 09:02:48.836380 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9a0e144a-6298-4092-86fc-ba45b986d6ec-scripts\") pod \"glance-default-single-0\" (UID: \"9a0e144a-6298-4092-86fc-ba45b986d6ec\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 01 09:02:48 crc kubenswrapper[4813]: I1201 09:02:48.836880 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9a0e144a-6298-4092-86fc-ba45b986d6ec-config-data\") pod \"glance-default-single-0\" (UID: \"9a0e144a-6298-4092-86fc-ba45b986d6ec\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 01 09:02:48 crc kubenswrapper[4813]: I1201 09:02:48.859143 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage16-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage16-crc\") pod \"glance-default-single-0\" (UID: \"9a0e144a-6298-4092-86fc-ba45b986d6ec\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 01 09:02:48 crc kubenswrapper[4813]: I1201 09:02:48.862615 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tj5p6\" (UniqueName: \"kubernetes.io/projected/9a0e144a-6298-4092-86fc-ba45b986d6ec-kube-api-access-tj5p6\") pod \"glance-default-single-0\" (UID: \"9a0e144a-6298-4092-86fc-ba45b986d6ec\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 01 09:02:48 crc kubenswrapper[4813]: I1201 09:02:48.872859 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-single-0\" (UID: \"9a0e144a-6298-4092-86fc-ba45b986d6ec\") " pod="glance-kuttl-tests/glance-default-single-0" Dec 01 09:02:48 crc kubenswrapper[4813]: I1201 09:02:48.900523 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Dec 01 09:02:49 crc kubenswrapper[4813]: I1201 09:02:49.346473 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Dec 01 09:02:49 crc kubenswrapper[4813]: I1201 09:02:49.524479 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"9a0e144a-6298-4092-86fc-ba45b986d6ec","Type":"ContainerStarted","Data":"9c4e446fc54267ec7444deb2562c0511cf39dac043ea8128c8e381d5edf0d1f4"} Dec 01 09:02:50 crc kubenswrapper[4813]: I1201 09:02:50.405315 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c16ea1be-8dc4-42c6-b4b6-e8ae21599dd3" path="/var/lib/kubelet/pods/c16ea1be-8dc4-42c6-b4b6-e8ae21599dd3/volumes" Dec 01 09:02:50 crc kubenswrapper[4813]: I1201 09:02:50.534470 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"9a0e144a-6298-4092-86fc-ba45b986d6ec","Type":"ContainerStarted","Data":"c0918867756a6c3013f771a8e9cbf8e78fd40a1e5f1cc6fa2576633699e9e100"} Dec 01 09:02:50 crc kubenswrapper[4813]: I1201 09:02:50.534515 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"9a0e144a-6298-4092-86fc-ba45b986d6ec","Type":"ContainerStarted","Data":"db306b76d1e4d93ad74ca563b646090f0237e7fdc176fc18b82ff6a6cc3e5413"} Dec 01 09:02:50 crc kubenswrapper[4813]: I1201 09:02:50.555927 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-default-single-0" podStartSLOduration=2.55590005 podStartE2EDuration="2.55590005s" podCreationTimestamp="2025-12-01 09:02:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 09:02:50.55516953 +0000 UTC m=+1355.178091146" watchObservedRunningTime="2025-12-01 09:02:50.55590005 +0000 UTC m=+1355.178821646" Dec 01 09:02:58 crc kubenswrapper[4813]: I1201 09:02:58.901635 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-single-0" Dec 01 09:02:58 crc kubenswrapper[4813]: I1201 09:02:58.902312 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-single-0" Dec 01 09:02:58 crc kubenswrapper[4813]: I1201 09:02:58.926275 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-single-0" Dec 01 09:02:58 crc kubenswrapper[4813]: I1201 09:02:58.943794 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-single-0" Dec 01 09:02:59 crc kubenswrapper[4813]: I1201 09:02:59.618632 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-single-0" Dec 01 09:02:59 crc kubenswrapper[4813]: I1201 09:02:59.618757 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-single-0" Dec 01 09:03:01 crc kubenswrapper[4813]: I1201 09:03:01.907929 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-single-0" Dec 01 09:03:01 crc kubenswrapper[4813]: I1201 09:03:01.909122 4813 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 01 09:03:02 crc kubenswrapper[4813]: I1201 09:03:02.089775 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-single-0" Dec 01 09:03:05 crc kubenswrapper[4813]: I1201 09:03:05.081600 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-single-2"] Dec 01 09:03:05 crc kubenswrapper[4813]: I1201 09:03:05.084087 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-2" Dec 01 09:03:05 crc kubenswrapper[4813]: I1201 09:03:05.088434 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-single-1"] Dec 01 09:03:05 crc kubenswrapper[4813]: I1201 09:03:05.089760 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-1" Dec 01 09:03:05 crc kubenswrapper[4813]: I1201 09:03:05.100626 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-single-2"] Dec 01 09:03:05 crc kubenswrapper[4813]: I1201 09:03:05.109830 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-single-1"] Dec 01 09:03:05 crc kubenswrapper[4813]: I1201 09:03:05.224444 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage20-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage20-crc\") pod \"glance-default-single-2\" (UID: \"9ea2713f-188e-4e62-9b40-16ec13591054\") " pod="glance-kuttl-tests/glance-default-single-2" Dec 01 09:03:05 crc kubenswrapper[4813]: I1201 09:03:05.224581 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/9ea2713f-188e-4e62-9b40-16ec13591054-sys\") pod \"glance-default-single-2\" (UID: \"9ea2713f-188e-4e62-9b40-16ec13591054\") " pod="glance-kuttl-tests/glance-default-single-2" Dec 01 09:03:05 crc kubenswrapper[4813]: I1201 09:03:05.224643 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage14-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage14-crc\") pod \"glance-default-single-1\" (UID: \"10f3f59e-3661-4248-aac9-8061818231ef\") " pod="glance-kuttl-tests/glance-default-single-1" Dec 01 09:03:05 crc kubenswrapper[4813]: I1201 09:03:05.224707 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/9ea2713f-188e-4e62-9b40-16ec13591054-dev\") pod \"glance-default-single-2\" (UID: \"9ea2713f-188e-4e62-9b40-16ec13591054\") " pod="glance-kuttl-tests/glance-default-single-2" Dec 01 09:03:05 crc kubenswrapper[4813]: I1201 09:03:05.224904 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/10f3f59e-3661-4248-aac9-8061818231ef-var-locks-brick\") pod \"glance-default-single-1\" (UID: \"10f3f59e-3661-4248-aac9-8061818231ef\") " pod="glance-kuttl-tests/glance-default-single-1" Dec 01 09:03:05 crc kubenswrapper[4813]: I1201 09:03:05.225013 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/9ea2713f-188e-4e62-9b40-16ec13591054-lib-modules\") pod \"glance-default-single-2\" (UID: \"9ea2713f-188e-4e62-9b40-16ec13591054\") " pod="glance-kuttl-tests/glance-default-single-2" Dec 01 09:03:05 crc kubenswrapper[4813]: I1201 09:03:05.225067 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/10f3f59e-3661-4248-aac9-8061818231ef-sys\") pod \"glance-default-single-1\" (UID: \"10f3f59e-3661-4248-aac9-8061818231ef\") " pod="glance-kuttl-tests/glance-default-single-1" Dec 01 09:03:05 crc kubenswrapper[4813]: I1201 09:03:05.225136 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4r7ld\" (UniqueName: \"kubernetes.io/projected/9ea2713f-188e-4e62-9b40-16ec13591054-kube-api-access-4r7ld\") pod \"glance-default-single-2\" (UID: \"9ea2713f-188e-4e62-9b40-16ec13591054\") " pod="glance-kuttl-tests/glance-default-single-2" Dec 01 09:03:05 crc kubenswrapper[4813]: I1201 09:03:05.225176 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/10f3f59e-3661-4248-aac9-8061818231ef-etc-nvme\") pod \"glance-default-single-1\" (UID: \"10f3f59e-3661-4248-aac9-8061818231ef\") " pod="glance-kuttl-tests/glance-default-single-1" Dec 01 09:03:05 crc kubenswrapper[4813]: I1201 09:03:05.225227 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-single-2\" (UID: \"9ea2713f-188e-4e62-9b40-16ec13591054\") " pod="glance-kuttl-tests/glance-default-single-2" Dec 01 09:03:05 crc kubenswrapper[4813]: I1201 09:03:05.225265 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9ea2713f-188e-4e62-9b40-16ec13591054-scripts\") pod \"glance-default-single-2\" (UID: \"9ea2713f-188e-4e62-9b40-16ec13591054\") " pod="glance-kuttl-tests/glance-default-single-2" Dec 01 09:03:05 crc kubenswrapper[4813]: I1201 09:03:05.225315 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/10f3f59e-3661-4248-aac9-8061818231ef-scripts\") pod \"glance-default-single-1\" (UID: \"10f3f59e-3661-4248-aac9-8061818231ef\") " pod="glance-kuttl-tests/glance-default-single-1" Dec 01 09:03:05 crc kubenswrapper[4813]: I1201 09:03:05.225355 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/10f3f59e-3661-4248-aac9-8061818231ef-dev\") pod \"glance-default-single-1\" (UID: \"10f3f59e-3661-4248-aac9-8061818231ef\") " pod="glance-kuttl-tests/glance-default-single-1" Dec 01 09:03:05 crc kubenswrapper[4813]: I1201 09:03:05.225464 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/10f3f59e-3661-4248-aac9-8061818231ef-etc-iscsi\") pod \"glance-default-single-1\" (UID: \"10f3f59e-3661-4248-aac9-8061818231ef\") " pod="glance-kuttl-tests/glance-default-single-1" Dec 01 09:03:05 crc kubenswrapper[4813]: I1201 09:03:05.225522 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/9ea2713f-188e-4e62-9b40-16ec13591054-httpd-run\") pod \"glance-default-single-2\" (UID: \"9ea2713f-188e-4e62-9b40-16ec13591054\") " pod="glance-kuttl-tests/glance-default-single-2" Dec 01 09:03:05 crc kubenswrapper[4813]: I1201 09:03:05.225573 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/9ea2713f-188e-4e62-9b40-16ec13591054-etc-iscsi\") pod \"glance-default-single-2\" (UID: \"9ea2713f-188e-4e62-9b40-16ec13591054\") " pod="glance-kuttl-tests/glance-default-single-2" Dec 01 09:03:05 crc kubenswrapper[4813]: I1201 09:03:05.225666 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/9ea2713f-188e-4e62-9b40-16ec13591054-run\") pod \"glance-default-single-2\" (UID: \"9ea2713f-188e-4e62-9b40-16ec13591054\") " pod="glance-kuttl-tests/glance-default-single-2" Dec 01 09:03:05 crc kubenswrapper[4813]: I1201 09:03:05.225742 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/10f3f59e-3661-4248-aac9-8061818231ef-config-data\") pod \"glance-default-single-1\" (UID: \"10f3f59e-3661-4248-aac9-8061818231ef\") " pod="glance-kuttl-tests/glance-default-single-1" Dec 01 09:03:05 crc kubenswrapper[4813]: I1201 09:03:05.225792 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/9ea2713f-188e-4e62-9b40-16ec13591054-etc-nvme\") pod \"glance-default-single-2\" (UID: \"9ea2713f-188e-4e62-9b40-16ec13591054\") " pod="glance-kuttl-tests/glance-default-single-2" Dec 01 09:03:05 crc kubenswrapper[4813]: I1201 09:03:05.225828 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/10f3f59e-3661-4248-aac9-8061818231ef-httpd-run\") pod \"glance-default-single-1\" (UID: \"10f3f59e-3661-4248-aac9-8061818231ef\") " pod="glance-kuttl-tests/glance-default-single-1" Dec 01 09:03:05 crc kubenswrapper[4813]: I1201 09:03:05.225875 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/9ea2713f-188e-4e62-9b40-16ec13591054-var-locks-brick\") pod \"glance-default-single-2\" (UID: \"9ea2713f-188e-4e62-9b40-16ec13591054\") " pod="glance-kuttl-tests/glance-default-single-2" Dec 01 09:03:05 crc kubenswrapper[4813]: I1201 09:03:05.226074 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/10f3f59e-3661-4248-aac9-8061818231ef-lib-modules\") pod \"glance-default-single-1\" (UID: \"10f3f59e-3661-4248-aac9-8061818231ef\") " pod="glance-kuttl-tests/glance-default-single-1" Dec 01 09:03:05 crc kubenswrapper[4813]: I1201 09:03:05.226142 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9ea2713f-188e-4e62-9b40-16ec13591054-config-data\") pod \"glance-default-single-2\" (UID: \"9ea2713f-188e-4e62-9b40-16ec13591054\") " pod="glance-kuttl-tests/glance-default-single-2" Dec 01 09:03:05 crc kubenswrapper[4813]: I1201 09:03:05.226196 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/10f3f59e-3661-4248-aac9-8061818231ef-logs\") pod \"glance-default-single-1\" (UID: \"10f3f59e-3661-4248-aac9-8061818231ef\") " pod="glance-kuttl-tests/glance-default-single-1" Dec 01 09:03:05 crc kubenswrapper[4813]: I1201 09:03:05.226269 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5tpkj\" (UniqueName: \"kubernetes.io/projected/10f3f59e-3661-4248-aac9-8061818231ef-kube-api-access-5tpkj\") pod \"glance-default-single-1\" (UID: \"10f3f59e-3661-4248-aac9-8061818231ef\") " pod="glance-kuttl-tests/glance-default-single-1" Dec 01 09:03:05 crc kubenswrapper[4813]: I1201 09:03:05.226446 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-single-1\" (UID: \"10f3f59e-3661-4248-aac9-8061818231ef\") " pod="glance-kuttl-tests/glance-default-single-1" Dec 01 09:03:05 crc kubenswrapper[4813]: I1201 09:03:05.226538 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9ea2713f-188e-4e62-9b40-16ec13591054-logs\") pod \"glance-default-single-2\" (UID: \"9ea2713f-188e-4e62-9b40-16ec13591054\") " pod="glance-kuttl-tests/glance-default-single-2" Dec 01 09:03:05 crc kubenswrapper[4813]: I1201 09:03:05.226615 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/10f3f59e-3661-4248-aac9-8061818231ef-run\") pod \"glance-default-single-1\" (UID: \"10f3f59e-3661-4248-aac9-8061818231ef\") " pod="glance-kuttl-tests/glance-default-single-1" Dec 01 09:03:05 crc kubenswrapper[4813]: I1201 09:03:05.328484 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/10f3f59e-3661-4248-aac9-8061818231ef-etc-iscsi\") pod \"glance-default-single-1\" (UID: \"10f3f59e-3661-4248-aac9-8061818231ef\") " pod="glance-kuttl-tests/glance-default-single-1" Dec 01 09:03:05 crc kubenswrapper[4813]: I1201 09:03:05.328530 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/10f3f59e-3661-4248-aac9-8061818231ef-dev\") pod \"glance-default-single-1\" (UID: \"10f3f59e-3661-4248-aac9-8061818231ef\") " pod="glance-kuttl-tests/glance-default-single-1" Dec 01 09:03:05 crc kubenswrapper[4813]: I1201 09:03:05.328568 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/9ea2713f-188e-4e62-9b40-16ec13591054-httpd-run\") pod \"glance-default-single-2\" (UID: \"9ea2713f-188e-4e62-9b40-16ec13591054\") " pod="glance-kuttl-tests/glance-default-single-2" Dec 01 09:03:05 crc kubenswrapper[4813]: I1201 09:03:05.328587 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/9ea2713f-188e-4e62-9b40-16ec13591054-etc-iscsi\") pod \"glance-default-single-2\" (UID: \"9ea2713f-188e-4e62-9b40-16ec13591054\") " pod="glance-kuttl-tests/glance-default-single-2" Dec 01 09:03:05 crc kubenswrapper[4813]: I1201 09:03:05.328619 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/9ea2713f-188e-4e62-9b40-16ec13591054-run\") pod \"glance-default-single-2\" (UID: \"9ea2713f-188e-4e62-9b40-16ec13591054\") " pod="glance-kuttl-tests/glance-default-single-2" Dec 01 09:03:05 crc kubenswrapper[4813]: I1201 09:03:05.328646 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/10f3f59e-3661-4248-aac9-8061818231ef-dev\") pod \"glance-default-single-1\" (UID: \"10f3f59e-3661-4248-aac9-8061818231ef\") " pod="glance-kuttl-tests/glance-default-single-1" Dec 01 09:03:05 crc kubenswrapper[4813]: I1201 09:03:05.328653 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/10f3f59e-3661-4248-aac9-8061818231ef-config-data\") pod \"glance-default-single-1\" (UID: \"10f3f59e-3661-4248-aac9-8061818231ef\") " pod="glance-kuttl-tests/glance-default-single-1" Dec 01 09:03:05 crc kubenswrapper[4813]: I1201 09:03:05.328679 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/9ea2713f-188e-4e62-9b40-16ec13591054-run\") pod \"glance-default-single-2\" (UID: \"9ea2713f-188e-4e62-9b40-16ec13591054\") " pod="glance-kuttl-tests/glance-default-single-2" Dec 01 09:03:05 crc kubenswrapper[4813]: I1201 09:03:05.328700 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/9ea2713f-188e-4e62-9b40-16ec13591054-etc-iscsi\") pod \"glance-default-single-2\" (UID: \"9ea2713f-188e-4e62-9b40-16ec13591054\") " pod="glance-kuttl-tests/glance-default-single-2" Dec 01 09:03:05 crc kubenswrapper[4813]: I1201 09:03:05.328682 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/9ea2713f-188e-4e62-9b40-16ec13591054-etc-nvme\") pod \"glance-default-single-2\" (UID: \"9ea2713f-188e-4e62-9b40-16ec13591054\") " pod="glance-kuttl-tests/glance-default-single-2" Dec 01 09:03:05 crc kubenswrapper[4813]: I1201 09:03:05.328740 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/9ea2713f-188e-4e62-9b40-16ec13591054-etc-nvme\") pod \"glance-default-single-2\" (UID: \"9ea2713f-188e-4e62-9b40-16ec13591054\") " pod="glance-kuttl-tests/glance-default-single-2" Dec 01 09:03:05 crc kubenswrapper[4813]: I1201 09:03:05.328748 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/10f3f59e-3661-4248-aac9-8061818231ef-httpd-run\") pod \"glance-default-single-1\" (UID: \"10f3f59e-3661-4248-aac9-8061818231ef\") " pod="glance-kuttl-tests/glance-default-single-1" Dec 01 09:03:05 crc kubenswrapper[4813]: I1201 09:03:05.328647 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/10f3f59e-3661-4248-aac9-8061818231ef-etc-iscsi\") pod \"glance-default-single-1\" (UID: \"10f3f59e-3661-4248-aac9-8061818231ef\") " pod="glance-kuttl-tests/glance-default-single-1" Dec 01 09:03:05 crc kubenswrapper[4813]: I1201 09:03:05.328796 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/9ea2713f-188e-4e62-9b40-16ec13591054-var-locks-brick\") pod \"glance-default-single-2\" (UID: \"9ea2713f-188e-4e62-9b40-16ec13591054\") " pod="glance-kuttl-tests/glance-default-single-2" Dec 01 09:03:05 crc kubenswrapper[4813]: I1201 09:03:05.328852 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/10f3f59e-3661-4248-aac9-8061818231ef-lib-modules\") pod \"glance-default-single-1\" (UID: \"10f3f59e-3661-4248-aac9-8061818231ef\") " pod="glance-kuttl-tests/glance-default-single-1" Dec 01 09:03:05 crc kubenswrapper[4813]: I1201 09:03:05.328878 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9ea2713f-188e-4e62-9b40-16ec13591054-config-data\") pod \"glance-default-single-2\" (UID: \"9ea2713f-188e-4e62-9b40-16ec13591054\") " pod="glance-kuttl-tests/glance-default-single-2" Dec 01 09:03:05 crc kubenswrapper[4813]: I1201 09:03:05.328894 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/9ea2713f-188e-4e62-9b40-16ec13591054-var-locks-brick\") pod \"glance-default-single-2\" (UID: \"9ea2713f-188e-4e62-9b40-16ec13591054\") " pod="glance-kuttl-tests/glance-default-single-2" Dec 01 09:03:05 crc kubenswrapper[4813]: I1201 09:03:05.328918 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/10f3f59e-3661-4248-aac9-8061818231ef-lib-modules\") pod \"glance-default-single-1\" (UID: \"10f3f59e-3661-4248-aac9-8061818231ef\") " pod="glance-kuttl-tests/glance-default-single-1" Dec 01 09:03:05 crc kubenswrapper[4813]: I1201 09:03:05.328939 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/10f3f59e-3661-4248-aac9-8061818231ef-logs\") pod \"glance-default-single-1\" (UID: \"10f3f59e-3661-4248-aac9-8061818231ef\") " pod="glance-kuttl-tests/glance-default-single-1" Dec 01 09:03:05 crc kubenswrapper[4813]: I1201 09:03:05.328991 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5tpkj\" (UniqueName: \"kubernetes.io/projected/10f3f59e-3661-4248-aac9-8061818231ef-kube-api-access-5tpkj\") pod \"glance-default-single-1\" (UID: \"10f3f59e-3661-4248-aac9-8061818231ef\") " pod="glance-kuttl-tests/glance-default-single-1" Dec 01 09:03:05 crc kubenswrapper[4813]: I1201 09:03:05.329013 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9ea2713f-188e-4e62-9b40-16ec13591054-logs\") pod \"glance-default-single-2\" (UID: \"9ea2713f-188e-4e62-9b40-16ec13591054\") " pod="glance-kuttl-tests/glance-default-single-2" Dec 01 09:03:05 crc kubenswrapper[4813]: I1201 09:03:05.329132 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/10f3f59e-3661-4248-aac9-8061818231ef-run\") pod \"glance-default-single-1\" (UID: \"10f3f59e-3661-4248-aac9-8061818231ef\") " pod="glance-kuttl-tests/glance-default-single-1" Dec 01 09:03:05 crc kubenswrapper[4813]: I1201 09:03:05.329230 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/9ea2713f-188e-4e62-9b40-16ec13591054-httpd-run\") pod \"glance-default-single-2\" (UID: \"9ea2713f-188e-4e62-9b40-16ec13591054\") " pod="glance-kuttl-tests/glance-default-single-2" Dec 01 09:03:05 crc kubenswrapper[4813]: I1201 09:03:05.329240 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/10f3f59e-3661-4248-aac9-8061818231ef-httpd-run\") pod \"glance-default-single-1\" (UID: \"10f3f59e-3661-4248-aac9-8061818231ef\") " pod="glance-kuttl-tests/glance-default-single-1" Dec 01 09:03:05 crc kubenswrapper[4813]: I1201 09:03:05.329449 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/10f3f59e-3661-4248-aac9-8061818231ef-logs\") pod \"glance-default-single-1\" (UID: \"10f3f59e-3661-4248-aac9-8061818231ef\") " pod="glance-kuttl-tests/glance-default-single-1" Dec 01 09:03:05 crc kubenswrapper[4813]: I1201 09:03:05.329448 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9ea2713f-188e-4e62-9b40-16ec13591054-logs\") pod \"glance-default-single-2\" (UID: \"9ea2713f-188e-4e62-9b40-16ec13591054\") " pod="glance-kuttl-tests/glance-default-single-2" Dec 01 09:03:05 crc kubenswrapper[4813]: I1201 09:03:05.329538 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/10f3f59e-3661-4248-aac9-8061818231ef-run\") pod \"glance-default-single-1\" (UID: \"10f3f59e-3661-4248-aac9-8061818231ef\") " pod="glance-kuttl-tests/glance-default-single-1" Dec 01 09:03:05 crc kubenswrapper[4813]: I1201 09:03:05.329645 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/9ea2713f-188e-4e62-9b40-16ec13591054-sys\") pod \"glance-default-single-2\" (UID: \"9ea2713f-188e-4e62-9b40-16ec13591054\") " pod="glance-kuttl-tests/glance-default-single-2" Dec 01 09:03:05 crc kubenswrapper[4813]: I1201 09:03:05.329707 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/9ea2713f-188e-4e62-9b40-16ec13591054-dev\") pod \"glance-default-single-2\" (UID: \"9ea2713f-188e-4e62-9b40-16ec13591054\") " pod="glance-kuttl-tests/glance-default-single-2" Dec 01 09:03:05 crc kubenswrapper[4813]: I1201 09:03:05.329729 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/9ea2713f-188e-4e62-9b40-16ec13591054-sys\") pod \"glance-default-single-2\" (UID: \"9ea2713f-188e-4e62-9b40-16ec13591054\") " pod="glance-kuttl-tests/glance-default-single-2" Dec 01 09:03:05 crc kubenswrapper[4813]: I1201 09:03:05.329741 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/10f3f59e-3661-4248-aac9-8061818231ef-var-locks-brick\") pod \"glance-default-single-1\" (UID: \"10f3f59e-3661-4248-aac9-8061818231ef\") " pod="glance-kuttl-tests/glance-default-single-1" Dec 01 09:03:05 crc kubenswrapper[4813]: I1201 09:03:05.329789 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/9ea2713f-188e-4e62-9b40-16ec13591054-lib-modules\") pod \"glance-default-single-2\" (UID: \"9ea2713f-188e-4e62-9b40-16ec13591054\") " pod="glance-kuttl-tests/glance-default-single-2" Dec 01 09:03:05 crc kubenswrapper[4813]: I1201 09:03:05.329806 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/10f3f59e-3661-4248-aac9-8061818231ef-sys\") pod \"glance-default-single-1\" (UID: \"10f3f59e-3661-4248-aac9-8061818231ef\") " pod="glance-kuttl-tests/glance-default-single-1" Dec 01 09:03:05 crc kubenswrapper[4813]: I1201 09:03:05.329845 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4r7ld\" (UniqueName: \"kubernetes.io/projected/9ea2713f-188e-4e62-9b40-16ec13591054-kube-api-access-4r7ld\") pod \"glance-default-single-2\" (UID: \"9ea2713f-188e-4e62-9b40-16ec13591054\") " pod="glance-kuttl-tests/glance-default-single-2" Dec 01 09:03:05 crc kubenswrapper[4813]: I1201 09:03:05.329863 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/10f3f59e-3661-4248-aac9-8061818231ef-etc-nvme\") pod \"glance-default-single-1\" (UID: \"10f3f59e-3661-4248-aac9-8061818231ef\") " pod="glance-kuttl-tests/glance-default-single-1" Dec 01 09:03:05 crc kubenswrapper[4813]: I1201 09:03:05.329860 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/10f3f59e-3661-4248-aac9-8061818231ef-var-locks-brick\") pod \"glance-default-single-1\" (UID: \"10f3f59e-3661-4248-aac9-8061818231ef\") " pod="glance-kuttl-tests/glance-default-single-1" Dec 01 09:03:05 crc kubenswrapper[4813]: I1201 09:03:05.329851 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/9ea2713f-188e-4e62-9b40-16ec13591054-dev\") pod \"glance-default-single-2\" (UID: \"9ea2713f-188e-4e62-9b40-16ec13591054\") " pod="glance-kuttl-tests/glance-default-single-2" Dec 01 09:03:05 crc kubenswrapper[4813]: I1201 09:03:05.329931 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/10f3f59e-3661-4248-aac9-8061818231ef-etc-nvme\") pod \"glance-default-single-1\" (UID: \"10f3f59e-3661-4248-aac9-8061818231ef\") " pod="glance-kuttl-tests/glance-default-single-1" Dec 01 09:03:05 crc kubenswrapper[4813]: I1201 09:03:05.329896 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9ea2713f-188e-4e62-9b40-16ec13591054-scripts\") pod \"glance-default-single-2\" (UID: \"9ea2713f-188e-4e62-9b40-16ec13591054\") " pod="glance-kuttl-tests/glance-default-single-2" Dec 01 09:03:05 crc kubenswrapper[4813]: I1201 09:03:05.329988 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/10f3f59e-3661-4248-aac9-8061818231ef-scripts\") pod \"glance-default-single-1\" (UID: \"10f3f59e-3661-4248-aac9-8061818231ef\") " pod="glance-kuttl-tests/glance-default-single-1" Dec 01 09:03:05 crc kubenswrapper[4813]: I1201 09:03:05.330019 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/9ea2713f-188e-4e62-9b40-16ec13591054-lib-modules\") pod \"glance-default-single-2\" (UID: \"9ea2713f-188e-4e62-9b40-16ec13591054\") " pod="glance-kuttl-tests/glance-default-single-2" Dec 01 09:03:05 crc kubenswrapper[4813]: I1201 09:03:05.330064 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/10f3f59e-3661-4248-aac9-8061818231ef-sys\") pod \"glance-default-single-1\" (UID: \"10f3f59e-3661-4248-aac9-8061818231ef\") " pod="glance-kuttl-tests/glance-default-single-1" Dec 01 09:03:05 crc kubenswrapper[4813]: I1201 09:03:05.335188 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/10f3f59e-3661-4248-aac9-8061818231ef-config-data\") pod \"glance-default-single-1\" (UID: \"10f3f59e-3661-4248-aac9-8061818231ef\") " pod="glance-kuttl-tests/glance-default-single-1" Dec 01 09:03:05 crc kubenswrapper[4813]: I1201 09:03:05.339053 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9ea2713f-188e-4e62-9b40-16ec13591054-config-data\") pod \"glance-default-single-2\" (UID: \"9ea2713f-188e-4e62-9b40-16ec13591054\") " pod="glance-kuttl-tests/glance-default-single-2" Dec 01 09:03:05 crc kubenswrapper[4813]: I1201 09:03:05.344636 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/10f3f59e-3661-4248-aac9-8061818231ef-scripts\") pod \"glance-default-single-1\" (UID: \"10f3f59e-3661-4248-aac9-8061818231ef\") " pod="glance-kuttl-tests/glance-default-single-1" Dec 01 09:03:05 crc kubenswrapper[4813]: I1201 09:03:05.348994 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9ea2713f-188e-4e62-9b40-16ec13591054-scripts\") pod \"glance-default-single-2\" (UID: \"9ea2713f-188e-4e62-9b40-16ec13591054\") " pod="glance-kuttl-tests/glance-default-single-2" Dec 01 09:03:05 crc kubenswrapper[4813]: I1201 09:03:05.355698 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5tpkj\" (UniqueName: \"kubernetes.io/projected/10f3f59e-3661-4248-aac9-8061818231ef-kube-api-access-5tpkj\") pod \"glance-default-single-1\" (UID: \"10f3f59e-3661-4248-aac9-8061818231ef\") " pod="glance-kuttl-tests/glance-default-single-1" Dec 01 09:03:05 crc kubenswrapper[4813]: I1201 09:03:05.365396 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4r7ld\" (UniqueName: \"kubernetes.io/projected/9ea2713f-188e-4e62-9b40-16ec13591054-kube-api-access-4r7ld\") pod \"glance-default-single-2\" (UID: \"9ea2713f-188e-4e62-9b40-16ec13591054\") " pod="glance-kuttl-tests/glance-default-single-2" Dec 01 09:03:05 crc kubenswrapper[4813]: I1201 09:03:05.431381 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-single-1\" (UID: \"10f3f59e-3661-4248-aac9-8061818231ef\") " pod="glance-kuttl-tests/glance-default-single-1" Dec 01 09:03:05 crc kubenswrapper[4813]: I1201 09:03:05.431454 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage20-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage20-crc\") pod \"glance-default-single-2\" (UID: \"9ea2713f-188e-4e62-9b40-16ec13591054\") " pod="glance-kuttl-tests/glance-default-single-2" Dec 01 09:03:05 crc kubenswrapper[4813]: I1201 09:03:05.431475 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage14-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage14-crc\") pod \"glance-default-single-1\" (UID: \"10f3f59e-3661-4248-aac9-8061818231ef\") " pod="glance-kuttl-tests/glance-default-single-1" Dec 01 09:03:05 crc kubenswrapper[4813]: I1201 09:03:05.431517 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-single-2\" (UID: \"9ea2713f-188e-4e62-9b40-16ec13591054\") " pod="glance-kuttl-tests/glance-default-single-2" Dec 01 09:03:05 crc kubenswrapper[4813]: I1201 09:03:05.431798 4813 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-single-1\" (UID: \"10f3f59e-3661-4248-aac9-8061818231ef\") device mount path \"/mnt/openstack/pv04\"" pod="glance-kuttl-tests/glance-default-single-1" Dec 01 09:03:05 crc kubenswrapper[4813]: I1201 09:03:05.431825 4813 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-single-2\" (UID: \"9ea2713f-188e-4e62-9b40-16ec13591054\") device mount path \"/mnt/openstack/pv03\"" pod="glance-kuttl-tests/glance-default-single-2" Dec 01 09:03:05 crc kubenswrapper[4813]: I1201 09:03:05.432016 4813 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage14-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage14-crc\") pod \"glance-default-single-1\" (UID: \"10f3f59e-3661-4248-aac9-8061818231ef\") device mount path \"/mnt/openstack/pv14\"" pod="glance-kuttl-tests/glance-default-single-1" Dec 01 09:03:05 crc kubenswrapper[4813]: I1201 09:03:05.432146 4813 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage20-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage20-crc\") pod \"glance-default-single-2\" (UID: \"9ea2713f-188e-4e62-9b40-16ec13591054\") device mount path \"/mnt/openstack/pv20\"" pod="glance-kuttl-tests/glance-default-single-2" Dec 01 09:03:05 crc kubenswrapper[4813]: I1201 09:03:05.457676 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-single-1\" (UID: \"10f3f59e-3661-4248-aac9-8061818231ef\") " pod="glance-kuttl-tests/glance-default-single-1" Dec 01 09:03:05 crc kubenswrapper[4813]: I1201 09:03:05.457700 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage14-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage14-crc\") pod \"glance-default-single-1\" (UID: \"10f3f59e-3661-4248-aac9-8061818231ef\") " pod="glance-kuttl-tests/glance-default-single-1" Dec 01 09:03:05 crc kubenswrapper[4813]: I1201 09:03:05.467384 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-single-2\" (UID: \"9ea2713f-188e-4e62-9b40-16ec13591054\") " pod="glance-kuttl-tests/glance-default-single-2" Dec 01 09:03:05 crc kubenswrapper[4813]: I1201 09:03:05.470820 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage20-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage20-crc\") pod \"glance-default-single-2\" (UID: \"9ea2713f-188e-4e62-9b40-16ec13591054\") " pod="glance-kuttl-tests/glance-default-single-2" Dec 01 09:03:05 crc kubenswrapper[4813]: I1201 09:03:05.711526 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-2" Dec 01 09:03:05 crc kubenswrapper[4813]: I1201 09:03:05.719246 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-1" Dec 01 09:03:06 crc kubenswrapper[4813]: I1201 09:03:06.194789 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-single-2"] Dec 01 09:03:06 crc kubenswrapper[4813]: W1201 09:03:06.202763 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9ea2713f_188e_4e62_9b40_16ec13591054.slice/crio-68cfe8f5f0108f59c200f92786563cd402ce9cda00f60022569e36dc510546bf WatchSource:0}: Error finding container 68cfe8f5f0108f59c200f92786563cd402ce9cda00f60022569e36dc510546bf: Status 404 returned error can't find the container with id 68cfe8f5f0108f59c200f92786563cd402ce9cda00f60022569e36dc510546bf Dec 01 09:03:06 crc kubenswrapper[4813]: I1201 09:03:06.260427 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-single-1"] Dec 01 09:03:06 crc kubenswrapper[4813]: I1201 09:03:06.696613 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-1" event={"ID":"10f3f59e-3661-4248-aac9-8061818231ef","Type":"ContainerStarted","Data":"3648cf5c6b62c15900c0c0a9b222f7cb38febe28fcb0eb01aa379760c6a970dd"} Dec 01 09:03:06 crc kubenswrapper[4813]: I1201 09:03:06.698410 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-2" event={"ID":"9ea2713f-188e-4e62-9b40-16ec13591054","Type":"ContainerStarted","Data":"68cfe8f5f0108f59c200f92786563cd402ce9cda00f60022569e36dc510546bf"} Dec 01 09:03:07 crc kubenswrapper[4813]: I1201 09:03:07.710877 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-1" event={"ID":"10f3f59e-3661-4248-aac9-8061818231ef","Type":"ContainerStarted","Data":"b20ba6eebbf841b9359d66d0d998610f2e698c070ec037b6d5be9b967bddf12d"} Dec 01 09:03:07 crc kubenswrapper[4813]: I1201 09:03:07.711407 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-1" event={"ID":"10f3f59e-3661-4248-aac9-8061818231ef","Type":"ContainerStarted","Data":"f5c0e2adef65f771285627cca9da712e378aee961ba1d0fae96bf38563a2c944"} Dec 01 09:03:07 crc kubenswrapper[4813]: I1201 09:03:07.714079 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-2" event={"ID":"9ea2713f-188e-4e62-9b40-16ec13591054","Type":"ContainerStarted","Data":"f46ce823c2b24ece3992a56d6daf0f8e40f3aa7ef37923a58876bfa8a3c1ff8f"} Dec 01 09:03:07 crc kubenswrapper[4813]: I1201 09:03:07.714134 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-2" event={"ID":"9ea2713f-188e-4e62-9b40-16ec13591054","Type":"ContainerStarted","Data":"430e108d4efc387388987e16d6db65e8a705c387bac53da344d1cda566beada2"} Dec 01 09:03:07 crc kubenswrapper[4813]: I1201 09:03:07.742086 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-default-single-1" podStartSLOduration=3.7420518830000002 podStartE2EDuration="3.742051883s" podCreationTimestamp="2025-12-01 09:03:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 09:03:07.737437673 +0000 UTC m=+1372.360359269" watchObservedRunningTime="2025-12-01 09:03:07.742051883 +0000 UTC m=+1372.364973469" Dec 01 09:03:07 crc kubenswrapper[4813]: I1201 09:03:07.770028 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-default-single-2" podStartSLOduration=3.7700041730000002 podStartE2EDuration="3.770004173s" podCreationTimestamp="2025-12-01 09:03:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 09:03:07.766840803 +0000 UTC m=+1372.389762409" watchObservedRunningTime="2025-12-01 09:03:07.770004173 +0000 UTC m=+1372.392925759" Dec 01 09:03:15 crc kubenswrapper[4813]: I1201 09:03:15.712069 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-single-2" Dec 01 09:03:15 crc kubenswrapper[4813]: I1201 09:03:15.712602 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-single-2" Dec 01 09:03:15 crc kubenswrapper[4813]: I1201 09:03:15.719454 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-single-1" Dec 01 09:03:15 crc kubenswrapper[4813]: I1201 09:03:15.719544 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-single-1" Dec 01 09:03:15 crc kubenswrapper[4813]: I1201 09:03:15.742120 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-single-2" Dec 01 09:03:15 crc kubenswrapper[4813]: I1201 09:03:15.743617 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-single-1" Dec 01 09:03:15 crc kubenswrapper[4813]: I1201 09:03:15.784441 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-single-2" Dec 01 09:03:15 crc kubenswrapper[4813]: I1201 09:03:15.784501 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-single-1" Dec 01 09:03:15 crc kubenswrapper[4813]: I1201 09:03:15.790841 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-single-1" Dec 01 09:03:15 crc kubenswrapper[4813]: I1201 09:03:15.794277 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-single-2" Dec 01 09:03:16 crc kubenswrapper[4813]: I1201 09:03:16.790937 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-single-1" Dec 01 09:03:16 crc kubenswrapper[4813]: I1201 09:03:16.791277 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-single-2" Dec 01 09:03:17 crc kubenswrapper[4813]: I1201 09:03:17.799116 4813 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 01 09:03:17 crc kubenswrapper[4813]: I1201 09:03:17.799228 4813 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 01 09:03:17 crc kubenswrapper[4813]: I1201 09:03:17.819335 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-single-2" Dec 01 09:03:17 crc kubenswrapper[4813]: I1201 09:03:17.964683 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-single-1" Dec 01 09:03:17 crc kubenswrapper[4813]: I1201 09:03:17.967283 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-single-1" Dec 01 09:03:18 crc kubenswrapper[4813]: I1201 09:03:18.723001 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-single-2" Dec 01 09:03:19 crc kubenswrapper[4813]: I1201 09:03:19.499982 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-single-2"] Dec 01 09:03:19 crc kubenswrapper[4813]: I1201 09:03:19.507223 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-single-1"] Dec 01 09:03:20 crc kubenswrapper[4813]: I1201 09:03:20.827106 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-single-2" podUID="9ea2713f-188e-4e62-9b40-16ec13591054" containerName="glance-log" containerID="cri-o://430e108d4efc387388987e16d6db65e8a705c387bac53da344d1cda566beada2" gracePeriod=30 Dec 01 09:03:20 crc kubenswrapper[4813]: I1201 09:03:20.827222 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-single-1" podUID="10f3f59e-3661-4248-aac9-8061818231ef" containerName="glance-log" containerID="cri-o://f5c0e2adef65f771285627cca9da712e378aee961ba1d0fae96bf38563a2c944" gracePeriod=30 Dec 01 09:03:20 crc kubenswrapper[4813]: I1201 09:03:20.827299 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-single-1" podUID="10f3f59e-3661-4248-aac9-8061818231ef" containerName="glance-httpd" containerID="cri-o://b20ba6eebbf841b9359d66d0d998610f2e698c070ec037b6d5be9b967bddf12d" gracePeriod=30 Dec 01 09:03:20 crc kubenswrapper[4813]: I1201 09:03:20.827362 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-single-2" podUID="9ea2713f-188e-4e62-9b40-16ec13591054" containerName="glance-httpd" containerID="cri-o://f46ce823c2b24ece3992a56d6daf0f8e40f3aa7ef37923a58876bfa8a3c1ff8f" gracePeriod=30 Dec 01 09:03:20 crc kubenswrapper[4813]: I1201 09:03:20.836430 4813 prober.go:107] "Probe failed" probeType="Readiness" pod="glance-kuttl-tests/glance-default-single-2" podUID="9ea2713f-188e-4e62-9b40-16ec13591054" containerName="glance-log" probeResult="failure" output="Get \"http://10.217.0.138:9292/healthcheck\": EOF" Dec 01 09:03:21 crc kubenswrapper[4813]: I1201 09:03:21.837424 4813 generic.go:334] "Generic (PLEG): container finished" podID="10f3f59e-3661-4248-aac9-8061818231ef" containerID="f5c0e2adef65f771285627cca9da712e378aee961ba1d0fae96bf38563a2c944" exitCode=143 Dec 01 09:03:21 crc kubenswrapper[4813]: I1201 09:03:21.837684 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-1" event={"ID":"10f3f59e-3661-4248-aac9-8061818231ef","Type":"ContainerDied","Data":"f5c0e2adef65f771285627cca9da712e378aee961ba1d0fae96bf38563a2c944"} Dec 01 09:03:21 crc kubenswrapper[4813]: I1201 09:03:21.840614 4813 generic.go:334] "Generic (PLEG): container finished" podID="9ea2713f-188e-4e62-9b40-16ec13591054" containerID="430e108d4efc387388987e16d6db65e8a705c387bac53da344d1cda566beada2" exitCode=143 Dec 01 09:03:21 crc kubenswrapper[4813]: I1201 09:03:21.840651 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-2" event={"ID":"9ea2713f-188e-4e62-9b40-16ec13591054","Type":"ContainerDied","Data":"430e108d4efc387388987e16d6db65e8a705c387bac53da344d1cda566beada2"} Dec 01 09:03:24 crc kubenswrapper[4813]: I1201 09:03:24.670019 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-2" Dec 01 09:03:24 crc kubenswrapper[4813]: I1201 09:03:24.865071 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9ea2713f-188e-4e62-9b40-16ec13591054-config-data\") pod \"9ea2713f-188e-4e62-9b40-16ec13591054\" (UID: \"9ea2713f-188e-4e62-9b40-16ec13591054\") " Dec 01 09:03:24 crc kubenswrapper[4813]: I1201 09:03:24.865135 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/9ea2713f-188e-4e62-9b40-16ec13591054-lib-modules\") pod \"9ea2713f-188e-4e62-9b40-16ec13591054\" (UID: \"9ea2713f-188e-4e62-9b40-16ec13591054\") " Dec 01 09:03:24 crc kubenswrapper[4813]: I1201 09:03:24.865157 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/9ea2713f-188e-4e62-9b40-16ec13591054-etc-nvme\") pod \"9ea2713f-188e-4e62-9b40-16ec13591054\" (UID: \"9ea2713f-188e-4e62-9b40-16ec13591054\") " Dec 01 09:03:24 crc kubenswrapper[4813]: I1201 09:03:24.865208 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/9ea2713f-188e-4e62-9b40-16ec13591054-dev\") pod \"9ea2713f-188e-4e62-9b40-16ec13591054\" (UID: \"9ea2713f-188e-4e62-9b40-16ec13591054\") " Dec 01 09:03:24 crc kubenswrapper[4813]: I1201 09:03:24.865241 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/9ea2713f-188e-4e62-9b40-16ec13591054-httpd-run\") pod \"9ea2713f-188e-4e62-9b40-16ec13591054\" (UID: \"9ea2713f-188e-4e62-9b40-16ec13591054\") " Dec 01 09:03:24 crc kubenswrapper[4813]: I1201 09:03:24.865280 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance-cache\" (UniqueName: \"kubernetes.io/local-volume/local-storage20-crc\") pod \"9ea2713f-188e-4e62-9b40-16ec13591054\" (UID: \"9ea2713f-188e-4e62-9b40-16ec13591054\") " Dec 01 09:03:24 crc kubenswrapper[4813]: I1201 09:03:24.865274 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9ea2713f-188e-4e62-9b40-16ec13591054-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "9ea2713f-188e-4e62-9b40-16ec13591054" (UID: "9ea2713f-188e-4e62-9b40-16ec13591054"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 09:03:24 crc kubenswrapper[4813]: I1201 09:03:24.865310 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9ea2713f-188e-4e62-9b40-16ec13591054-logs\") pod \"9ea2713f-188e-4e62-9b40-16ec13591054\" (UID: \"9ea2713f-188e-4e62-9b40-16ec13591054\") " Dec 01 09:03:24 crc kubenswrapper[4813]: I1201 09:03:24.865274 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9ea2713f-188e-4e62-9b40-16ec13591054-etc-nvme" (OuterVolumeSpecName: "etc-nvme") pod "9ea2713f-188e-4e62-9b40-16ec13591054" (UID: "9ea2713f-188e-4e62-9b40-16ec13591054"). InnerVolumeSpecName "etc-nvme". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 09:03:24 crc kubenswrapper[4813]: I1201 09:03:24.865331 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/9ea2713f-188e-4e62-9b40-16ec13591054-run\") pod \"9ea2713f-188e-4e62-9b40-16ec13591054\" (UID: \"9ea2713f-188e-4e62-9b40-16ec13591054\") " Dec 01 09:03:24 crc kubenswrapper[4813]: I1201 09:03:24.865376 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9ea2713f-188e-4e62-9b40-16ec13591054-run" (OuterVolumeSpecName: "run") pod "9ea2713f-188e-4e62-9b40-16ec13591054" (UID: "9ea2713f-188e-4e62-9b40-16ec13591054"). InnerVolumeSpecName "run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 09:03:24 crc kubenswrapper[4813]: I1201 09:03:24.865406 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9ea2713f-188e-4e62-9b40-16ec13591054-scripts\") pod \"9ea2713f-188e-4e62-9b40-16ec13591054\" (UID: \"9ea2713f-188e-4e62-9b40-16ec13591054\") " Dec 01 09:03:24 crc kubenswrapper[4813]: I1201 09:03:24.865456 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/9ea2713f-188e-4e62-9b40-16ec13591054-etc-iscsi\") pod \"9ea2713f-188e-4e62-9b40-16ec13591054\" (UID: \"9ea2713f-188e-4e62-9b40-16ec13591054\") " Dec 01 09:03:24 crc kubenswrapper[4813]: I1201 09:03:24.865523 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/9ea2713f-188e-4e62-9b40-16ec13591054-var-locks-brick\") pod \"9ea2713f-188e-4e62-9b40-16ec13591054\" (UID: \"9ea2713f-188e-4e62-9b40-16ec13591054\") " Dec 01 09:03:24 crc kubenswrapper[4813]: I1201 09:03:24.865552 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"9ea2713f-188e-4e62-9b40-16ec13591054\" (UID: \"9ea2713f-188e-4e62-9b40-16ec13591054\") " Dec 01 09:03:24 crc kubenswrapper[4813]: I1201 09:03:24.865577 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/9ea2713f-188e-4e62-9b40-16ec13591054-sys\") pod \"9ea2713f-188e-4e62-9b40-16ec13591054\" (UID: \"9ea2713f-188e-4e62-9b40-16ec13591054\") " Dec 01 09:03:24 crc kubenswrapper[4813]: I1201 09:03:24.865584 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9ea2713f-188e-4e62-9b40-16ec13591054-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "9ea2713f-188e-4e62-9b40-16ec13591054" (UID: "9ea2713f-188e-4e62-9b40-16ec13591054"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 09:03:24 crc kubenswrapper[4813]: I1201 09:03:24.865657 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4r7ld\" (UniqueName: \"kubernetes.io/projected/9ea2713f-188e-4e62-9b40-16ec13591054-kube-api-access-4r7ld\") pod \"9ea2713f-188e-4e62-9b40-16ec13591054\" (UID: \"9ea2713f-188e-4e62-9b40-16ec13591054\") " Dec 01 09:03:24 crc kubenswrapper[4813]: I1201 09:03:24.865660 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9ea2713f-188e-4e62-9b40-16ec13591054-dev" (OuterVolumeSpecName: "dev") pod "9ea2713f-188e-4e62-9b40-16ec13591054" (UID: "9ea2713f-188e-4e62-9b40-16ec13591054"). InnerVolumeSpecName "dev". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 09:03:24 crc kubenswrapper[4813]: I1201 09:03:24.865783 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9ea2713f-188e-4e62-9b40-16ec13591054-var-locks-brick" (OuterVolumeSpecName: "var-locks-brick") pod "9ea2713f-188e-4e62-9b40-16ec13591054" (UID: "9ea2713f-188e-4e62-9b40-16ec13591054"). InnerVolumeSpecName "var-locks-brick". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 09:03:24 crc kubenswrapper[4813]: I1201 09:03:24.865832 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9ea2713f-188e-4e62-9b40-16ec13591054-etc-iscsi" (OuterVolumeSpecName: "etc-iscsi") pod "9ea2713f-188e-4e62-9b40-16ec13591054" (UID: "9ea2713f-188e-4e62-9b40-16ec13591054"). InnerVolumeSpecName "etc-iscsi". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 09:03:24 crc kubenswrapper[4813]: I1201 09:03:24.865924 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9ea2713f-188e-4e62-9b40-16ec13591054-sys" (OuterVolumeSpecName: "sys") pod "9ea2713f-188e-4e62-9b40-16ec13591054" (UID: "9ea2713f-188e-4e62-9b40-16ec13591054"). InnerVolumeSpecName "sys". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 09:03:24 crc kubenswrapper[4813]: I1201 09:03:24.866031 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9ea2713f-188e-4e62-9b40-16ec13591054-logs" (OuterVolumeSpecName: "logs") pod "9ea2713f-188e-4e62-9b40-16ec13591054" (UID: "9ea2713f-188e-4e62-9b40-16ec13591054"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 09:03:24 crc kubenswrapper[4813]: I1201 09:03:24.866948 4813 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9ea2713f-188e-4e62-9b40-16ec13591054-logs\") on node \"crc\" DevicePath \"\"" Dec 01 09:03:24 crc kubenswrapper[4813]: I1201 09:03:24.866985 4813 reconciler_common.go:293] "Volume detached for volume \"run\" (UniqueName: \"kubernetes.io/host-path/9ea2713f-188e-4e62-9b40-16ec13591054-run\") on node \"crc\" DevicePath \"\"" Dec 01 09:03:24 crc kubenswrapper[4813]: I1201 09:03:24.867014 4813 reconciler_common.go:293] "Volume detached for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/9ea2713f-188e-4e62-9b40-16ec13591054-etc-iscsi\") on node \"crc\" DevicePath \"\"" Dec 01 09:03:24 crc kubenswrapper[4813]: I1201 09:03:24.867024 4813 reconciler_common.go:293] "Volume detached for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/9ea2713f-188e-4e62-9b40-16ec13591054-var-locks-brick\") on node \"crc\" DevicePath \"\"" Dec 01 09:03:24 crc kubenswrapper[4813]: I1201 09:03:24.867038 4813 reconciler_common.go:293] "Volume detached for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/9ea2713f-188e-4e62-9b40-16ec13591054-sys\") on node \"crc\" DevicePath \"\"" Dec 01 09:03:24 crc kubenswrapper[4813]: I1201 09:03:24.867048 4813 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/9ea2713f-188e-4e62-9b40-16ec13591054-lib-modules\") on node \"crc\" DevicePath \"\"" Dec 01 09:03:24 crc kubenswrapper[4813]: I1201 09:03:24.867060 4813 reconciler_common.go:293] "Volume detached for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/9ea2713f-188e-4e62-9b40-16ec13591054-etc-nvme\") on node \"crc\" DevicePath \"\"" Dec 01 09:03:24 crc kubenswrapper[4813]: I1201 09:03:24.867070 4813 reconciler_common.go:293] "Volume detached for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/9ea2713f-188e-4e62-9b40-16ec13591054-dev\") on node \"crc\" DevicePath \"\"" Dec 01 09:03:24 crc kubenswrapper[4813]: I1201 09:03:24.867079 4813 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/9ea2713f-188e-4e62-9b40-16ec13591054-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 01 09:03:24 crc kubenswrapper[4813]: I1201 09:03:24.868856 4813 generic.go:334] "Generic (PLEG): container finished" podID="9ea2713f-188e-4e62-9b40-16ec13591054" containerID="f46ce823c2b24ece3992a56d6daf0f8e40f3aa7ef37923a58876bfa8a3c1ff8f" exitCode=0 Dec 01 09:03:24 crc kubenswrapper[4813]: I1201 09:03:24.868931 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-2" event={"ID":"9ea2713f-188e-4e62-9b40-16ec13591054","Type":"ContainerDied","Data":"f46ce823c2b24ece3992a56d6daf0f8e40f3aa7ef37923a58876bfa8a3c1ff8f"} Dec 01 09:03:24 crc kubenswrapper[4813]: I1201 09:03:24.868948 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-2" Dec 01 09:03:24 crc kubenswrapper[4813]: I1201 09:03:24.868976 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-2" event={"ID":"9ea2713f-188e-4e62-9b40-16ec13591054","Type":"ContainerDied","Data":"68cfe8f5f0108f59c200f92786563cd402ce9cda00f60022569e36dc510546bf"} Dec 01 09:03:24 crc kubenswrapper[4813]: I1201 09:03:24.869032 4813 scope.go:117] "RemoveContainer" containerID="f46ce823c2b24ece3992a56d6daf0f8e40f3aa7ef37923a58876bfa8a3c1ff8f" Dec 01 09:03:24 crc kubenswrapper[4813]: I1201 09:03:24.871940 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9ea2713f-188e-4e62-9b40-16ec13591054-scripts" (OuterVolumeSpecName: "scripts") pod "9ea2713f-188e-4e62-9b40-16ec13591054" (UID: "9ea2713f-188e-4e62-9b40-16ec13591054"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:03:24 crc kubenswrapper[4813]: I1201 09:03:24.872513 4813 generic.go:334] "Generic (PLEG): container finished" podID="10f3f59e-3661-4248-aac9-8061818231ef" containerID="b20ba6eebbf841b9359d66d0d998610f2e698c070ec037b6d5be9b967bddf12d" exitCode=0 Dec 01 09:03:24 crc kubenswrapper[4813]: I1201 09:03:24.872563 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-1" event={"ID":"10f3f59e-3661-4248-aac9-8061818231ef","Type":"ContainerDied","Data":"b20ba6eebbf841b9359d66d0d998610f2e698c070ec037b6d5be9b967bddf12d"} Dec 01 09:03:24 crc kubenswrapper[4813]: I1201 09:03:24.872764 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage03-crc" (OuterVolumeSpecName: "glance") pod "9ea2713f-188e-4e62-9b40-16ec13591054" (UID: "9ea2713f-188e-4e62-9b40-16ec13591054"). InnerVolumeSpecName "local-storage03-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 01 09:03:24 crc kubenswrapper[4813]: I1201 09:03:24.874412 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9ea2713f-188e-4e62-9b40-16ec13591054-kube-api-access-4r7ld" (OuterVolumeSpecName: "kube-api-access-4r7ld") pod "9ea2713f-188e-4e62-9b40-16ec13591054" (UID: "9ea2713f-188e-4e62-9b40-16ec13591054"). InnerVolumeSpecName "kube-api-access-4r7ld". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:03:24 crc kubenswrapper[4813]: I1201 09:03:24.875383 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage20-crc" (OuterVolumeSpecName: "glance-cache") pod "9ea2713f-188e-4e62-9b40-16ec13591054" (UID: "9ea2713f-188e-4e62-9b40-16ec13591054"). InnerVolumeSpecName "local-storage20-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 01 09:03:24 crc kubenswrapper[4813]: I1201 09:03:24.908308 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9ea2713f-188e-4e62-9b40-16ec13591054-config-data" (OuterVolumeSpecName: "config-data") pod "9ea2713f-188e-4e62-9b40-16ec13591054" (UID: "9ea2713f-188e-4e62-9b40-16ec13591054"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:03:24 crc kubenswrapper[4813]: I1201 09:03:24.916377 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-1" Dec 01 09:03:24 crc kubenswrapper[4813]: I1201 09:03:24.920179 4813 scope.go:117] "RemoveContainer" containerID="430e108d4efc387388987e16d6db65e8a705c387bac53da344d1cda566beada2" Dec 01 09:03:24 crc kubenswrapper[4813]: I1201 09:03:24.938342 4813 scope.go:117] "RemoveContainer" containerID="f46ce823c2b24ece3992a56d6daf0f8e40f3aa7ef37923a58876bfa8a3c1ff8f" Dec 01 09:03:24 crc kubenswrapper[4813]: E1201 09:03:24.941954 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f46ce823c2b24ece3992a56d6daf0f8e40f3aa7ef37923a58876bfa8a3c1ff8f\": container with ID starting with f46ce823c2b24ece3992a56d6daf0f8e40f3aa7ef37923a58876bfa8a3c1ff8f not found: ID does not exist" containerID="f46ce823c2b24ece3992a56d6daf0f8e40f3aa7ef37923a58876bfa8a3c1ff8f" Dec 01 09:03:24 crc kubenswrapper[4813]: I1201 09:03:24.942034 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f46ce823c2b24ece3992a56d6daf0f8e40f3aa7ef37923a58876bfa8a3c1ff8f"} err="failed to get container status \"f46ce823c2b24ece3992a56d6daf0f8e40f3aa7ef37923a58876bfa8a3c1ff8f\": rpc error: code = NotFound desc = could not find container \"f46ce823c2b24ece3992a56d6daf0f8e40f3aa7ef37923a58876bfa8a3c1ff8f\": container with ID starting with f46ce823c2b24ece3992a56d6daf0f8e40f3aa7ef37923a58876bfa8a3c1ff8f not found: ID does not exist" Dec 01 09:03:24 crc kubenswrapper[4813]: I1201 09:03:24.942069 4813 scope.go:117] "RemoveContainer" containerID="430e108d4efc387388987e16d6db65e8a705c387bac53da344d1cda566beada2" Dec 01 09:03:24 crc kubenswrapper[4813]: E1201 09:03:24.942645 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"430e108d4efc387388987e16d6db65e8a705c387bac53da344d1cda566beada2\": container with ID starting with 430e108d4efc387388987e16d6db65e8a705c387bac53da344d1cda566beada2 not found: ID does not exist" containerID="430e108d4efc387388987e16d6db65e8a705c387bac53da344d1cda566beada2" Dec 01 09:03:24 crc kubenswrapper[4813]: I1201 09:03:24.942678 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"430e108d4efc387388987e16d6db65e8a705c387bac53da344d1cda566beada2"} err="failed to get container status \"430e108d4efc387388987e16d6db65e8a705c387bac53da344d1cda566beada2\": rpc error: code = NotFound desc = could not find container \"430e108d4efc387388987e16d6db65e8a705c387bac53da344d1cda566beada2\": container with ID starting with 430e108d4efc387388987e16d6db65e8a705c387bac53da344d1cda566beada2 not found: ID does not exist" Dec 01 09:03:24 crc kubenswrapper[4813]: I1201 09:03:24.985098 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/10f3f59e-3661-4248-aac9-8061818231ef-etc-nvme\") pod \"10f3f59e-3661-4248-aac9-8061818231ef\" (UID: \"10f3f59e-3661-4248-aac9-8061818231ef\") " Dec 01 09:03:24 crc kubenswrapper[4813]: I1201 09:03:24.985179 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage14-crc\") pod \"10f3f59e-3661-4248-aac9-8061818231ef\" (UID: \"10f3f59e-3661-4248-aac9-8061818231ef\") " Dec 01 09:03:24 crc kubenswrapper[4813]: I1201 09:03:24.985240 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/10f3f59e-3661-4248-aac9-8061818231ef-sys\") pod \"10f3f59e-3661-4248-aac9-8061818231ef\" (UID: \"10f3f59e-3661-4248-aac9-8061818231ef\") " Dec 01 09:03:24 crc kubenswrapper[4813]: I1201 09:03:24.985339 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/10f3f59e-3661-4248-aac9-8061818231ef-dev\") pod \"10f3f59e-3661-4248-aac9-8061818231ef\" (UID: \"10f3f59e-3661-4248-aac9-8061818231ef\") " Dec 01 09:03:24 crc kubenswrapper[4813]: I1201 09:03:24.985416 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/10f3f59e-3661-4248-aac9-8061818231ef-logs\") pod \"10f3f59e-3661-4248-aac9-8061818231ef\" (UID: \"10f3f59e-3661-4248-aac9-8061818231ef\") " Dec 01 09:03:24 crc kubenswrapper[4813]: I1201 09:03:24.985436 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/10f3f59e-3661-4248-aac9-8061818231ef-var-locks-brick\") pod \"10f3f59e-3661-4248-aac9-8061818231ef\" (UID: \"10f3f59e-3661-4248-aac9-8061818231ef\") " Dec 01 09:03:24 crc kubenswrapper[4813]: I1201 09:03:24.985450 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/10f3f59e-3661-4248-aac9-8061818231ef-etc-iscsi\") pod \"10f3f59e-3661-4248-aac9-8061818231ef\" (UID: \"10f3f59e-3661-4248-aac9-8061818231ef\") " Dec 01 09:03:24 crc kubenswrapper[4813]: I1201 09:03:24.985479 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/10f3f59e-3661-4248-aac9-8061818231ef-httpd-run\") pod \"10f3f59e-3661-4248-aac9-8061818231ef\" (UID: \"10f3f59e-3661-4248-aac9-8061818231ef\") " Dec 01 09:03:24 crc kubenswrapper[4813]: I1201 09:03:24.985508 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5tpkj\" (UniqueName: \"kubernetes.io/projected/10f3f59e-3661-4248-aac9-8061818231ef-kube-api-access-5tpkj\") pod \"10f3f59e-3661-4248-aac9-8061818231ef\" (UID: \"10f3f59e-3661-4248-aac9-8061818231ef\") " Dec 01 09:03:24 crc kubenswrapper[4813]: I1201 09:03:24.985543 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/10f3f59e-3661-4248-aac9-8061818231ef-dev" (OuterVolumeSpecName: "dev") pod "10f3f59e-3661-4248-aac9-8061818231ef" (UID: "10f3f59e-3661-4248-aac9-8061818231ef"). InnerVolumeSpecName "dev". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 09:03:24 crc kubenswrapper[4813]: I1201 09:03:24.985560 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/10f3f59e-3661-4248-aac9-8061818231ef-scripts\") pod \"10f3f59e-3661-4248-aac9-8061818231ef\" (UID: \"10f3f59e-3661-4248-aac9-8061818231ef\") " Dec 01 09:03:24 crc kubenswrapper[4813]: I1201 09:03:24.985620 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/10f3f59e-3661-4248-aac9-8061818231ef-config-data\") pod \"10f3f59e-3661-4248-aac9-8061818231ef\" (UID: \"10f3f59e-3661-4248-aac9-8061818231ef\") " Dec 01 09:03:24 crc kubenswrapper[4813]: I1201 09:03:24.985640 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance-cache\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"10f3f59e-3661-4248-aac9-8061818231ef\" (UID: \"10f3f59e-3661-4248-aac9-8061818231ef\") " Dec 01 09:03:24 crc kubenswrapper[4813]: I1201 09:03:24.985659 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/10f3f59e-3661-4248-aac9-8061818231ef-lib-modules\") pod \"10f3f59e-3661-4248-aac9-8061818231ef\" (UID: \"10f3f59e-3661-4248-aac9-8061818231ef\") " Dec 01 09:03:24 crc kubenswrapper[4813]: I1201 09:03:24.985678 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/10f3f59e-3661-4248-aac9-8061818231ef-run\") pod \"10f3f59e-3661-4248-aac9-8061818231ef\" (UID: \"10f3f59e-3661-4248-aac9-8061818231ef\") " Dec 01 09:03:24 crc kubenswrapper[4813]: I1201 09:03:24.985870 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/10f3f59e-3661-4248-aac9-8061818231ef-sys" (OuterVolumeSpecName: "sys") pod "10f3f59e-3661-4248-aac9-8061818231ef" (UID: "10f3f59e-3661-4248-aac9-8061818231ef"). InnerVolumeSpecName "sys". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 09:03:24 crc kubenswrapper[4813]: I1201 09:03:24.985912 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/10f3f59e-3661-4248-aac9-8061818231ef-run" (OuterVolumeSpecName: "run") pod "10f3f59e-3661-4248-aac9-8061818231ef" (UID: "10f3f59e-3661-4248-aac9-8061818231ef"). InnerVolumeSpecName "run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 09:03:24 crc kubenswrapper[4813]: I1201 09:03:24.986041 4813 reconciler_common.go:293] "Volume detached for volume \"run\" (UniqueName: \"kubernetes.io/host-path/10f3f59e-3661-4248-aac9-8061818231ef-run\") on node \"crc\" DevicePath \"\"" Dec 01 09:03:24 crc kubenswrapper[4813]: I1201 09:03:24.986069 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4r7ld\" (UniqueName: \"kubernetes.io/projected/9ea2713f-188e-4e62-9b40-16ec13591054-kube-api-access-4r7ld\") on node \"crc\" DevicePath \"\"" Dec 01 09:03:24 crc kubenswrapper[4813]: I1201 09:03:24.986082 4813 reconciler_common.go:293] "Volume detached for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/10f3f59e-3661-4248-aac9-8061818231ef-sys\") on node \"crc\" DevicePath \"\"" Dec 01 09:03:24 crc kubenswrapper[4813]: I1201 09:03:24.986094 4813 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9ea2713f-188e-4e62-9b40-16ec13591054-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 09:03:24 crc kubenswrapper[4813]: I1201 09:03:24.986107 4813 reconciler_common.go:293] "Volume detached for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/10f3f59e-3661-4248-aac9-8061818231ef-dev\") on node \"crc\" DevicePath \"\"" Dec 01 09:03:24 crc kubenswrapper[4813]: I1201 09:03:24.986153 4813 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage20-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage20-crc\") on node \"crc\" " Dec 01 09:03:24 crc kubenswrapper[4813]: I1201 09:03:24.986166 4813 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9ea2713f-188e-4e62-9b40-16ec13591054-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 09:03:24 crc kubenswrapper[4813]: I1201 09:03:24.986182 4813 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" " Dec 01 09:03:24 crc kubenswrapper[4813]: I1201 09:03:24.987341 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/10f3f59e-3661-4248-aac9-8061818231ef-etc-iscsi" (OuterVolumeSpecName: "etc-iscsi") pod "10f3f59e-3661-4248-aac9-8061818231ef" (UID: "10f3f59e-3661-4248-aac9-8061818231ef"). InnerVolumeSpecName "etc-iscsi". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 09:03:24 crc kubenswrapper[4813]: I1201 09:03:24.994789 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/10f3f59e-3661-4248-aac9-8061818231ef-var-locks-brick" (OuterVolumeSpecName: "var-locks-brick") pod "10f3f59e-3661-4248-aac9-8061818231ef" (UID: "10f3f59e-3661-4248-aac9-8061818231ef"). InnerVolumeSpecName "var-locks-brick". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 09:03:24 crc kubenswrapper[4813]: I1201 09:03:24.995189 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/10f3f59e-3661-4248-aac9-8061818231ef-logs" (OuterVolumeSpecName: "logs") pod "10f3f59e-3661-4248-aac9-8061818231ef" (UID: "10f3f59e-3661-4248-aac9-8061818231ef"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 09:03:24 crc kubenswrapper[4813]: I1201 09:03:24.995722 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/10f3f59e-3661-4248-aac9-8061818231ef-scripts" (OuterVolumeSpecName: "scripts") pod "10f3f59e-3661-4248-aac9-8061818231ef" (UID: "10f3f59e-3661-4248-aac9-8061818231ef"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:03:24 crc kubenswrapper[4813]: I1201 09:03:24.999177 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/10f3f59e-3661-4248-aac9-8061818231ef-etc-nvme" (OuterVolumeSpecName: "etc-nvme") pod "10f3f59e-3661-4248-aac9-8061818231ef" (UID: "10f3f59e-3661-4248-aac9-8061818231ef"). InnerVolumeSpecName "etc-nvme". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 09:03:24 crc kubenswrapper[4813]: I1201 09:03:24.999297 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/10f3f59e-3661-4248-aac9-8061818231ef-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "10f3f59e-3661-4248-aac9-8061818231ef" (UID: "10f3f59e-3661-4248-aac9-8061818231ef"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 09:03:24 crc kubenswrapper[4813]: I1201 09:03:24.999571 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/10f3f59e-3661-4248-aac9-8061818231ef-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "10f3f59e-3661-4248-aac9-8061818231ef" (UID: "10f3f59e-3661-4248-aac9-8061818231ef"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 09:03:25 crc kubenswrapper[4813]: I1201 09:03:25.000322 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage04-crc" (OuterVolumeSpecName: "glance-cache") pod "10f3f59e-3661-4248-aac9-8061818231ef" (UID: "10f3f59e-3661-4248-aac9-8061818231ef"). InnerVolumeSpecName "local-storage04-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 01 09:03:25 crc kubenswrapper[4813]: I1201 09:03:25.001520 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage14-crc" (OuterVolumeSpecName: "glance") pod "10f3f59e-3661-4248-aac9-8061818231ef" (UID: "10f3f59e-3661-4248-aac9-8061818231ef"). InnerVolumeSpecName "local-storage14-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 01 09:03:25 crc kubenswrapper[4813]: I1201 09:03:25.019445 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/10f3f59e-3661-4248-aac9-8061818231ef-kube-api-access-5tpkj" (OuterVolumeSpecName: "kube-api-access-5tpkj") pod "10f3f59e-3661-4248-aac9-8061818231ef" (UID: "10f3f59e-3661-4248-aac9-8061818231ef"). InnerVolumeSpecName "kube-api-access-5tpkj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:03:25 crc kubenswrapper[4813]: I1201 09:03:25.023909 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/10f3f59e-3661-4248-aac9-8061818231ef-config-data" (OuterVolumeSpecName: "config-data") pod "10f3f59e-3661-4248-aac9-8061818231ef" (UID: "10f3f59e-3661-4248-aac9-8061818231ef"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:03:25 crc kubenswrapper[4813]: I1201 09:03:25.024822 4813 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage03-crc" (UniqueName: "kubernetes.io/local-volume/local-storage03-crc") on node "crc" Dec 01 09:03:25 crc kubenswrapper[4813]: I1201 09:03:25.031474 4813 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage20-crc" (UniqueName: "kubernetes.io/local-volume/local-storage20-crc") on node "crc" Dec 01 09:03:25 crc kubenswrapper[4813]: I1201 09:03:25.087621 4813 reconciler_common.go:293] "Volume detached for volume \"local-storage20-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage20-crc\") on node \"crc\" DevicePath \"\"" Dec 01 09:03:25 crc kubenswrapper[4813]: I1201 09:03:25.087652 4813 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/10f3f59e-3661-4248-aac9-8061818231ef-logs\") on node \"crc\" DevicePath \"\"" Dec 01 09:03:25 crc kubenswrapper[4813]: I1201 09:03:25.087665 4813 reconciler_common.go:293] "Volume detached for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/10f3f59e-3661-4248-aac9-8061818231ef-var-locks-brick\") on node \"crc\" DevicePath \"\"" Dec 01 09:03:25 crc kubenswrapper[4813]: I1201 09:03:25.087675 4813 reconciler_common.go:293] "Volume detached for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/10f3f59e-3661-4248-aac9-8061818231ef-etc-iscsi\") on node \"crc\" DevicePath \"\"" Dec 01 09:03:25 crc kubenswrapper[4813]: I1201 09:03:25.087685 4813 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/10f3f59e-3661-4248-aac9-8061818231ef-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 01 09:03:25 crc kubenswrapper[4813]: I1201 09:03:25.087694 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5tpkj\" (UniqueName: \"kubernetes.io/projected/10f3f59e-3661-4248-aac9-8061818231ef-kube-api-access-5tpkj\") on node \"crc\" DevicePath \"\"" Dec 01 09:03:25 crc kubenswrapper[4813]: I1201 09:03:25.087703 4813 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/10f3f59e-3661-4248-aac9-8061818231ef-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 09:03:25 crc kubenswrapper[4813]: I1201 09:03:25.087711 4813 reconciler_common.go:293] "Volume detached for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" DevicePath \"\"" Dec 01 09:03:25 crc kubenswrapper[4813]: I1201 09:03:25.087719 4813 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/10f3f59e-3661-4248-aac9-8061818231ef-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 09:03:25 crc kubenswrapper[4813]: I1201 09:03:25.087753 4813 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" " Dec 01 09:03:25 crc kubenswrapper[4813]: I1201 09:03:25.087762 4813 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/10f3f59e-3661-4248-aac9-8061818231ef-lib-modules\") on node \"crc\" DevicePath \"\"" Dec 01 09:03:25 crc kubenswrapper[4813]: I1201 09:03:25.087770 4813 reconciler_common.go:293] "Volume detached for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/10f3f59e-3661-4248-aac9-8061818231ef-etc-nvme\") on node \"crc\" DevicePath \"\"" Dec 01 09:03:25 crc kubenswrapper[4813]: I1201 09:03:25.087783 4813 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage14-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage14-crc\") on node \"crc\" " Dec 01 09:03:25 crc kubenswrapper[4813]: I1201 09:03:25.106342 4813 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage14-crc" (UniqueName: "kubernetes.io/local-volume/local-storage14-crc") on node "crc" Dec 01 09:03:25 crc kubenswrapper[4813]: I1201 09:03:25.113216 4813 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage04-crc" (UniqueName: "kubernetes.io/local-volume/local-storage04-crc") on node "crc" Dec 01 09:03:25 crc kubenswrapper[4813]: I1201 09:03:25.211831 4813 reconciler_common.go:293] "Volume detached for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" DevicePath \"\"" Dec 01 09:03:25 crc kubenswrapper[4813]: I1201 09:03:25.211860 4813 reconciler_common.go:293] "Volume detached for volume \"local-storage14-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage14-crc\") on node \"crc\" DevicePath \"\"" Dec 01 09:03:25 crc kubenswrapper[4813]: I1201 09:03:25.247681 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-single-2"] Dec 01 09:03:25 crc kubenswrapper[4813]: I1201 09:03:25.259622 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-default-single-2"] Dec 01 09:03:25 crc kubenswrapper[4813]: I1201 09:03:25.885315 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-1" event={"ID":"10f3f59e-3661-4248-aac9-8061818231ef","Type":"ContainerDied","Data":"3648cf5c6b62c15900c0c0a9b222f7cb38febe28fcb0eb01aa379760c6a970dd"} Dec 01 09:03:25 crc kubenswrapper[4813]: I1201 09:03:25.885413 4813 scope.go:117] "RemoveContainer" containerID="b20ba6eebbf841b9359d66d0d998610f2e698c070ec037b6d5be9b967bddf12d" Dec 01 09:03:25 crc kubenswrapper[4813]: I1201 09:03:25.885400 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-1" Dec 01 09:03:25 crc kubenswrapper[4813]: I1201 09:03:25.915943 4813 scope.go:117] "RemoveContainer" containerID="f5c0e2adef65f771285627cca9da712e378aee961ba1d0fae96bf38563a2c944" Dec 01 09:03:25 crc kubenswrapper[4813]: I1201 09:03:25.921111 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-single-1"] Dec 01 09:03:25 crc kubenswrapper[4813]: I1201 09:03:25.930522 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-default-single-1"] Dec 01 09:03:26 crc kubenswrapper[4813]: I1201 09:03:26.410351 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="10f3f59e-3661-4248-aac9-8061818231ef" path="/var/lib/kubelet/pods/10f3f59e-3661-4248-aac9-8061818231ef/volumes" Dec 01 09:03:26 crc kubenswrapper[4813]: I1201 09:03:26.411743 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9ea2713f-188e-4e62-9b40-16ec13591054" path="/var/lib/kubelet/pods/9ea2713f-188e-4e62-9b40-16ec13591054/volumes" Dec 01 09:03:26 crc kubenswrapper[4813]: I1201 09:03:26.819425 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Dec 01 09:03:26 crc kubenswrapper[4813]: I1201 09:03:26.819738 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-single-0" podUID="9a0e144a-6298-4092-86fc-ba45b986d6ec" containerName="glance-log" containerID="cri-o://db306b76d1e4d93ad74ca563b646090f0237e7fdc176fc18b82ff6a6cc3e5413" gracePeriod=30 Dec 01 09:03:26 crc kubenswrapper[4813]: I1201 09:03:26.819808 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-single-0" podUID="9a0e144a-6298-4092-86fc-ba45b986d6ec" containerName="glance-httpd" containerID="cri-o://c0918867756a6c3013f771a8e9cbf8e78fd40a1e5f1cc6fa2576633699e9e100" gracePeriod=30 Dec 01 09:03:27 crc kubenswrapper[4813]: I1201 09:03:27.916160 4813 generic.go:334] "Generic (PLEG): container finished" podID="9a0e144a-6298-4092-86fc-ba45b986d6ec" containerID="db306b76d1e4d93ad74ca563b646090f0237e7fdc176fc18b82ff6a6cc3e5413" exitCode=143 Dec 01 09:03:27 crc kubenswrapper[4813]: I1201 09:03:27.916266 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"9a0e144a-6298-4092-86fc-ba45b986d6ec","Type":"ContainerDied","Data":"db306b76d1e4d93ad74ca563b646090f0237e7fdc176fc18b82ff6a6cc3e5413"} Dec 01 09:03:30 crc kubenswrapper[4813]: I1201 09:03:30.741572 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Dec 01 09:03:30 crc kubenswrapper[4813]: I1201 09:03:30.902525 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9a0e144a-6298-4092-86fc-ba45b986d6ec-config-data\") pod \"9a0e144a-6298-4092-86fc-ba45b986d6ec\" (UID: \"9a0e144a-6298-4092-86fc-ba45b986d6ec\") " Dec 01 09:03:30 crc kubenswrapper[4813]: I1201 09:03:30.902793 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance-cache\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"9a0e144a-6298-4092-86fc-ba45b986d6ec\" (UID: \"9a0e144a-6298-4092-86fc-ba45b986d6ec\") " Dec 01 09:03:30 crc kubenswrapper[4813]: I1201 09:03:30.902816 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/9a0e144a-6298-4092-86fc-ba45b986d6ec-run\") pod \"9a0e144a-6298-4092-86fc-ba45b986d6ec\" (UID: \"9a0e144a-6298-4092-86fc-ba45b986d6ec\") " Dec 01 09:03:30 crc kubenswrapper[4813]: I1201 09:03:30.902842 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/9a0e144a-6298-4092-86fc-ba45b986d6ec-dev\") pod \"9a0e144a-6298-4092-86fc-ba45b986d6ec\" (UID: \"9a0e144a-6298-4092-86fc-ba45b986d6ec\") " Dec 01 09:03:30 crc kubenswrapper[4813]: I1201 09:03:30.902883 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9a0e144a-6298-4092-86fc-ba45b986d6ec-scripts\") pod \"9a0e144a-6298-4092-86fc-ba45b986d6ec\" (UID: \"9a0e144a-6298-4092-86fc-ba45b986d6ec\") " Dec 01 09:03:30 crc kubenswrapper[4813]: I1201 09:03:30.902903 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/9a0e144a-6298-4092-86fc-ba45b986d6ec-var-locks-brick\") pod \"9a0e144a-6298-4092-86fc-ba45b986d6ec\" (UID: \"9a0e144a-6298-4092-86fc-ba45b986d6ec\") " Dec 01 09:03:30 crc kubenswrapper[4813]: I1201 09:03:30.902947 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/9a0e144a-6298-4092-86fc-ba45b986d6ec-etc-nvme\") pod \"9a0e144a-6298-4092-86fc-ba45b986d6ec\" (UID: \"9a0e144a-6298-4092-86fc-ba45b986d6ec\") " Dec 01 09:03:30 crc kubenswrapper[4813]: I1201 09:03:30.902978 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9a0e144a-6298-4092-86fc-ba45b986d6ec-dev" (OuterVolumeSpecName: "dev") pod "9a0e144a-6298-4092-86fc-ba45b986d6ec" (UID: "9a0e144a-6298-4092-86fc-ba45b986d6ec"). InnerVolumeSpecName "dev". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 09:03:30 crc kubenswrapper[4813]: I1201 09:03:30.903014 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9a0e144a-6298-4092-86fc-ba45b986d6ec-var-locks-brick" (OuterVolumeSpecName: "var-locks-brick") pod "9a0e144a-6298-4092-86fc-ba45b986d6ec" (UID: "9a0e144a-6298-4092-86fc-ba45b986d6ec"). InnerVolumeSpecName "var-locks-brick". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 09:03:30 crc kubenswrapper[4813]: I1201 09:03:30.902977 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9a0e144a-6298-4092-86fc-ba45b986d6ec-run" (OuterVolumeSpecName: "run") pod "9a0e144a-6298-4092-86fc-ba45b986d6ec" (UID: "9a0e144a-6298-4092-86fc-ba45b986d6ec"). InnerVolumeSpecName "run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 09:03:30 crc kubenswrapper[4813]: I1201 09:03:30.902993 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9a0e144a-6298-4092-86fc-ba45b986d6ec-logs\") pod \"9a0e144a-6298-4092-86fc-ba45b986d6ec\" (UID: \"9a0e144a-6298-4092-86fc-ba45b986d6ec\") " Dec 01 09:03:30 crc kubenswrapper[4813]: I1201 09:03:30.903116 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tj5p6\" (UniqueName: \"kubernetes.io/projected/9a0e144a-6298-4092-86fc-ba45b986d6ec-kube-api-access-tj5p6\") pod \"9a0e144a-6298-4092-86fc-ba45b986d6ec\" (UID: \"9a0e144a-6298-4092-86fc-ba45b986d6ec\") " Dec 01 09:03:30 crc kubenswrapper[4813]: I1201 09:03:30.903230 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/9a0e144a-6298-4092-86fc-ba45b986d6ec-httpd-run\") pod \"9a0e144a-6298-4092-86fc-ba45b986d6ec\" (UID: \"9a0e144a-6298-4092-86fc-ba45b986d6ec\") " Dec 01 09:03:30 crc kubenswrapper[4813]: I1201 09:03:30.903258 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/9a0e144a-6298-4092-86fc-ba45b986d6ec-lib-modules\") pod \"9a0e144a-6298-4092-86fc-ba45b986d6ec\" (UID: \"9a0e144a-6298-4092-86fc-ba45b986d6ec\") " Dec 01 09:03:30 crc kubenswrapper[4813]: I1201 09:03:30.903242 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9a0e144a-6298-4092-86fc-ba45b986d6ec-etc-nvme" (OuterVolumeSpecName: "etc-nvme") pod "9a0e144a-6298-4092-86fc-ba45b986d6ec" (UID: "9a0e144a-6298-4092-86fc-ba45b986d6ec"). InnerVolumeSpecName "etc-nvme". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 09:03:30 crc kubenswrapper[4813]: I1201 09:03:30.903280 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9a0e144a-6298-4092-86fc-ba45b986d6ec-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "9a0e144a-6298-4092-86fc-ba45b986d6ec" (UID: "9a0e144a-6298-4092-86fc-ba45b986d6ec"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 09:03:30 crc kubenswrapper[4813]: I1201 09:03:30.903308 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9a0e144a-6298-4092-86fc-ba45b986d6ec-etc-iscsi" (OuterVolumeSpecName: "etc-iscsi") pod "9a0e144a-6298-4092-86fc-ba45b986d6ec" (UID: "9a0e144a-6298-4092-86fc-ba45b986d6ec"). InnerVolumeSpecName "etc-iscsi". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 09:03:30 crc kubenswrapper[4813]: I1201 09:03:30.903286 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/9a0e144a-6298-4092-86fc-ba45b986d6ec-etc-iscsi\") pod \"9a0e144a-6298-4092-86fc-ba45b986d6ec\" (UID: \"9a0e144a-6298-4092-86fc-ba45b986d6ec\") " Dec 01 09:03:30 crc kubenswrapper[4813]: I1201 09:03:30.903406 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage16-crc\") pod \"9a0e144a-6298-4092-86fc-ba45b986d6ec\" (UID: \"9a0e144a-6298-4092-86fc-ba45b986d6ec\") " Dec 01 09:03:30 crc kubenswrapper[4813]: I1201 09:03:30.903434 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/9a0e144a-6298-4092-86fc-ba45b986d6ec-sys\") pod \"9a0e144a-6298-4092-86fc-ba45b986d6ec\" (UID: \"9a0e144a-6298-4092-86fc-ba45b986d6ec\") " Dec 01 09:03:30 crc kubenswrapper[4813]: I1201 09:03:30.903471 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9a0e144a-6298-4092-86fc-ba45b986d6ec-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "9a0e144a-6298-4092-86fc-ba45b986d6ec" (UID: "9a0e144a-6298-4092-86fc-ba45b986d6ec"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 09:03:30 crc kubenswrapper[4813]: I1201 09:03:30.903616 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9a0e144a-6298-4092-86fc-ba45b986d6ec-sys" (OuterVolumeSpecName: "sys") pod "9a0e144a-6298-4092-86fc-ba45b986d6ec" (UID: "9a0e144a-6298-4092-86fc-ba45b986d6ec"). InnerVolumeSpecName "sys". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 09:03:30 crc kubenswrapper[4813]: I1201 09:03:30.903707 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9a0e144a-6298-4092-86fc-ba45b986d6ec-logs" (OuterVolumeSpecName: "logs") pod "9a0e144a-6298-4092-86fc-ba45b986d6ec" (UID: "9a0e144a-6298-4092-86fc-ba45b986d6ec"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 09:03:30 crc kubenswrapper[4813]: I1201 09:03:30.904100 4813 reconciler_common.go:293] "Volume detached for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/9a0e144a-6298-4092-86fc-ba45b986d6ec-var-locks-brick\") on node \"crc\" DevicePath \"\"" Dec 01 09:03:30 crc kubenswrapper[4813]: I1201 09:03:30.904120 4813 reconciler_common.go:293] "Volume detached for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/9a0e144a-6298-4092-86fc-ba45b986d6ec-etc-nvme\") on node \"crc\" DevicePath \"\"" Dec 01 09:03:30 crc kubenswrapper[4813]: I1201 09:03:30.904133 4813 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9a0e144a-6298-4092-86fc-ba45b986d6ec-logs\") on node \"crc\" DevicePath \"\"" Dec 01 09:03:30 crc kubenswrapper[4813]: I1201 09:03:30.904145 4813 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/9a0e144a-6298-4092-86fc-ba45b986d6ec-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 01 09:03:30 crc kubenswrapper[4813]: I1201 09:03:30.904156 4813 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/9a0e144a-6298-4092-86fc-ba45b986d6ec-lib-modules\") on node \"crc\" DevicePath \"\"" Dec 01 09:03:30 crc kubenswrapper[4813]: I1201 09:03:30.904167 4813 reconciler_common.go:293] "Volume detached for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/9a0e144a-6298-4092-86fc-ba45b986d6ec-etc-iscsi\") on node \"crc\" DevicePath \"\"" Dec 01 09:03:30 crc kubenswrapper[4813]: I1201 09:03:30.904178 4813 reconciler_common.go:293] "Volume detached for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/9a0e144a-6298-4092-86fc-ba45b986d6ec-sys\") on node \"crc\" DevicePath \"\"" Dec 01 09:03:30 crc kubenswrapper[4813]: I1201 09:03:30.904191 4813 reconciler_common.go:293] "Volume detached for volume \"run\" (UniqueName: \"kubernetes.io/host-path/9a0e144a-6298-4092-86fc-ba45b986d6ec-run\") on node \"crc\" DevicePath \"\"" Dec 01 09:03:30 crc kubenswrapper[4813]: I1201 09:03:30.904202 4813 reconciler_common.go:293] "Volume detached for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/9a0e144a-6298-4092-86fc-ba45b986d6ec-dev\") on node \"crc\" DevicePath \"\"" Dec 01 09:03:30 crc kubenswrapper[4813]: I1201 09:03:30.908416 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage16-crc" (OuterVolumeSpecName: "glance") pod "9a0e144a-6298-4092-86fc-ba45b986d6ec" (UID: "9a0e144a-6298-4092-86fc-ba45b986d6ec"). InnerVolumeSpecName "local-storage16-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 01 09:03:30 crc kubenswrapper[4813]: I1201 09:03:30.908776 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage10-crc" (OuterVolumeSpecName: "glance-cache") pod "9a0e144a-6298-4092-86fc-ba45b986d6ec" (UID: "9a0e144a-6298-4092-86fc-ba45b986d6ec"). InnerVolumeSpecName "local-storage10-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 01 09:03:30 crc kubenswrapper[4813]: I1201 09:03:30.908938 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9a0e144a-6298-4092-86fc-ba45b986d6ec-scripts" (OuterVolumeSpecName: "scripts") pod "9a0e144a-6298-4092-86fc-ba45b986d6ec" (UID: "9a0e144a-6298-4092-86fc-ba45b986d6ec"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:03:30 crc kubenswrapper[4813]: I1201 09:03:30.909595 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9a0e144a-6298-4092-86fc-ba45b986d6ec-kube-api-access-tj5p6" (OuterVolumeSpecName: "kube-api-access-tj5p6") pod "9a0e144a-6298-4092-86fc-ba45b986d6ec" (UID: "9a0e144a-6298-4092-86fc-ba45b986d6ec"). InnerVolumeSpecName "kube-api-access-tj5p6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:03:30 crc kubenswrapper[4813]: I1201 09:03:30.944097 4813 generic.go:334] "Generic (PLEG): container finished" podID="9a0e144a-6298-4092-86fc-ba45b986d6ec" containerID="c0918867756a6c3013f771a8e9cbf8e78fd40a1e5f1cc6fa2576633699e9e100" exitCode=0 Dec 01 09:03:30 crc kubenswrapper[4813]: I1201 09:03:30.944157 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"9a0e144a-6298-4092-86fc-ba45b986d6ec","Type":"ContainerDied","Data":"c0918867756a6c3013f771a8e9cbf8e78fd40a1e5f1cc6fa2576633699e9e100"} Dec 01 09:03:30 crc kubenswrapper[4813]: I1201 09:03:30.944196 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"9a0e144a-6298-4092-86fc-ba45b986d6ec","Type":"ContainerDied","Data":"9c4e446fc54267ec7444deb2562c0511cf39dac043ea8128c8e381d5edf0d1f4"} Dec 01 09:03:30 crc kubenswrapper[4813]: I1201 09:03:30.944224 4813 scope.go:117] "RemoveContainer" containerID="c0918867756a6c3013f771a8e9cbf8e78fd40a1e5f1cc6fa2576633699e9e100" Dec 01 09:03:30 crc kubenswrapper[4813]: I1201 09:03:30.944433 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Dec 01 09:03:30 crc kubenswrapper[4813]: I1201 09:03:30.951713 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9a0e144a-6298-4092-86fc-ba45b986d6ec-config-data" (OuterVolumeSpecName: "config-data") pod "9a0e144a-6298-4092-86fc-ba45b986d6ec" (UID: "9a0e144a-6298-4092-86fc-ba45b986d6ec"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:03:31 crc kubenswrapper[4813]: I1201 09:03:31.003556 4813 scope.go:117] "RemoveContainer" containerID="db306b76d1e4d93ad74ca563b646090f0237e7fdc176fc18b82ff6a6cc3e5413" Dec 01 09:03:31 crc kubenswrapper[4813]: I1201 09:03:31.005200 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tj5p6\" (UniqueName: \"kubernetes.io/projected/9a0e144a-6298-4092-86fc-ba45b986d6ec-kube-api-access-tj5p6\") on node \"crc\" DevicePath \"\"" Dec 01 09:03:31 crc kubenswrapper[4813]: I1201 09:03:31.005238 4813 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage16-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage16-crc\") on node \"crc\" " Dec 01 09:03:31 crc kubenswrapper[4813]: I1201 09:03:31.005248 4813 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9a0e144a-6298-4092-86fc-ba45b986d6ec-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 09:03:31 crc kubenswrapper[4813]: I1201 09:03:31.005264 4813 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" " Dec 01 09:03:31 crc kubenswrapper[4813]: I1201 09:03:31.005274 4813 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9a0e144a-6298-4092-86fc-ba45b986d6ec-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 09:03:31 crc kubenswrapper[4813]: I1201 09:03:31.020151 4813 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage10-crc" (UniqueName: "kubernetes.io/local-volume/local-storage10-crc") on node "crc" Dec 01 09:03:31 crc kubenswrapper[4813]: I1201 09:03:31.025487 4813 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage16-crc" (UniqueName: "kubernetes.io/local-volume/local-storage16-crc") on node "crc" Dec 01 09:03:31 crc kubenswrapper[4813]: I1201 09:03:31.032672 4813 scope.go:117] "RemoveContainer" containerID="c0918867756a6c3013f771a8e9cbf8e78fd40a1e5f1cc6fa2576633699e9e100" Dec 01 09:03:31 crc kubenswrapper[4813]: E1201 09:03:31.034192 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c0918867756a6c3013f771a8e9cbf8e78fd40a1e5f1cc6fa2576633699e9e100\": container with ID starting with c0918867756a6c3013f771a8e9cbf8e78fd40a1e5f1cc6fa2576633699e9e100 not found: ID does not exist" containerID="c0918867756a6c3013f771a8e9cbf8e78fd40a1e5f1cc6fa2576633699e9e100" Dec 01 09:03:31 crc kubenswrapper[4813]: I1201 09:03:31.034241 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c0918867756a6c3013f771a8e9cbf8e78fd40a1e5f1cc6fa2576633699e9e100"} err="failed to get container status \"c0918867756a6c3013f771a8e9cbf8e78fd40a1e5f1cc6fa2576633699e9e100\": rpc error: code = NotFound desc = could not find container \"c0918867756a6c3013f771a8e9cbf8e78fd40a1e5f1cc6fa2576633699e9e100\": container with ID starting with c0918867756a6c3013f771a8e9cbf8e78fd40a1e5f1cc6fa2576633699e9e100 not found: ID does not exist" Dec 01 09:03:31 crc kubenswrapper[4813]: I1201 09:03:31.034280 4813 scope.go:117] "RemoveContainer" containerID="db306b76d1e4d93ad74ca563b646090f0237e7fdc176fc18b82ff6a6cc3e5413" Dec 01 09:03:31 crc kubenswrapper[4813]: E1201 09:03:31.034908 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"db306b76d1e4d93ad74ca563b646090f0237e7fdc176fc18b82ff6a6cc3e5413\": container with ID starting with db306b76d1e4d93ad74ca563b646090f0237e7fdc176fc18b82ff6a6cc3e5413 not found: ID does not exist" containerID="db306b76d1e4d93ad74ca563b646090f0237e7fdc176fc18b82ff6a6cc3e5413" Dec 01 09:03:31 crc kubenswrapper[4813]: I1201 09:03:31.034958 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"db306b76d1e4d93ad74ca563b646090f0237e7fdc176fc18b82ff6a6cc3e5413"} err="failed to get container status \"db306b76d1e4d93ad74ca563b646090f0237e7fdc176fc18b82ff6a6cc3e5413\": rpc error: code = NotFound desc = could not find container \"db306b76d1e4d93ad74ca563b646090f0237e7fdc176fc18b82ff6a6cc3e5413\": container with ID starting with db306b76d1e4d93ad74ca563b646090f0237e7fdc176fc18b82ff6a6cc3e5413 not found: ID does not exist" Dec 01 09:03:31 crc kubenswrapper[4813]: I1201 09:03:31.107072 4813 reconciler_common.go:293] "Volume detached for volume \"local-storage16-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage16-crc\") on node \"crc\" DevicePath \"\"" Dec 01 09:03:31 crc kubenswrapper[4813]: I1201 09:03:31.107108 4813 reconciler_common.go:293] "Volume detached for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" DevicePath \"\"" Dec 01 09:03:31 crc kubenswrapper[4813]: I1201 09:03:31.284801 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Dec 01 09:03:31 crc kubenswrapper[4813]: I1201 09:03:31.290796 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Dec 01 09:03:32 crc kubenswrapper[4813]: I1201 09:03:32.222702 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-db-sync-rdsjx"] Dec 01 09:03:32 crc kubenswrapper[4813]: I1201 09:03:32.248144 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-db-sync-rdsjx"] Dec 01 09:03:32 crc kubenswrapper[4813]: I1201 09:03:32.272100 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glancec6e3-account-delete-kcfsh"] Dec 01 09:03:32 crc kubenswrapper[4813]: E1201 09:03:32.272516 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a0e144a-6298-4092-86fc-ba45b986d6ec" containerName="glance-log" Dec 01 09:03:32 crc kubenswrapper[4813]: I1201 09:03:32.272551 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a0e144a-6298-4092-86fc-ba45b986d6ec" containerName="glance-log" Dec 01 09:03:32 crc kubenswrapper[4813]: E1201 09:03:32.272573 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a0e144a-6298-4092-86fc-ba45b986d6ec" containerName="glance-httpd" Dec 01 09:03:32 crc kubenswrapper[4813]: I1201 09:03:32.272584 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a0e144a-6298-4092-86fc-ba45b986d6ec" containerName="glance-httpd" Dec 01 09:03:32 crc kubenswrapper[4813]: E1201 09:03:32.272597 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="10f3f59e-3661-4248-aac9-8061818231ef" containerName="glance-log" Dec 01 09:03:32 crc kubenswrapper[4813]: I1201 09:03:32.272605 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="10f3f59e-3661-4248-aac9-8061818231ef" containerName="glance-log" Dec 01 09:03:32 crc kubenswrapper[4813]: E1201 09:03:32.272620 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9ea2713f-188e-4e62-9b40-16ec13591054" containerName="glance-httpd" Dec 01 09:03:32 crc kubenswrapper[4813]: I1201 09:03:32.272627 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="9ea2713f-188e-4e62-9b40-16ec13591054" containerName="glance-httpd" Dec 01 09:03:32 crc kubenswrapper[4813]: E1201 09:03:32.272642 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9ea2713f-188e-4e62-9b40-16ec13591054" containerName="glance-log" Dec 01 09:03:32 crc kubenswrapper[4813]: I1201 09:03:32.272650 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="9ea2713f-188e-4e62-9b40-16ec13591054" containerName="glance-log" Dec 01 09:03:32 crc kubenswrapper[4813]: E1201 09:03:32.272663 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="10f3f59e-3661-4248-aac9-8061818231ef" containerName="glance-httpd" Dec 01 09:03:32 crc kubenswrapper[4813]: I1201 09:03:32.272670 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="10f3f59e-3661-4248-aac9-8061818231ef" containerName="glance-httpd" Dec 01 09:03:32 crc kubenswrapper[4813]: I1201 09:03:32.272893 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="9ea2713f-188e-4e62-9b40-16ec13591054" containerName="glance-log" Dec 01 09:03:32 crc kubenswrapper[4813]: I1201 09:03:32.272918 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="9ea2713f-188e-4e62-9b40-16ec13591054" containerName="glance-httpd" Dec 01 09:03:32 crc kubenswrapper[4813]: I1201 09:03:32.272931 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="9a0e144a-6298-4092-86fc-ba45b986d6ec" containerName="glance-log" Dec 01 09:03:32 crc kubenswrapper[4813]: I1201 09:03:32.273000 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="10f3f59e-3661-4248-aac9-8061818231ef" containerName="glance-httpd" Dec 01 09:03:32 crc kubenswrapper[4813]: I1201 09:03:32.273012 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="10f3f59e-3661-4248-aac9-8061818231ef" containerName="glance-log" Dec 01 09:03:32 crc kubenswrapper[4813]: I1201 09:03:32.273024 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="9a0e144a-6298-4092-86fc-ba45b986d6ec" containerName="glance-httpd" Dec 01 09:03:32 crc kubenswrapper[4813]: I1201 09:03:32.273698 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glancec6e3-account-delete-kcfsh" Dec 01 09:03:32 crc kubenswrapper[4813]: I1201 09:03:32.293596 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glancec6e3-account-delete-kcfsh"] Dec 01 09:03:32 crc kubenswrapper[4813]: I1201 09:03:32.402767 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9a0e144a-6298-4092-86fc-ba45b986d6ec" path="/var/lib/kubelet/pods/9a0e144a-6298-4092-86fc-ba45b986d6ec/volumes" Dec 01 09:03:32 crc kubenswrapper[4813]: I1201 09:03:32.403366 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9ffc7257-c24e-4234-a314-e9a61f9e436f" path="/var/lib/kubelet/pods/9ffc7257-c24e-4234-a314-e9a61f9e436f/volumes" Dec 01 09:03:32 crc kubenswrapper[4813]: I1201 09:03:32.423475 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/57cfd932-24f2-4afa-867b-4dfe7a4ed005-operator-scripts\") pod \"glancec6e3-account-delete-kcfsh\" (UID: \"57cfd932-24f2-4afa-867b-4dfe7a4ed005\") " pod="glance-kuttl-tests/glancec6e3-account-delete-kcfsh" Dec 01 09:03:32 crc kubenswrapper[4813]: I1201 09:03:32.423568 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6ghqx\" (UniqueName: \"kubernetes.io/projected/57cfd932-24f2-4afa-867b-4dfe7a4ed005-kube-api-access-6ghqx\") pod \"glancec6e3-account-delete-kcfsh\" (UID: \"57cfd932-24f2-4afa-867b-4dfe7a4ed005\") " pod="glance-kuttl-tests/glancec6e3-account-delete-kcfsh" Dec 01 09:03:32 crc kubenswrapper[4813]: I1201 09:03:32.525041 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6ghqx\" (UniqueName: \"kubernetes.io/projected/57cfd932-24f2-4afa-867b-4dfe7a4ed005-kube-api-access-6ghqx\") pod \"glancec6e3-account-delete-kcfsh\" (UID: \"57cfd932-24f2-4afa-867b-4dfe7a4ed005\") " pod="glance-kuttl-tests/glancec6e3-account-delete-kcfsh" Dec 01 09:03:32 crc kubenswrapper[4813]: I1201 09:03:32.526435 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/57cfd932-24f2-4afa-867b-4dfe7a4ed005-operator-scripts\") pod \"glancec6e3-account-delete-kcfsh\" (UID: \"57cfd932-24f2-4afa-867b-4dfe7a4ed005\") " pod="glance-kuttl-tests/glancec6e3-account-delete-kcfsh" Dec 01 09:03:32 crc kubenswrapper[4813]: I1201 09:03:32.526810 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/57cfd932-24f2-4afa-867b-4dfe7a4ed005-operator-scripts\") pod \"glancec6e3-account-delete-kcfsh\" (UID: \"57cfd932-24f2-4afa-867b-4dfe7a4ed005\") " pod="glance-kuttl-tests/glancec6e3-account-delete-kcfsh" Dec 01 09:03:32 crc kubenswrapper[4813]: I1201 09:03:32.555738 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6ghqx\" (UniqueName: \"kubernetes.io/projected/57cfd932-24f2-4afa-867b-4dfe7a4ed005-kube-api-access-6ghqx\") pod \"glancec6e3-account-delete-kcfsh\" (UID: \"57cfd932-24f2-4afa-867b-4dfe7a4ed005\") " pod="glance-kuttl-tests/glancec6e3-account-delete-kcfsh" Dec 01 09:03:32 crc kubenswrapper[4813]: I1201 09:03:32.588680 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glancec6e3-account-delete-kcfsh" Dec 01 09:03:33 crc kubenswrapper[4813]: I1201 09:03:33.033764 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glancec6e3-account-delete-kcfsh"] Dec 01 09:03:33 crc kubenswrapper[4813]: W1201 09:03:33.042785 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod57cfd932_24f2_4afa_867b_4dfe7a4ed005.slice/crio-2f7acd9b934b1bced6c412825eb5852c7016d1803fd97c814617dfb4e15f89aa WatchSource:0}: Error finding container 2f7acd9b934b1bced6c412825eb5852c7016d1803fd97c814617dfb4e15f89aa: Status 404 returned error can't find the container with id 2f7acd9b934b1bced6c412825eb5852c7016d1803fd97c814617dfb4e15f89aa Dec 01 09:03:33 crc kubenswrapper[4813]: I1201 09:03:33.968173 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glancec6e3-account-delete-kcfsh" event={"ID":"57cfd932-24f2-4afa-867b-4dfe7a4ed005","Type":"ContainerStarted","Data":"1c35642330f47379d43b8119514c1a7daa3e3e9f0dd54bef1d33b46a9d211985"} Dec 01 09:03:33 crc kubenswrapper[4813]: I1201 09:03:33.968543 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glancec6e3-account-delete-kcfsh" event={"ID":"57cfd932-24f2-4afa-867b-4dfe7a4ed005","Type":"ContainerStarted","Data":"2f7acd9b934b1bced6c412825eb5852c7016d1803fd97c814617dfb4e15f89aa"} Dec 01 09:03:33 crc kubenswrapper[4813]: I1201 09:03:33.985215 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glancec6e3-account-delete-kcfsh" podStartSLOduration=1.985187378 podStartE2EDuration="1.985187378s" podCreationTimestamp="2025-12-01 09:03:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 09:03:33.983519911 +0000 UTC m=+1398.606441497" watchObservedRunningTime="2025-12-01 09:03:33.985187378 +0000 UTC m=+1398.608108964" Dec 01 09:03:35 crc kubenswrapper[4813]: I1201 09:03:35.512140 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/openstackclient"] Dec 01 09:03:35 crc kubenswrapper[4813]: I1201 09:03:35.513635 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/openstackclient" Dec 01 09:03:35 crc kubenswrapper[4813]: I1201 09:03:35.516455 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"openstack-config" Dec 01 09:03:35 crc kubenswrapper[4813]: I1201 09:03:35.520452 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"openstack-scripts-9db6gc427h" Dec 01 09:03:35 crc kubenswrapper[4813]: I1201 09:03:35.520487 4813 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"default-dockercfg-r6mcj" Dec 01 09:03:35 crc kubenswrapper[4813]: I1201 09:03:35.520504 4813 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"openstack-config-secret" Dec 01 09:03:35 crc kubenswrapper[4813]: I1201 09:03:35.526563 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/openstackclient"] Dec 01 09:03:35 crc kubenswrapper[4813]: I1201 09:03:35.678078 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g4hv4\" (UniqueName: \"kubernetes.io/projected/4d0873f6-c7e7-4a9b-bd6e-a0263fd2bc5a-kube-api-access-g4hv4\") pod \"openstackclient\" (UID: \"4d0873f6-c7e7-4a9b-bd6e-a0263fd2bc5a\") " pod="glance-kuttl-tests/openstackclient" Dec 01 09:03:35 crc kubenswrapper[4813]: I1201 09:03:35.678135 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-scripts\" (UniqueName: \"kubernetes.io/configmap/4d0873f6-c7e7-4a9b-bd6e-a0263fd2bc5a-openstack-scripts\") pod \"openstackclient\" (UID: \"4d0873f6-c7e7-4a9b-bd6e-a0263fd2bc5a\") " pod="glance-kuttl-tests/openstackclient" Dec 01 09:03:35 crc kubenswrapper[4813]: I1201 09:03:35.678197 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/4d0873f6-c7e7-4a9b-bd6e-a0263fd2bc5a-openstack-config-secret\") pod \"openstackclient\" (UID: \"4d0873f6-c7e7-4a9b-bd6e-a0263fd2bc5a\") " pod="glance-kuttl-tests/openstackclient" Dec 01 09:03:35 crc kubenswrapper[4813]: I1201 09:03:35.678215 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/4d0873f6-c7e7-4a9b-bd6e-a0263fd2bc5a-openstack-config\") pod \"openstackclient\" (UID: \"4d0873f6-c7e7-4a9b-bd6e-a0263fd2bc5a\") " pod="glance-kuttl-tests/openstackclient" Dec 01 09:03:35 crc kubenswrapper[4813]: I1201 09:03:35.779624 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g4hv4\" (UniqueName: \"kubernetes.io/projected/4d0873f6-c7e7-4a9b-bd6e-a0263fd2bc5a-kube-api-access-g4hv4\") pod \"openstackclient\" (UID: \"4d0873f6-c7e7-4a9b-bd6e-a0263fd2bc5a\") " pod="glance-kuttl-tests/openstackclient" Dec 01 09:03:35 crc kubenswrapper[4813]: I1201 09:03:35.779682 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-scripts\" (UniqueName: \"kubernetes.io/configmap/4d0873f6-c7e7-4a9b-bd6e-a0263fd2bc5a-openstack-scripts\") pod \"openstackclient\" (UID: \"4d0873f6-c7e7-4a9b-bd6e-a0263fd2bc5a\") " pod="glance-kuttl-tests/openstackclient" Dec 01 09:03:35 crc kubenswrapper[4813]: I1201 09:03:35.779750 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/4d0873f6-c7e7-4a9b-bd6e-a0263fd2bc5a-openstack-config-secret\") pod \"openstackclient\" (UID: \"4d0873f6-c7e7-4a9b-bd6e-a0263fd2bc5a\") " pod="glance-kuttl-tests/openstackclient" Dec 01 09:03:35 crc kubenswrapper[4813]: I1201 09:03:35.779791 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/4d0873f6-c7e7-4a9b-bd6e-a0263fd2bc5a-openstack-config\") pod \"openstackclient\" (UID: \"4d0873f6-c7e7-4a9b-bd6e-a0263fd2bc5a\") " pod="glance-kuttl-tests/openstackclient" Dec 01 09:03:35 crc kubenswrapper[4813]: I1201 09:03:35.781369 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/4d0873f6-c7e7-4a9b-bd6e-a0263fd2bc5a-openstack-config\") pod \"openstackclient\" (UID: \"4d0873f6-c7e7-4a9b-bd6e-a0263fd2bc5a\") " pod="glance-kuttl-tests/openstackclient" Dec 01 09:03:35 crc kubenswrapper[4813]: I1201 09:03:35.782600 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-scripts\" (UniqueName: \"kubernetes.io/configmap/4d0873f6-c7e7-4a9b-bd6e-a0263fd2bc5a-openstack-scripts\") pod \"openstackclient\" (UID: \"4d0873f6-c7e7-4a9b-bd6e-a0263fd2bc5a\") " pod="glance-kuttl-tests/openstackclient" Dec 01 09:03:35 crc kubenswrapper[4813]: I1201 09:03:35.787355 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/4d0873f6-c7e7-4a9b-bd6e-a0263fd2bc5a-openstack-config-secret\") pod \"openstackclient\" (UID: \"4d0873f6-c7e7-4a9b-bd6e-a0263fd2bc5a\") " pod="glance-kuttl-tests/openstackclient" Dec 01 09:03:35 crc kubenswrapper[4813]: I1201 09:03:35.796474 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g4hv4\" (UniqueName: \"kubernetes.io/projected/4d0873f6-c7e7-4a9b-bd6e-a0263fd2bc5a-kube-api-access-g4hv4\") pod \"openstackclient\" (UID: \"4d0873f6-c7e7-4a9b-bd6e-a0263fd2bc5a\") " pod="glance-kuttl-tests/openstackclient" Dec 01 09:03:35 crc kubenswrapper[4813]: I1201 09:03:35.844883 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/openstackclient" Dec 01 09:03:35 crc kubenswrapper[4813]: I1201 09:03:35.984929 4813 generic.go:334] "Generic (PLEG): container finished" podID="57cfd932-24f2-4afa-867b-4dfe7a4ed005" containerID="1c35642330f47379d43b8119514c1a7daa3e3e9f0dd54bef1d33b46a9d211985" exitCode=0 Dec 01 09:03:35 crc kubenswrapper[4813]: I1201 09:03:35.985244 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glancec6e3-account-delete-kcfsh" event={"ID":"57cfd932-24f2-4afa-867b-4dfe7a4ed005","Type":"ContainerDied","Data":"1c35642330f47379d43b8119514c1a7daa3e3e9f0dd54bef1d33b46a9d211985"} Dec 01 09:03:36 crc kubenswrapper[4813]: I1201 09:03:36.070361 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/openstackclient"] Dec 01 09:03:37 crc kubenswrapper[4813]: I1201 09:03:37.000802 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstackclient" event={"ID":"4d0873f6-c7e7-4a9b-bd6e-a0263fd2bc5a","Type":"ContainerStarted","Data":"2a2a923cac5075245c80fdf031dd3c61fd96a1b67bde407b41b650932dfc7b23"} Dec 01 09:03:37 crc kubenswrapper[4813]: I1201 09:03:37.003041 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstackclient" event={"ID":"4d0873f6-c7e7-4a9b-bd6e-a0263fd2bc5a","Type":"ContainerStarted","Data":"50df357971be6487725cf98c6ed4b38a73cc8fbd946e44f23ba23dfd24200a24"} Dec 01 09:03:37 crc kubenswrapper[4813]: I1201 09:03:37.026733 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/openstackclient" podStartSLOduration=2.026700564 podStartE2EDuration="2.026700564s" podCreationTimestamp="2025-12-01 09:03:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 09:03:37.021491097 +0000 UTC m=+1401.644412733" watchObservedRunningTime="2025-12-01 09:03:37.026700564 +0000 UTC m=+1401.649622190" Dec 01 09:03:37 crc kubenswrapper[4813]: I1201 09:03:37.322266 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glancec6e3-account-delete-kcfsh" Dec 01 09:03:37 crc kubenswrapper[4813]: I1201 09:03:37.406297 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/57cfd932-24f2-4afa-867b-4dfe7a4ed005-operator-scripts\") pod \"57cfd932-24f2-4afa-867b-4dfe7a4ed005\" (UID: \"57cfd932-24f2-4afa-867b-4dfe7a4ed005\") " Dec 01 09:03:37 crc kubenswrapper[4813]: I1201 09:03:37.406449 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ghqx\" (UniqueName: \"kubernetes.io/projected/57cfd932-24f2-4afa-867b-4dfe7a4ed005-kube-api-access-6ghqx\") pod \"57cfd932-24f2-4afa-867b-4dfe7a4ed005\" (UID: \"57cfd932-24f2-4afa-867b-4dfe7a4ed005\") " Dec 01 09:03:37 crc kubenswrapper[4813]: I1201 09:03:37.406903 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/57cfd932-24f2-4afa-867b-4dfe7a4ed005-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "57cfd932-24f2-4afa-867b-4dfe7a4ed005" (UID: "57cfd932-24f2-4afa-867b-4dfe7a4ed005"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 09:03:37 crc kubenswrapper[4813]: I1201 09:03:37.411403 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57cfd932-24f2-4afa-867b-4dfe7a4ed005-kube-api-access-6ghqx" (OuterVolumeSpecName: "kube-api-access-6ghqx") pod "57cfd932-24f2-4afa-867b-4dfe7a4ed005" (UID: "57cfd932-24f2-4afa-867b-4dfe7a4ed005"). InnerVolumeSpecName "kube-api-access-6ghqx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:03:37 crc kubenswrapper[4813]: I1201 09:03:37.507941 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ghqx\" (UniqueName: \"kubernetes.io/projected/57cfd932-24f2-4afa-867b-4dfe7a4ed005-kube-api-access-6ghqx\") on node \"crc\" DevicePath \"\"" Dec 01 09:03:37 crc kubenswrapper[4813]: I1201 09:03:37.508332 4813 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/57cfd932-24f2-4afa-867b-4dfe7a4ed005-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 09:03:38 crc kubenswrapper[4813]: I1201 09:03:38.013788 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glancec6e3-account-delete-kcfsh" event={"ID":"57cfd932-24f2-4afa-867b-4dfe7a4ed005","Type":"ContainerDied","Data":"2f7acd9b934b1bced6c412825eb5852c7016d1803fd97c814617dfb4e15f89aa"} Dec 01 09:03:38 crc kubenswrapper[4813]: I1201 09:03:38.013929 4813 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2f7acd9b934b1bced6c412825eb5852c7016d1803fd97c814617dfb4e15f89aa" Dec 01 09:03:38 crc kubenswrapper[4813]: I1201 09:03:38.014203 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glancec6e3-account-delete-kcfsh" Dec 01 09:03:42 crc kubenswrapper[4813]: I1201 09:03:42.316173 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-db-create-p5mf4"] Dec 01 09:03:42 crc kubenswrapper[4813]: I1201 09:03:42.323330 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-c6e3-account-create-update-h46sk"] Dec 01 09:03:42 crc kubenswrapper[4813]: I1201 09:03:42.330459 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glancec6e3-account-delete-kcfsh"] Dec 01 09:03:42 crc kubenswrapper[4813]: I1201 09:03:42.336616 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-c6e3-account-create-update-h46sk"] Dec 01 09:03:42 crc kubenswrapper[4813]: I1201 09:03:42.342798 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-db-create-p5mf4"] Dec 01 09:03:42 crc kubenswrapper[4813]: I1201 09:03:42.349073 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glancec6e3-account-delete-kcfsh"] Dec 01 09:03:42 crc kubenswrapper[4813]: I1201 09:03:42.403946 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="11bd67f4-fce9-479e-858b-44e6b1829d81" path="/var/lib/kubelet/pods/11bd67f4-fce9-479e-858b-44e6b1829d81/volumes" Dec 01 09:03:42 crc kubenswrapper[4813]: I1201 09:03:42.404469 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57cfd932-24f2-4afa-867b-4dfe7a4ed005" path="/var/lib/kubelet/pods/57cfd932-24f2-4afa-867b-4dfe7a4ed005/volumes" Dec 01 09:03:42 crc kubenswrapper[4813]: I1201 09:03:42.404945 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d3dfa26e-2774-4129-9bfc-5cf40aa3e942" path="/var/lib/kubelet/pods/d3dfa26e-2774-4129-9bfc-5cf40aa3e942/volumes" Dec 01 09:03:42 crc kubenswrapper[4813]: I1201 09:03:42.405414 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-db-create-fcjt9"] Dec 01 09:03:42 crc kubenswrapper[4813]: E1201 09:03:42.405768 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="57cfd932-24f2-4afa-867b-4dfe7a4ed005" containerName="mariadb-account-delete" Dec 01 09:03:42 crc kubenswrapper[4813]: I1201 09:03:42.405803 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="57cfd932-24f2-4afa-867b-4dfe7a4ed005" containerName="mariadb-account-delete" Dec 01 09:03:42 crc kubenswrapper[4813]: I1201 09:03:42.406049 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="57cfd932-24f2-4afa-867b-4dfe7a4ed005" containerName="mariadb-account-delete" Dec 01 09:03:42 crc kubenswrapper[4813]: I1201 09:03:42.406657 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-create-fcjt9"] Dec 01 09:03:42 crc kubenswrapper[4813]: I1201 09:03:42.406763 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-fcjt9" Dec 01 09:03:42 crc kubenswrapper[4813]: I1201 09:03:42.495152 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-a9c4-account-create-update-4g8nj"] Dec 01 09:03:42 crc kubenswrapper[4813]: I1201 09:03:42.496264 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-a9c4-account-create-update-4g8nj" Dec 01 09:03:42 crc kubenswrapper[4813]: I1201 09:03:42.499397 4813 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-db-secret" Dec 01 09:03:42 crc kubenswrapper[4813]: I1201 09:03:42.510200 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-a9c4-account-create-update-4g8nj"] Dec 01 09:03:42 crc kubenswrapper[4813]: I1201 09:03:42.568405 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/77e7d212-07ac-4c29-acb2-4e83bcf6d7e0-operator-scripts\") pod \"glance-db-create-fcjt9\" (UID: \"77e7d212-07ac-4c29-acb2-4e83bcf6d7e0\") " pod="glance-kuttl-tests/glance-db-create-fcjt9" Dec 01 09:03:42 crc kubenswrapper[4813]: I1201 09:03:42.568585 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5psb5\" (UniqueName: \"kubernetes.io/projected/77e7d212-07ac-4c29-acb2-4e83bcf6d7e0-kube-api-access-5psb5\") pod \"glance-db-create-fcjt9\" (UID: \"77e7d212-07ac-4c29-acb2-4e83bcf6d7e0\") " pod="glance-kuttl-tests/glance-db-create-fcjt9" Dec 01 09:03:42 crc kubenswrapper[4813]: I1201 09:03:42.669689 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5psb5\" (UniqueName: \"kubernetes.io/projected/77e7d212-07ac-4c29-acb2-4e83bcf6d7e0-kube-api-access-5psb5\") pod \"glance-db-create-fcjt9\" (UID: \"77e7d212-07ac-4c29-acb2-4e83bcf6d7e0\") " pod="glance-kuttl-tests/glance-db-create-fcjt9" Dec 01 09:03:42 crc kubenswrapper[4813]: I1201 09:03:42.670228 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/77e7d212-07ac-4c29-acb2-4e83bcf6d7e0-operator-scripts\") pod \"glance-db-create-fcjt9\" (UID: \"77e7d212-07ac-4c29-acb2-4e83bcf6d7e0\") " pod="glance-kuttl-tests/glance-db-create-fcjt9" Dec 01 09:03:42 crc kubenswrapper[4813]: I1201 09:03:42.670925 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/77e7d212-07ac-4c29-acb2-4e83bcf6d7e0-operator-scripts\") pod \"glance-db-create-fcjt9\" (UID: \"77e7d212-07ac-4c29-acb2-4e83bcf6d7e0\") " pod="glance-kuttl-tests/glance-db-create-fcjt9" Dec 01 09:03:42 crc kubenswrapper[4813]: I1201 09:03:42.671073 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p7fdz\" (UniqueName: \"kubernetes.io/projected/3de75509-b8b6-4b97-b08e-77057f86ea77-kube-api-access-p7fdz\") pod \"glance-a9c4-account-create-update-4g8nj\" (UID: \"3de75509-b8b6-4b97-b08e-77057f86ea77\") " pod="glance-kuttl-tests/glance-a9c4-account-create-update-4g8nj" Dec 01 09:03:42 crc kubenswrapper[4813]: I1201 09:03:42.671101 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3de75509-b8b6-4b97-b08e-77057f86ea77-operator-scripts\") pod \"glance-a9c4-account-create-update-4g8nj\" (UID: \"3de75509-b8b6-4b97-b08e-77057f86ea77\") " pod="glance-kuttl-tests/glance-a9c4-account-create-update-4g8nj" Dec 01 09:03:42 crc kubenswrapper[4813]: I1201 09:03:42.698048 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5psb5\" (UniqueName: \"kubernetes.io/projected/77e7d212-07ac-4c29-acb2-4e83bcf6d7e0-kube-api-access-5psb5\") pod \"glance-db-create-fcjt9\" (UID: \"77e7d212-07ac-4c29-acb2-4e83bcf6d7e0\") " pod="glance-kuttl-tests/glance-db-create-fcjt9" Dec 01 09:03:42 crc kubenswrapper[4813]: I1201 09:03:42.726669 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-fcjt9" Dec 01 09:03:42 crc kubenswrapper[4813]: I1201 09:03:42.772460 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p7fdz\" (UniqueName: \"kubernetes.io/projected/3de75509-b8b6-4b97-b08e-77057f86ea77-kube-api-access-p7fdz\") pod \"glance-a9c4-account-create-update-4g8nj\" (UID: \"3de75509-b8b6-4b97-b08e-77057f86ea77\") " pod="glance-kuttl-tests/glance-a9c4-account-create-update-4g8nj" Dec 01 09:03:42 crc kubenswrapper[4813]: I1201 09:03:42.772553 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3de75509-b8b6-4b97-b08e-77057f86ea77-operator-scripts\") pod \"glance-a9c4-account-create-update-4g8nj\" (UID: \"3de75509-b8b6-4b97-b08e-77057f86ea77\") " pod="glance-kuttl-tests/glance-a9c4-account-create-update-4g8nj" Dec 01 09:03:42 crc kubenswrapper[4813]: I1201 09:03:42.773813 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3de75509-b8b6-4b97-b08e-77057f86ea77-operator-scripts\") pod \"glance-a9c4-account-create-update-4g8nj\" (UID: \"3de75509-b8b6-4b97-b08e-77057f86ea77\") " pod="glance-kuttl-tests/glance-a9c4-account-create-update-4g8nj" Dec 01 09:03:42 crc kubenswrapper[4813]: I1201 09:03:42.797929 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p7fdz\" (UniqueName: \"kubernetes.io/projected/3de75509-b8b6-4b97-b08e-77057f86ea77-kube-api-access-p7fdz\") pod \"glance-a9c4-account-create-update-4g8nj\" (UID: \"3de75509-b8b6-4b97-b08e-77057f86ea77\") " pod="glance-kuttl-tests/glance-a9c4-account-create-update-4g8nj" Dec 01 09:03:42 crc kubenswrapper[4813]: I1201 09:03:42.813396 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-a9c4-account-create-update-4g8nj" Dec 01 09:03:42 crc kubenswrapper[4813]: I1201 09:03:42.975481 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-create-fcjt9"] Dec 01 09:03:43 crc kubenswrapper[4813]: I1201 09:03:43.064997 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-create-fcjt9" event={"ID":"77e7d212-07ac-4c29-acb2-4e83bcf6d7e0","Type":"ContainerStarted","Data":"173fcde8444f0a27bef2ec5e2aa285f25d75f64f1b413819b5715931a38dd281"} Dec 01 09:03:43 crc kubenswrapper[4813]: I1201 09:03:43.272180 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-a9c4-account-create-update-4g8nj"] Dec 01 09:03:43 crc kubenswrapper[4813]: W1201 09:03:43.275329 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3de75509_b8b6_4b97_b08e_77057f86ea77.slice/crio-b8fd02416ad672618072daad45cc7cc7b6f4dbf231cdf27ed2dd719e162a9703 WatchSource:0}: Error finding container b8fd02416ad672618072daad45cc7cc7b6f4dbf231cdf27ed2dd719e162a9703: Status 404 returned error can't find the container with id b8fd02416ad672618072daad45cc7cc7b6f4dbf231cdf27ed2dd719e162a9703 Dec 01 09:03:44 crc kubenswrapper[4813]: I1201 09:03:44.075721 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-a9c4-account-create-update-4g8nj" event={"ID":"3de75509-b8b6-4b97-b08e-77057f86ea77","Type":"ContainerStarted","Data":"158ce31cbd9e2c3e30d880e6f28ca453fb84f4568471c7d8b694016d66c31583"} Dec 01 09:03:44 crc kubenswrapper[4813]: I1201 09:03:44.075792 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-a9c4-account-create-update-4g8nj" event={"ID":"3de75509-b8b6-4b97-b08e-77057f86ea77","Type":"ContainerStarted","Data":"b8fd02416ad672618072daad45cc7cc7b6f4dbf231cdf27ed2dd719e162a9703"} Dec 01 09:03:44 crc kubenswrapper[4813]: I1201 09:03:44.079259 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-create-fcjt9" event={"ID":"77e7d212-07ac-4c29-acb2-4e83bcf6d7e0","Type":"ContainerStarted","Data":"e560ea1af90413a541ed0e779818533d49870856d37b3a88dfb8b16e0583f993"} Dec 01 09:03:44 crc kubenswrapper[4813]: I1201 09:03:44.093756 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-a9c4-account-create-update-4g8nj" podStartSLOduration=2.093546318 podStartE2EDuration="2.093546318s" podCreationTimestamp="2025-12-01 09:03:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 09:03:44.091349166 +0000 UTC m=+1408.714270762" watchObservedRunningTime="2025-12-01 09:03:44.093546318 +0000 UTC m=+1408.716467914" Dec 01 09:03:45 crc kubenswrapper[4813]: I1201 09:03:45.089285 4813 generic.go:334] "Generic (PLEG): container finished" podID="3de75509-b8b6-4b97-b08e-77057f86ea77" containerID="158ce31cbd9e2c3e30d880e6f28ca453fb84f4568471c7d8b694016d66c31583" exitCode=0 Dec 01 09:03:45 crc kubenswrapper[4813]: I1201 09:03:45.089657 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-a9c4-account-create-update-4g8nj" event={"ID":"3de75509-b8b6-4b97-b08e-77057f86ea77","Type":"ContainerDied","Data":"158ce31cbd9e2c3e30d880e6f28ca453fb84f4568471c7d8b694016d66c31583"} Dec 01 09:03:45 crc kubenswrapper[4813]: I1201 09:03:45.092253 4813 generic.go:334] "Generic (PLEG): container finished" podID="77e7d212-07ac-4c29-acb2-4e83bcf6d7e0" containerID="e560ea1af90413a541ed0e779818533d49870856d37b3a88dfb8b16e0583f993" exitCode=0 Dec 01 09:03:45 crc kubenswrapper[4813]: I1201 09:03:45.092288 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-create-fcjt9" event={"ID":"77e7d212-07ac-4c29-acb2-4e83bcf6d7e0","Type":"ContainerDied","Data":"e560ea1af90413a541ed0e779818533d49870856d37b3a88dfb8b16e0583f993"} Dec 01 09:03:45 crc kubenswrapper[4813]: I1201 09:03:45.115787 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-db-create-fcjt9" podStartSLOduration=3.115755143 podStartE2EDuration="3.115755143s" podCreationTimestamp="2025-12-01 09:03:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 09:03:44.119602534 +0000 UTC m=+1408.742524140" watchObservedRunningTime="2025-12-01 09:03:45.115755143 +0000 UTC m=+1409.738676739" Dec 01 09:03:46 crc kubenswrapper[4813]: I1201 09:03:46.476398 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-fcjt9" Dec 01 09:03:46 crc kubenswrapper[4813]: I1201 09:03:46.487603 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-a9c4-account-create-update-4g8nj" Dec 01 09:03:46 crc kubenswrapper[4813]: I1201 09:03:46.493333 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3de75509-b8b6-4b97-b08e-77057f86ea77-operator-scripts\") pod \"3de75509-b8b6-4b97-b08e-77057f86ea77\" (UID: \"3de75509-b8b6-4b97-b08e-77057f86ea77\") " Dec 01 09:03:46 crc kubenswrapper[4813]: I1201 09:03:46.493475 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/77e7d212-07ac-4c29-acb2-4e83bcf6d7e0-operator-scripts\") pod \"77e7d212-07ac-4c29-acb2-4e83bcf6d7e0\" (UID: \"77e7d212-07ac-4c29-acb2-4e83bcf6d7e0\") " Dec 01 09:03:46 crc kubenswrapper[4813]: I1201 09:03:46.493552 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5psb5\" (UniqueName: \"kubernetes.io/projected/77e7d212-07ac-4c29-acb2-4e83bcf6d7e0-kube-api-access-5psb5\") pod \"77e7d212-07ac-4c29-acb2-4e83bcf6d7e0\" (UID: \"77e7d212-07ac-4c29-acb2-4e83bcf6d7e0\") " Dec 01 09:03:46 crc kubenswrapper[4813]: I1201 09:03:46.493600 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p7fdz\" (UniqueName: \"kubernetes.io/projected/3de75509-b8b6-4b97-b08e-77057f86ea77-kube-api-access-p7fdz\") pod \"3de75509-b8b6-4b97-b08e-77057f86ea77\" (UID: \"3de75509-b8b6-4b97-b08e-77057f86ea77\") " Dec 01 09:03:46 crc kubenswrapper[4813]: I1201 09:03:46.494557 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/77e7d212-07ac-4c29-acb2-4e83bcf6d7e0-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "77e7d212-07ac-4c29-acb2-4e83bcf6d7e0" (UID: "77e7d212-07ac-4c29-acb2-4e83bcf6d7e0"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 09:03:46 crc kubenswrapper[4813]: I1201 09:03:46.501057 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3de75509-b8b6-4b97-b08e-77057f86ea77-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "3de75509-b8b6-4b97-b08e-77057f86ea77" (UID: "3de75509-b8b6-4b97-b08e-77057f86ea77"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 09:03:46 crc kubenswrapper[4813]: I1201 09:03:46.503168 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3de75509-b8b6-4b97-b08e-77057f86ea77-kube-api-access-p7fdz" (OuterVolumeSpecName: "kube-api-access-p7fdz") pod "3de75509-b8b6-4b97-b08e-77057f86ea77" (UID: "3de75509-b8b6-4b97-b08e-77057f86ea77"). InnerVolumeSpecName "kube-api-access-p7fdz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:03:46 crc kubenswrapper[4813]: I1201 09:03:46.503943 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/77e7d212-07ac-4c29-acb2-4e83bcf6d7e0-kube-api-access-5psb5" (OuterVolumeSpecName: "kube-api-access-5psb5") pod "77e7d212-07ac-4c29-acb2-4e83bcf6d7e0" (UID: "77e7d212-07ac-4c29-acb2-4e83bcf6d7e0"). InnerVolumeSpecName "kube-api-access-5psb5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:03:46 crc kubenswrapper[4813]: I1201 09:03:46.595507 4813 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/77e7d212-07ac-4c29-acb2-4e83bcf6d7e0-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 09:03:46 crc kubenswrapper[4813]: I1201 09:03:46.595551 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5psb5\" (UniqueName: \"kubernetes.io/projected/77e7d212-07ac-4c29-acb2-4e83bcf6d7e0-kube-api-access-5psb5\") on node \"crc\" DevicePath \"\"" Dec 01 09:03:46 crc kubenswrapper[4813]: I1201 09:03:46.595567 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p7fdz\" (UniqueName: \"kubernetes.io/projected/3de75509-b8b6-4b97-b08e-77057f86ea77-kube-api-access-p7fdz\") on node \"crc\" DevicePath \"\"" Dec 01 09:03:46 crc kubenswrapper[4813]: I1201 09:03:46.595584 4813 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3de75509-b8b6-4b97-b08e-77057f86ea77-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 09:03:47 crc kubenswrapper[4813]: I1201 09:03:47.109280 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-create-fcjt9" event={"ID":"77e7d212-07ac-4c29-acb2-4e83bcf6d7e0","Type":"ContainerDied","Data":"173fcde8444f0a27bef2ec5e2aa285f25d75f64f1b413819b5715931a38dd281"} Dec 01 09:03:47 crc kubenswrapper[4813]: I1201 09:03:47.109318 4813 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="173fcde8444f0a27bef2ec5e2aa285f25d75f64f1b413819b5715931a38dd281" Dec 01 09:03:47 crc kubenswrapper[4813]: I1201 09:03:47.109399 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-fcjt9" Dec 01 09:03:47 crc kubenswrapper[4813]: I1201 09:03:47.113262 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-a9c4-account-create-update-4g8nj" event={"ID":"3de75509-b8b6-4b97-b08e-77057f86ea77","Type":"ContainerDied","Data":"b8fd02416ad672618072daad45cc7cc7b6f4dbf231cdf27ed2dd719e162a9703"} Dec 01 09:03:47 crc kubenswrapper[4813]: I1201 09:03:47.113317 4813 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b8fd02416ad672618072daad45cc7cc7b6f4dbf231cdf27ed2dd719e162a9703" Dec 01 09:03:47 crc kubenswrapper[4813]: I1201 09:03:47.113404 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-a9c4-account-create-update-4g8nj" Dec 01 09:03:47 crc kubenswrapper[4813]: I1201 09:03:47.768271 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-db-sync-78knl"] Dec 01 09:03:47 crc kubenswrapper[4813]: E1201 09:03:47.769588 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="77e7d212-07ac-4c29-acb2-4e83bcf6d7e0" containerName="mariadb-database-create" Dec 01 09:03:47 crc kubenswrapper[4813]: I1201 09:03:47.769688 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="77e7d212-07ac-4c29-acb2-4e83bcf6d7e0" containerName="mariadb-database-create" Dec 01 09:03:47 crc kubenswrapper[4813]: E1201 09:03:47.769818 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3de75509-b8b6-4b97-b08e-77057f86ea77" containerName="mariadb-account-create-update" Dec 01 09:03:47 crc kubenswrapper[4813]: I1201 09:03:47.769901 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="3de75509-b8b6-4b97-b08e-77057f86ea77" containerName="mariadb-account-create-update" Dec 01 09:03:47 crc kubenswrapper[4813]: I1201 09:03:47.770180 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="3de75509-b8b6-4b97-b08e-77057f86ea77" containerName="mariadb-account-create-update" Dec 01 09:03:47 crc kubenswrapper[4813]: I1201 09:03:47.770311 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="77e7d212-07ac-4c29-acb2-4e83bcf6d7e0" containerName="mariadb-database-create" Dec 01 09:03:47 crc kubenswrapper[4813]: I1201 09:03:47.771016 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-78knl" Dec 01 09:03:47 crc kubenswrapper[4813]: I1201 09:03:47.773231 4813 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-config-data" Dec 01 09:03:47 crc kubenswrapper[4813]: I1201 09:03:47.773323 4813 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-glance-dockercfg-xxf4v" Dec 01 09:03:47 crc kubenswrapper[4813]: I1201 09:03:47.781903 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-sync-78knl"] Dec 01 09:03:47 crc kubenswrapper[4813]: I1201 09:03:47.827786 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dgwhg\" (UniqueName: \"kubernetes.io/projected/40ca776e-7b84-4e84-b976-521ae258cf0e-kube-api-access-dgwhg\") pod \"glance-db-sync-78knl\" (UID: \"40ca776e-7b84-4e84-b976-521ae258cf0e\") " pod="glance-kuttl-tests/glance-db-sync-78knl" Dec 01 09:03:47 crc kubenswrapper[4813]: I1201 09:03:47.827856 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/40ca776e-7b84-4e84-b976-521ae258cf0e-db-sync-config-data\") pod \"glance-db-sync-78knl\" (UID: \"40ca776e-7b84-4e84-b976-521ae258cf0e\") " pod="glance-kuttl-tests/glance-db-sync-78knl" Dec 01 09:03:47 crc kubenswrapper[4813]: I1201 09:03:47.828023 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/40ca776e-7b84-4e84-b976-521ae258cf0e-config-data\") pod \"glance-db-sync-78knl\" (UID: \"40ca776e-7b84-4e84-b976-521ae258cf0e\") " pod="glance-kuttl-tests/glance-db-sync-78knl" Dec 01 09:03:47 crc kubenswrapper[4813]: I1201 09:03:47.929570 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dgwhg\" (UniqueName: \"kubernetes.io/projected/40ca776e-7b84-4e84-b976-521ae258cf0e-kube-api-access-dgwhg\") pod \"glance-db-sync-78knl\" (UID: \"40ca776e-7b84-4e84-b976-521ae258cf0e\") " pod="glance-kuttl-tests/glance-db-sync-78knl" Dec 01 09:03:47 crc kubenswrapper[4813]: I1201 09:03:47.929673 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/40ca776e-7b84-4e84-b976-521ae258cf0e-db-sync-config-data\") pod \"glance-db-sync-78knl\" (UID: \"40ca776e-7b84-4e84-b976-521ae258cf0e\") " pod="glance-kuttl-tests/glance-db-sync-78knl" Dec 01 09:03:47 crc kubenswrapper[4813]: I1201 09:03:47.929714 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/40ca776e-7b84-4e84-b976-521ae258cf0e-config-data\") pod \"glance-db-sync-78knl\" (UID: \"40ca776e-7b84-4e84-b976-521ae258cf0e\") " pod="glance-kuttl-tests/glance-db-sync-78knl" Dec 01 09:03:47 crc kubenswrapper[4813]: I1201 09:03:47.939843 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/40ca776e-7b84-4e84-b976-521ae258cf0e-db-sync-config-data\") pod \"glance-db-sync-78knl\" (UID: \"40ca776e-7b84-4e84-b976-521ae258cf0e\") " pod="glance-kuttl-tests/glance-db-sync-78knl" Dec 01 09:03:47 crc kubenswrapper[4813]: I1201 09:03:47.939992 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/40ca776e-7b84-4e84-b976-521ae258cf0e-config-data\") pod \"glance-db-sync-78knl\" (UID: \"40ca776e-7b84-4e84-b976-521ae258cf0e\") " pod="glance-kuttl-tests/glance-db-sync-78knl" Dec 01 09:03:47 crc kubenswrapper[4813]: I1201 09:03:47.946728 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dgwhg\" (UniqueName: \"kubernetes.io/projected/40ca776e-7b84-4e84-b976-521ae258cf0e-kube-api-access-dgwhg\") pod \"glance-db-sync-78knl\" (UID: \"40ca776e-7b84-4e84-b976-521ae258cf0e\") " pod="glance-kuttl-tests/glance-db-sync-78knl" Dec 01 09:03:48 crc kubenswrapper[4813]: I1201 09:03:48.085844 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-78knl" Dec 01 09:03:48 crc kubenswrapper[4813]: I1201 09:03:48.668174 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-sync-78knl"] Dec 01 09:03:49 crc kubenswrapper[4813]: I1201 09:03:49.137167 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-78knl" event={"ID":"40ca776e-7b84-4e84-b976-521ae258cf0e","Type":"ContainerStarted","Data":"93220d87896d1edb4e628fdb2ef73824efffd8adf819563d1cda4afa52202ead"} Dec 01 09:03:50 crc kubenswrapper[4813]: I1201 09:03:50.145353 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-78knl" event={"ID":"40ca776e-7b84-4e84-b976-521ae258cf0e","Type":"ContainerStarted","Data":"19c723c9091ce35ec9683cf2c52fb493beafe751cf1fe8498cf83f234b727873"} Dec 01 09:03:50 crc kubenswrapper[4813]: I1201 09:03:50.164856 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-db-sync-78knl" podStartSLOduration=3.164833068 podStartE2EDuration="3.164833068s" podCreationTimestamp="2025-12-01 09:03:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 09:03:50.162930325 +0000 UTC m=+1414.785851921" watchObservedRunningTime="2025-12-01 09:03:50.164833068 +0000 UTC m=+1414.787754654" Dec 01 09:03:53 crc kubenswrapper[4813]: I1201 09:03:53.192414 4813 generic.go:334] "Generic (PLEG): container finished" podID="40ca776e-7b84-4e84-b976-521ae258cf0e" containerID="19c723c9091ce35ec9683cf2c52fb493beafe751cf1fe8498cf83f234b727873" exitCode=0 Dec 01 09:03:53 crc kubenswrapper[4813]: I1201 09:03:53.192488 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-78knl" event={"ID":"40ca776e-7b84-4e84-b976-521ae258cf0e","Type":"ContainerDied","Data":"19c723c9091ce35ec9683cf2c52fb493beafe751cf1fe8498cf83f234b727873"} Dec 01 09:03:54 crc kubenswrapper[4813]: I1201 09:03:54.564424 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-78knl" Dec 01 09:03:54 crc kubenswrapper[4813]: I1201 09:03:54.720600 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/40ca776e-7b84-4e84-b976-521ae258cf0e-config-data\") pod \"40ca776e-7b84-4e84-b976-521ae258cf0e\" (UID: \"40ca776e-7b84-4e84-b976-521ae258cf0e\") " Dec 01 09:03:54 crc kubenswrapper[4813]: I1201 09:03:54.720694 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/40ca776e-7b84-4e84-b976-521ae258cf0e-db-sync-config-data\") pod \"40ca776e-7b84-4e84-b976-521ae258cf0e\" (UID: \"40ca776e-7b84-4e84-b976-521ae258cf0e\") " Dec 01 09:03:54 crc kubenswrapper[4813]: I1201 09:03:54.720798 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dgwhg\" (UniqueName: \"kubernetes.io/projected/40ca776e-7b84-4e84-b976-521ae258cf0e-kube-api-access-dgwhg\") pod \"40ca776e-7b84-4e84-b976-521ae258cf0e\" (UID: \"40ca776e-7b84-4e84-b976-521ae258cf0e\") " Dec 01 09:03:54 crc kubenswrapper[4813]: I1201 09:03:54.727183 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/40ca776e-7b84-4e84-b976-521ae258cf0e-kube-api-access-dgwhg" (OuterVolumeSpecName: "kube-api-access-dgwhg") pod "40ca776e-7b84-4e84-b976-521ae258cf0e" (UID: "40ca776e-7b84-4e84-b976-521ae258cf0e"). InnerVolumeSpecName "kube-api-access-dgwhg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:03:54 crc kubenswrapper[4813]: I1201 09:03:54.727616 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/40ca776e-7b84-4e84-b976-521ae258cf0e-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "40ca776e-7b84-4e84-b976-521ae258cf0e" (UID: "40ca776e-7b84-4e84-b976-521ae258cf0e"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:03:54 crc kubenswrapper[4813]: I1201 09:03:54.776250 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/40ca776e-7b84-4e84-b976-521ae258cf0e-config-data" (OuterVolumeSpecName: "config-data") pod "40ca776e-7b84-4e84-b976-521ae258cf0e" (UID: "40ca776e-7b84-4e84-b976-521ae258cf0e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:03:54 crc kubenswrapper[4813]: I1201 09:03:54.822520 4813 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/40ca776e-7b84-4e84-b976-521ae258cf0e-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 09:03:54 crc kubenswrapper[4813]: I1201 09:03:54.822568 4813 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/40ca776e-7b84-4e84-b976-521ae258cf0e-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 09:03:54 crc kubenswrapper[4813]: I1201 09:03:54.822585 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dgwhg\" (UniqueName: \"kubernetes.io/projected/40ca776e-7b84-4e84-b976-521ae258cf0e-kube-api-access-dgwhg\") on node \"crc\" DevicePath \"\"" Dec 01 09:03:55 crc kubenswrapper[4813]: I1201 09:03:55.219262 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-78knl" event={"ID":"40ca776e-7b84-4e84-b976-521ae258cf0e","Type":"ContainerDied","Data":"93220d87896d1edb4e628fdb2ef73824efffd8adf819563d1cda4afa52202ead"} Dec 01 09:03:55 crc kubenswrapper[4813]: I1201 09:03:55.219567 4813 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="93220d87896d1edb4e628fdb2ef73824efffd8adf819563d1cda4afa52202ead" Dec 01 09:03:55 crc kubenswrapper[4813]: I1201 09:03:55.219621 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-78knl" Dec 01 09:03:56 crc kubenswrapper[4813]: I1201 09:03:56.301362 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-external-api-1"] Dec 01 09:03:56 crc kubenswrapper[4813]: E1201 09:03:56.301648 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="40ca776e-7b84-4e84-b976-521ae258cf0e" containerName="glance-db-sync" Dec 01 09:03:56 crc kubenswrapper[4813]: I1201 09:03:56.301660 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="40ca776e-7b84-4e84-b976-521ae258cf0e" containerName="glance-db-sync" Dec 01 09:03:56 crc kubenswrapper[4813]: I1201 09:03:56.301803 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="40ca776e-7b84-4e84-b976-521ae258cf0e" containerName="glance-db-sync" Dec 01 09:03:56 crc kubenswrapper[4813]: I1201 09:03:56.302489 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-1" Dec 01 09:03:56 crc kubenswrapper[4813]: I1201 09:03:56.304575 4813 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-scripts" Dec 01 09:03:56 crc kubenswrapper[4813]: I1201 09:03:56.304660 4813 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-default-external-config-data" Dec 01 09:03:56 crc kubenswrapper[4813]: I1201 09:03:56.311524 4813 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-glance-dockercfg-xxf4v" Dec 01 09:03:56 crc kubenswrapper[4813]: I1201 09:03:56.382144 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-1"] Dec 01 09:03:56 crc kubenswrapper[4813]: I1201 09:03:56.474891 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/b4c39a03-2938-4b10-b430-369066c79ce4-run\") pod \"glance-default-external-api-1\" (UID: \"b4c39a03-2938-4b10-b430-369066c79ce4\") " pod="glance-kuttl-tests/glance-default-external-api-1" Dec 01 09:03:56 crc kubenswrapper[4813]: I1201 09:03:56.474978 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/b4c39a03-2938-4b10-b430-369066c79ce4-sys\") pod \"glance-default-external-api-1\" (UID: \"b4c39a03-2938-4b10-b430-369066c79ce4\") " pod="glance-kuttl-tests/glance-default-external-api-1" Dec 01 09:03:56 crc kubenswrapper[4813]: I1201 09:03:56.475020 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b4c39a03-2938-4b10-b430-369066c79ce4-httpd-run\") pod \"glance-default-external-api-1\" (UID: \"b4c39a03-2938-4b10-b430-369066c79ce4\") " pod="glance-kuttl-tests/glance-default-external-api-1" Dec 01 09:03:56 crc kubenswrapper[4813]: I1201 09:03:56.475113 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b4c39a03-2938-4b10-b430-369066c79ce4-scripts\") pod \"glance-default-external-api-1\" (UID: \"b4c39a03-2938-4b10-b430-369066c79ce4\") " pod="glance-kuttl-tests/glance-default-external-api-1" Dec 01 09:03:56 crc kubenswrapper[4813]: I1201 09:03:56.475180 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sgn7t\" (UniqueName: \"kubernetes.io/projected/b4c39a03-2938-4b10-b430-369066c79ce4-kube-api-access-sgn7t\") pod \"glance-default-external-api-1\" (UID: \"b4c39a03-2938-4b10-b430-369066c79ce4\") " pod="glance-kuttl-tests/glance-default-external-api-1" Dec 01 09:03:56 crc kubenswrapper[4813]: I1201 09:03:56.475239 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-external-api-1\" (UID: \"b4c39a03-2938-4b10-b430-369066c79ce4\") " pod="glance-kuttl-tests/glance-default-external-api-1" Dec 01 09:03:56 crc kubenswrapper[4813]: I1201 09:03:56.475279 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/b4c39a03-2938-4b10-b430-369066c79ce4-etc-nvme\") pod \"glance-default-external-api-1\" (UID: \"b4c39a03-2938-4b10-b430-369066c79ce4\") " pod="glance-kuttl-tests/glance-default-external-api-1" Dec 01 09:03:56 crc kubenswrapper[4813]: I1201 09:03:56.475297 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage14-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage14-crc\") pod \"glance-default-external-api-1\" (UID: \"b4c39a03-2938-4b10-b430-369066c79ce4\") " pod="glance-kuttl-tests/glance-default-external-api-1" Dec 01 09:03:56 crc kubenswrapper[4813]: I1201 09:03:56.475316 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/b4c39a03-2938-4b10-b430-369066c79ce4-lib-modules\") pod \"glance-default-external-api-1\" (UID: \"b4c39a03-2938-4b10-b430-369066c79ce4\") " pod="glance-kuttl-tests/glance-default-external-api-1" Dec 01 09:03:56 crc kubenswrapper[4813]: I1201 09:03:56.475341 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/b4c39a03-2938-4b10-b430-369066c79ce4-var-locks-brick\") pod \"glance-default-external-api-1\" (UID: \"b4c39a03-2938-4b10-b430-369066c79ce4\") " pod="glance-kuttl-tests/glance-default-external-api-1" Dec 01 09:03:56 crc kubenswrapper[4813]: I1201 09:03:56.475386 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b4c39a03-2938-4b10-b430-369066c79ce4-logs\") pod \"glance-default-external-api-1\" (UID: \"b4c39a03-2938-4b10-b430-369066c79ce4\") " pod="glance-kuttl-tests/glance-default-external-api-1" Dec 01 09:03:56 crc kubenswrapper[4813]: I1201 09:03:56.475452 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/b4c39a03-2938-4b10-b430-369066c79ce4-etc-iscsi\") pod \"glance-default-external-api-1\" (UID: \"b4c39a03-2938-4b10-b430-369066c79ce4\") " pod="glance-kuttl-tests/glance-default-external-api-1" Dec 01 09:03:56 crc kubenswrapper[4813]: I1201 09:03:56.475528 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/b4c39a03-2938-4b10-b430-369066c79ce4-dev\") pod \"glance-default-external-api-1\" (UID: \"b4c39a03-2938-4b10-b430-369066c79ce4\") " pod="glance-kuttl-tests/glance-default-external-api-1" Dec 01 09:03:56 crc kubenswrapper[4813]: I1201 09:03:56.475557 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b4c39a03-2938-4b10-b430-369066c79ce4-config-data\") pod \"glance-default-external-api-1\" (UID: \"b4c39a03-2938-4b10-b430-369066c79ce4\") " pod="glance-kuttl-tests/glance-default-external-api-1" Dec 01 09:03:56 crc kubenswrapper[4813]: I1201 09:03:56.522217 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-external-api-0"] Dec 01 09:03:56 crc kubenswrapper[4813]: I1201 09:03:56.523643 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-0" Dec 01 09:03:56 crc kubenswrapper[4813]: I1201 09:03:56.534219 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-0"] Dec 01 09:03:56 crc kubenswrapper[4813]: I1201 09:03:56.576953 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b4c39a03-2938-4b10-b430-369066c79ce4-logs\") pod \"glance-default-external-api-1\" (UID: \"b4c39a03-2938-4b10-b430-369066c79ce4\") " pod="glance-kuttl-tests/glance-default-external-api-1" Dec 01 09:03:56 crc kubenswrapper[4813]: I1201 09:03:56.577014 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/b4c39a03-2938-4b10-b430-369066c79ce4-etc-iscsi\") pod \"glance-default-external-api-1\" (UID: \"b4c39a03-2938-4b10-b430-369066c79ce4\") " pod="glance-kuttl-tests/glance-default-external-api-1" Dec 01 09:03:56 crc kubenswrapper[4813]: I1201 09:03:56.577053 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/b4c39a03-2938-4b10-b430-369066c79ce4-dev\") pod \"glance-default-external-api-1\" (UID: \"b4c39a03-2938-4b10-b430-369066c79ce4\") " pod="glance-kuttl-tests/glance-default-external-api-1" Dec 01 09:03:56 crc kubenswrapper[4813]: I1201 09:03:56.577088 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b4c39a03-2938-4b10-b430-369066c79ce4-config-data\") pod \"glance-default-external-api-1\" (UID: \"b4c39a03-2938-4b10-b430-369066c79ce4\") " pod="glance-kuttl-tests/glance-default-external-api-1" Dec 01 09:03:56 crc kubenswrapper[4813]: I1201 09:03:56.577114 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/b4c39a03-2938-4b10-b430-369066c79ce4-run\") pod \"glance-default-external-api-1\" (UID: \"b4c39a03-2938-4b10-b430-369066c79ce4\") " pod="glance-kuttl-tests/glance-default-external-api-1" Dec 01 09:03:56 crc kubenswrapper[4813]: I1201 09:03:56.577184 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/b4c39a03-2938-4b10-b430-369066c79ce4-sys\") pod \"glance-default-external-api-1\" (UID: \"b4c39a03-2938-4b10-b430-369066c79ce4\") " pod="glance-kuttl-tests/glance-default-external-api-1" Dec 01 09:03:56 crc kubenswrapper[4813]: I1201 09:03:56.577213 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b4c39a03-2938-4b10-b430-369066c79ce4-httpd-run\") pod \"glance-default-external-api-1\" (UID: \"b4c39a03-2938-4b10-b430-369066c79ce4\") " pod="glance-kuttl-tests/glance-default-external-api-1" Dec 01 09:03:56 crc kubenswrapper[4813]: I1201 09:03:56.577207 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/b4c39a03-2938-4b10-b430-369066c79ce4-etc-iscsi\") pod \"glance-default-external-api-1\" (UID: \"b4c39a03-2938-4b10-b430-369066c79ce4\") " pod="glance-kuttl-tests/glance-default-external-api-1" Dec 01 09:03:56 crc kubenswrapper[4813]: I1201 09:03:56.577232 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b4c39a03-2938-4b10-b430-369066c79ce4-scripts\") pod \"glance-default-external-api-1\" (UID: \"b4c39a03-2938-4b10-b430-369066c79ce4\") " pod="glance-kuttl-tests/glance-default-external-api-1" Dec 01 09:03:56 crc kubenswrapper[4813]: I1201 09:03:56.577305 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sgn7t\" (UniqueName: \"kubernetes.io/projected/b4c39a03-2938-4b10-b430-369066c79ce4-kube-api-access-sgn7t\") pod \"glance-default-external-api-1\" (UID: \"b4c39a03-2938-4b10-b430-369066c79ce4\") " pod="glance-kuttl-tests/glance-default-external-api-1" Dec 01 09:03:56 crc kubenswrapper[4813]: I1201 09:03:56.577346 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-external-api-1\" (UID: \"b4c39a03-2938-4b10-b430-369066c79ce4\") " pod="glance-kuttl-tests/glance-default-external-api-1" Dec 01 09:03:56 crc kubenswrapper[4813]: I1201 09:03:56.577403 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/b4c39a03-2938-4b10-b430-369066c79ce4-etc-nvme\") pod \"glance-default-external-api-1\" (UID: \"b4c39a03-2938-4b10-b430-369066c79ce4\") " pod="glance-kuttl-tests/glance-default-external-api-1" Dec 01 09:03:56 crc kubenswrapper[4813]: I1201 09:03:56.577432 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage14-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage14-crc\") pod \"glance-default-external-api-1\" (UID: \"b4c39a03-2938-4b10-b430-369066c79ce4\") " pod="glance-kuttl-tests/glance-default-external-api-1" Dec 01 09:03:56 crc kubenswrapper[4813]: I1201 09:03:56.577476 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b4c39a03-2938-4b10-b430-369066c79ce4-logs\") pod \"glance-default-external-api-1\" (UID: \"b4c39a03-2938-4b10-b430-369066c79ce4\") " pod="glance-kuttl-tests/glance-default-external-api-1" Dec 01 09:03:56 crc kubenswrapper[4813]: I1201 09:03:56.577512 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/b4c39a03-2938-4b10-b430-369066c79ce4-sys\") pod \"glance-default-external-api-1\" (UID: \"b4c39a03-2938-4b10-b430-369066c79ce4\") " pod="glance-kuttl-tests/glance-default-external-api-1" Dec 01 09:03:56 crc kubenswrapper[4813]: I1201 09:03:56.577225 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/b4c39a03-2938-4b10-b430-369066c79ce4-run\") pod \"glance-default-external-api-1\" (UID: \"b4c39a03-2938-4b10-b430-369066c79ce4\") " pod="glance-kuttl-tests/glance-default-external-api-1" Dec 01 09:03:56 crc kubenswrapper[4813]: I1201 09:03:56.577225 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/b4c39a03-2938-4b10-b430-369066c79ce4-dev\") pod \"glance-default-external-api-1\" (UID: \"b4c39a03-2938-4b10-b430-369066c79ce4\") " pod="glance-kuttl-tests/glance-default-external-api-1" Dec 01 09:03:56 crc kubenswrapper[4813]: I1201 09:03:56.578022 4813 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-external-api-1\" (UID: \"b4c39a03-2938-4b10-b430-369066c79ce4\") device mount path \"/mnt/openstack/pv04\"" pod="glance-kuttl-tests/glance-default-external-api-1" Dec 01 09:03:56 crc kubenswrapper[4813]: I1201 09:03:56.578197 4813 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage14-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage14-crc\") pod \"glance-default-external-api-1\" (UID: \"b4c39a03-2938-4b10-b430-369066c79ce4\") device mount path \"/mnt/openstack/pv14\"" pod="glance-kuttl-tests/glance-default-external-api-1" Dec 01 09:03:56 crc kubenswrapper[4813]: I1201 09:03:56.578240 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/b4c39a03-2938-4b10-b430-369066c79ce4-etc-nvme\") pod \"glance-default-external-api-1\" (UID: \"b4c39a03-2938-4b10-b430-369066c79ce4\") " pod="glance-kuttl-tests/glance-default-external-api-1" Dec 01 09:03:56 crc kubenswrapper[4813]: I1201 09:03:56.578329 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/b4c39a03-2938-4b10-b430-369066c79ce4-lib-modules\") pod \"glance-default-external-api-1\" (UID: \"b4c39a03-2938-4b10-b430-369066c79ce4\") " pod="glance-kuttl-tests/glance-default-external-api-1" Dec 01 09:03:56 crc kubenswrapper[4813]: I1201 09:03:56.578312 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/b4c39a03-2938-4b10-b430-369066c79ce4-lib-modules\") pod \"glance-default-external-api-1\" (UID: \"b4c39a03-2938-4b10-b430-369066c79ce4\") " pod="glance-kuttl-tests/glance-default-external-api-1" Dec 01 09:03:56 crc kubenswrapper[4813]: I1201 09:03:56.579021 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/b4c39a03-2938-4b10-b430-369066c79ce4-var-locks-brick\") pod \"glance-default-external-api-1\" (UID: \"b4c39a03-2938-4b10-b430-369066c79ce4\") " pod="glance-kuttl-tests/glance-default-external-api-1" Dec 01 09:03:56 crc kubenswrapper[4813]: I1201 09:03:56.579138 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/b4c39a03-2938-4b10-b430-369066c79ce4-var-locks-brick\") pod \"glance-default-external-api-1\" (UID: \"b4c39a03-2938-4b10-b430-369066c79ce4\") " pod="glance-kuttl-tests/glance-default-external-api-1" Dec 01 09:03:56 crc kubenswrapper[4813]: I1201 09:03:56.581114 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b4c39a03-2938-4b10-b430-369066c79ce4-httpd-run\") pod \"glance-default-external-api-1\" (UID: \"b4c39a03-2938-4b10-b430-369066c79ce4\") " pod="glance-kuttl-tests/glance-default-external-api-1" Dec 01 09:03:56 crc kubenswrapper[4813]: I1201 09:03:56.587825 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b4c39a03-2938-4b10-b430-369066c79ce4-scripts\") pod \"glance-default-external-api-1\" (UID: \"b4c39a03-2938-4b10-b430-369066c79ce4\") " pod="glance-kuttl-tests/glance-default-external-api-1" Dec 01 09:03:56 crc kubenswrapper[4813]: I1201 09:03:56.594148 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b4c39a03-2938-4b10-b430-369066c79ce4-config-data\") pod \"glance-default-external-api-1\" (UID: \"b4c39a03-2938-4b10-b430-369066c79ce4\") " pod="glance-kuttl-tests/glance-default-external-api-1" Dec 01 09:03:56 crc kubenswrapper[4813]: I1201 09:03:56.610758 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Dec 01 09:03:56 crc kubenswrapper[4813]: I1201 09:03:56.616290 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:03:56 crc kubenswrapper[4813]: I1201 09:03:56.611835 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage14-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage14-crc\") pod \"glance-default-external-api-1\" (UID: \"b4c39a03-2938-4b10-b430-369066c79ce4\") " pod="glance-kuttl-tests/glance-default-external-api-1" Dec 01 09:03:56 crc kubenswrapper[4813]: I1201 09:03:56.612944 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sgn7t\" (UniqueName: \"kubernetes.io/projected/b4c39a03-2938-4b10-b430-369066c79ce4-kube-api-access-sgn7t\") pod \"glance-default-external-api-1\" (UID: \"b4c39a03-2938-4b10-b430-369066c79ce4\") " pod="glance-kuttl-tests/glance-default-external-api-1" Dec 01 09:03:56 crc kubenswrapper[4813]: I1201 09:03:56.613358 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-external-api-1\" (UID: \"b4c39a03-2938-4b10-b430-369066c79ce4\") " pod="glance-kuttl-tests/glance-default-external-api-1" Dec 01 09:03:56 crc kubenswrapper[4813]: I1201 09:03:56.626677 4813 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-default-internal-config-data" Dec 01 09:03:56 crc kubenswrapper[4813]: I1201 09:03:56.630719 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Dec 01 09:03:56 crc kubenswrapper[4813]: I1201 09:03:56.634862 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-1"] Dec 01 09:03:56 crc kubenswrapper[4813]: I1201 09:03:56.636334 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-1" Dec 01 09:03:56 crc kubenswrapper[4813]: I1201 09:03:56.661689 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-1"] Dec 01 09:03:56 crc kubenswrapper[4813]: I1201 09:03:56.677691 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-1" Dec 01 09:03:56 crc kubenswrapper[4813]: I1201 09:03:56.684372 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/86363010-586b-4ee9-89e6-0549789e4f0a-dev\") pod \"glance-default-external-api-0\" (UID: \"86363010-586b-4ee9-89e6-0549789e4f0a\") " pod="glance-kuttl-tests/glance-default-external-api-0" Dec 01 09:03:56 crc kubenswrapper[4813]: I1201 09:03:56.684422 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/86363010-586b-4ee9-89e6-0549789e4f0a-lib-modules\") pod \"glance-default-external-api-0\" (UID: \"86363010-586b-4ee9-89e6-0549789e4f0a\") " pod="glance-kuttl-tests/glance-default-external-api-0" Dec 01 09:03:56 crc kubenswrapper[4813]: I1201 09:03:56.684439 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/86363010-586b-4ee9-89e6-0549789e4f0a-scripts\") pod \"glance-default-external-api-0\" (UID: \"86363010-586b-4ee9-89e6-0549789e4f0a\") " pod="glance-kuttl-tests/glance-default-external-api-0" Dec 01 09:03:56 crc kubenswrapper[4813]: I1201 09:03:56.684484 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/86363010-586b-4ee9-89e6-0549789e4f0a-etc-iscsi\") pod \"glance-default-external-api-0\" (UID: \"86363010-586b-4ee9-89e6-0549789e4f0a\") " pod="glance-kuttl-tests/glance-default-external-api-0" Dec 01 09:03:56 crc kubenswrapper[4813]: I1201 09:03:56.684503 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/86363010-586b-4ee9-89e6-0549789e4f0a-run\") pod \"glance-default-external-api-0\" (UID: \"86363010-586b-4ee9-89e6-0549789e4f0a\") " pod="glance-kuttl-tests/glance-default-external-api-0" Dec 01 09:03:56 crc kubenswrapper[4813]: I1201 09:03:56.684529 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/86363010-586b-4ee9-89e6-0549789e4f0a-var-locks-brick\") pod \"glance-default-external-api-0\" (UID: \"86363010-586b-4ee9-89e6-0549789e4f0a\") " pod="glance-kuttl-tests/glance-default-external-api-0" Dec 01 09:03:56 crc kubenswrapper[4813]: I1201 09:03:56.684546 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage16-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage16-crc\") pod \"glance-default-external-api-0\" (UID: \"86363010-586b-4ee9-89e6-0549789e4f0a\") " pod="glance-kuttl-tests/glance-default-external-api-0" Dec 01 09:03:56 crc kubenswrapper[4813]: I1201 09:03:56.684572 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage20-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage20-crc\") pod \"glance-default-external-api-0\" (UID: \"86363010-586b-4ee9-89e6-0549789e4f0a\") " pod="glance-kuttl-tests/glance-default-external-api-0" Dec 01 09:03:56 crc kubenswrapper[4813]: I1201 09:03:56.684591 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/86363010-586b-4ee9-89e6-0549789e4f0a-etc-nvme\") pod \"glance-default-external-api-0\" (UID: \"86363010-586b-4ee9-89e6-0549789e4f0a\") " pod="glance-kuttl-tests/glance-default-external-api-0" Dec 01 09:03:56 crc kubenswrapper[4813]: I1201 09:03:56.684614 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rd8wb\" (UniqueName: \"kubernetes.io/projected/86363010-586b-4ee9-89e6-0549789e4f0a-kube-api-access-rd8wb\") pod \"glance-default-external-api-0\" (UID: \"86363010-586b-4ee9-89e6-0549789e4f0a\") " pod="glance-kuttl-tests/glance-default-external-api-0" Dec 01 09:03:56 crc kubenswrapper[4813]: I1201 09:03:56.684649 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/86363010-586b-4ee9-89e6-0549789e4f0a-sys\") pod \"glance-default-external-api-0\" (UID: \"86363010-586b-4ee9-89e6-0549789e4f0a\") " pod="glance-kuttl-tests/glance-default-external-api-0" Dec 01 09:03:56 crc kubenswrapper[4813]: I1201 09:03:56.684671 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/86363010-586b-4ee9-89e6-0549789e4f0a-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"86363010-586b-4ee9-89e6-0549789e4f0a\") " pod="glance-kuttl-tests/glance-default-external-api-0" Dec 01 09:03:56 crc kubenswrapper[4813]: I1201 09:03:56.684691 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/86363010-586b-4ee9-89e6-0549789e4f0a-logs\") pod \"glance-default-external-api-0\" (UID: \"86363010-586b-4ee9-89e6-0549789e4f0a\") " pod="glance-kuttl-tests/glance-default-external-api-0" Dec 01 09:03:56 crc kubenswrapper[4813]: I1201 09:03:56.684708 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/86363010-586b-4ee9-89e6-0549789e4f0a-config-data\") pod \"glance-default-external-api-0\" (UID: \"86363010-586b-4ee9-89e6-0549789e4f0a\") " pod="glance-kuttl-tests/glance-default-external-api-0" Dec 01 09:03:56 crc kubenswrapper[4813]: I1201 09:03:56.791839 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-internal-api-0\" (UID: \"9402b27d-1b2c-4336-b2d7-e400772dc110\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:03:56 crc kubenswrapper[4813]: I1201 09:03:56.791886 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/83f41607-b074-4b35-b9d3-ec3199ae2ede-var-locks-brick\") pod \"glance-default-internal-api-1\" (UID: \"83f41607-b074-4b35-b9d3-ec3199ae2ede\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Dec 01 09:03:56 crc kubenswrapper[4813]: I1201 09:03:56.791915 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/86363010-586b-4ee9-89e6-0549789e4f0a-logs\") pod \"glance-default-external-api-0\" (UID: \"86363010-586b-4ee9-89e6-0549789e4f0a\") " pod="glance-kuttl-tests/glance-default-external-api-0" Dec 01 09:03:56 crc kubenswrapper[4813]: I1201 09:03:56.791944 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fwtdm\" (UniqueName: \"kubernetes.io/projected/83f41607-b074-4b35-b9d3-ec3199ae2ede-kube-api-access-fwtdm\") pod \"glance-default-internal-api-1\" (UID: \"83f41607-b074-4b35-b9d3-ec3199ae2ede\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Dec 01 09:03:56 crc kubenswrapper[4813]: I1201 09:03:56.791972 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/86363010-586b-4ee9-89e6-0549789e4f0a-config-data\") pod \"glance-default-external-api-0\" (UID: \"86363010-586b-4ee9-89e6-0549789e4f0a\") " pod="glance-kuttl-tests/glance-default-external-api-0" Dec 01 09:03:56 crc kubenswrapper[4813]: I1201 09:03:56.791989 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/9402b27d-1b2c-4336-b2d7-e400772dc110-run\") pod \"glance-default-internal-api-0\" (UID: \"9402b27d-1b2c-4336-b2d7-e400772dc110\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:03:56 crc kubenswrapper[4813]: I1201 09:03:56.792009 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/9402b27d-1b2c-4336-b2d7-e400772dc110-dev\") pod \"glance-default-internal-api-0\" (UID: \"9402b27d-1b2c-4336-b2d7-e400772dc110\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:03:56 crc kubenswrapper[4813]: I1201 09:03:56.792026 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/9402b27d-1b2c-4336-b2d7-e400772dc110-etc-nvme\") pod \"glance-default-internal-api-0\" (UID: \"9402b27d-1b2c-4336-b2d7-e400772dc110\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:03:56 crc kubenswrapper[4813]: I1201 09:03:56.792047 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/9402b27d-1b2c-4336-b2d7-e400772dc110-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"9402b27d-1b2c-4336-b2d7-e400772dc110\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:03:56 crc kubenswrapper[4813]: I1201 09:03:56.792064 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/86363010-586b-4ee9-89e6-0549789e4f0a-dev\") pod \"glance-default-external-api-0\" (UID: \"86363010-586b-4ee9-89e6-0549789e4f0a\") " pod="glance-kuttl-tests/glance-default-external-api-0" Dec 01 09:03:56 crc kubenswrapper[4813]: I1201 09:03:56.792084 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/9402b27d-1b2c-4336-b2d7-e400772dc110-sys\") pod \"glance-default-internal-api-0\" (UID: \"9402b27d-1b2c-4336-b2d7-e400772dc110\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:03:56 crc kubenswrapper[4813]: I1201 09:03:56.792106 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/86363010-586b-4ee9-89e6-0549789e4f0a-lib-modules\") pod \"glance-default-external-api-0\" (UID: \"86363010-586b-4ee9-89e6-0549789e4f0a\") " pod="glance-kuttl-tests/glance-default-external-api-0" Dec 01 09:03:56 crc kubenswrapper[4813]: I1201 09:03:56.792131 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/86363010-586b-4ee9-89e6-0549789e4f0a-scripts\") pod \"glance-default-external-api-0\" (UID: \"86363010-586b-4ee9-89e6-0549789e4f0a\") " pod="glance-kuttl-tests/glance-default-external-api-0" Dec 01 09:03:56 crc kubenswrapper[4813]: I1201 09:03:56.792160 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/83f41607-b074-4b35-b9d3-ec3199ae2ede-etc-iscsi\") pod \"glance-default-internal-api-1\" (UID: \"83f41607-b074-4b35-b9d3-ec3199ae2ede\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Dec 01 09:03:56 crc kubenswrapper[4813]: I1201 09:03:56.792179 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/83f41607-b074-4b35-b9d3-ec3199ae2ede-logs\") pod \"glance-default-internal-api-1\" (UID: \"83f41607-b074-4b35-b9d3-ec3199ae2ede\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Dec 01 09:03:56 crc kubenswrapper[4813]: I1201 09:03:56.792203 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/83f41607-b074-4b35-b9d3-ec3199ae2ede-run\") pod \"glance-default-internal-api-1\" (UID: \"83f41607-b074-4b35-b9d3-ec3199ae2ede\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Dec 01 09:03:56 crc kubenswrapper[4813]: I1201 09:03:56.792219 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/9402b27d-1b2c-4336-b2d7-e400772dc110-lib-modules\") pod \"glance-default-internal-api-0\" (UID: \"9402b27d-1b2c-4336-b2d7-e400772dc110\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:03:56 crc kubenswrapper[4813]: I1201 09:03:56.792286 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/86363010-586b-4ee9-89e6-0549789e4f0a-etc-iscsi\") pod \"glance-default-external-api-0\" (UID: \"86363010-586b-4ee9-89e6-0549789e4f0a\") " pod="glance-kuttl-tests/glance-default-external-api-0" Dec 01 09:03:56 crc kubenswrapper[4813]: I1201 09:03:56.792321 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/86363010-586b-4ee9-89e6-0549789e4f0a-lib-modules\") pod \"glance-default-external-api-0\" (UID: \"86363010-586b-4ee9-89e6-0549789e4f0a\") " pod="glance-kuttl-tests/glance-default-external-api-0" Dec 01 09:03:56 crc kubenswrapper[4813]: I1201 09:03:56.792350 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/83f41607-b074-4b35-b9d3-ec3199ae2ede-etc-nvme\") pod \"glance-default-internal-api-1\" (UID: \"83f41607-b074-4b35-b9d3-ec3199ae2ede\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Dec 01 09:03:56 crc kubenswrapper[4813]: I1201 09:03:56.792424 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/86363010-586b-4ee9-89e6-0549789e4f0a-run\") pod \"glance-default-external-api-0\" (UID: \"86363010-586b-4ee9-89e6-0549789e4f0a\") " pod="glance-kuttl-tests/glance-default-external-api-0" Dec 01 09:03:56 crc kubenswrapper[4813]: I1201 09:03:56.792473 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/83f41607-b074-4b35-b9d3-ec3199ae2ede-sys\") pod \"glance-default-internal-api-1\" (UID: \"83f41607-b074-4b35-b9d3-ec3199ae2ede\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Dec 01 09:03:56 crc kubenswrapper[4813]: I1201 09:03:56.792493 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/86363010-586b-4ee9-89e6-0549789e4f0a-logs\") pod \"glance-default-external-api-0\" (UID: \"86363010-586b-4ee9-89e6-0549789e4f0a\") " pod="glance-kuttl-tests/glance-default-external-api-0" Dec 01 09:03:56 crc kubenswrapper[4813]: I1201 09:03:56.792519 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/9402b27d-1b2c-4336-b2d7-e400772dc110-etc-iscsi\") pod \"glance-default-internal-api-0\" (UID: \"9402b27d-1b2c-4336-b2d7-e400772dc110\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:03:56 crc kubenswrapper[4813]: I1201 09:03:56.792558 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/83f41607-b074-4b35-b9d3-ec3199ae2ede-config-data\") pod \"glance-default-internal-api-1\" (UID: \"83f41607-b074-4b35-b9d3-ec3199ae2ede\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Dec 01 09:03:56 crc kubenswrapper[4813]: I1201 09:03:56.792585 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/83f41607-b074-4b35-b9d3-ec3199ae2ede-dev\") pod \"glance-default-internal-api-1\" (UID: \"83f41607-b074-4b35-b9d3-ec3199ae2ede\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Dec 01 09:03:56 crc kubenswrapper[4813]: I1201 09:03:56.792586 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/86363010-586b-4ee9-89e6-0549789e4f0a-etc-iscsi\") pod \"glance-default-external-api-0\" (UID: \"86363010-586b-4ee9-89e6-0549789e4f0a\") " pod="glance-kuttl-tests/glance-default-external-api-0" Dec 01 09:03:56 crc kubenswrapper[4813]: I1201 09:03:56.792300 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/86363010-586b-4ee9-89e6-0549789e4f0a-dev\") pod \"glance-default-external-api-0\" (UID: \"86363010-586b-4ee9-89e6-0549789e4f0a\") " pod="glance-kuttl-tests/glance-default-external-api-0" Dec 01 09:03:56 crc kubenswrapper[4813]: I1201 09:03:56.792652 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/86363010-586b-4ee9-89e6-0549789e4f0a-run\") pod \"glance-default-external-api-0\" (UID: \"86363010-586b-4ee9-89e6-0549789e4f0a\") " pod="glance-kuttl-tests/glance-default-external-api-0" Dec 01 09:03:56 crc kubenswrapper[4813]: I1201 09:03:56.792697 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/86363010-586b-4ee9-89e6-0549789e4f0a-var-locks-brick\") pod \"glance-default-external-api-0\" (UID: \"86363010-586b-4ee9-89e6-0549789e4f0a\") " pod="glance-kuttl-tests/glance-default-external-api-0" Dec 01 09:03:56 crc kubenswrapper[4813]: I1201 09:03:56.792785 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/86363010-586b-4ee9-89e6-0549789e4f0a-var-locks-brick\") pod \"glance-default-external-api-0\" (UID: \"86363010-586b-4ee9-89e6-0549789e4f0a\") " pod="glance-kuttl-tests/glance-default-external-api-0" Dec 01 09:03:56 crc kubenswrapper[4813]: I1201 09:03:56.792869 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage16-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage16-crc\") pod \"glance-default-external-api-0\" (UID: \"86363010-586b-4ee9-89e6-0549789e4f0a\") " pod="glance-kuttl-tests/glance-default-external-api-0" Dec 01 09:03:56 crc kubenswrapper[4813]: I1201 09:03:56.792905 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/83f41607-b074-4b35-b9d3-ec3199ae2ede-httpd-run\") pod \"glance-default-internal-api-1\" (UID: \"83f41607-b074-4b35-b9d3-ec3199ae2ede\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Dec 01 09:03:56 crc kubenswrapper[4813]: I1201 09:03:56.792955 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9402b27d-1b2c-4336-b2d7-e400772dc110-logs\") pod \"glance-default-internal-api-0\" (UID: \"9402b27d-1b2c-4336-b2d7-e400772dc110\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:03:56 crc kubenswrapper[4813]: I1201 09:03:56.793023 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage20-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage20-crc\") pod \"glance-default-external-api-0\" (UID: \"86363010-586b-4ee9-89e6-0549789e4f0a\") " pod="glance-kuttl-tests/glance-default-external-api-0" Dec 01 09:03:56 crc kubenswrapper[4813]: I1201 09:03:56.793053 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/9402b27d-1b2c-4336-b2d7-e400772dc110-var-locks-brick\") pod \"glance-default-internal-api-0\" (UID: \"9402b27d-1b2c-4336-b2d7-e400772dc110\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:03:56 crc kubenswrapper[4813]: I1201 09:03:56.793107 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/86363010-586b-4ee9-89e6-0549789e4f0a-etc-nvme\") pod \"glance-default-external-api-0\" (UID: \"86363010-586b-4ee9-89e6-0549789e4f0a\") " pod="glance-kuttl-tests/glance-default-external-api-0" Dec 01 09:03:56 crc kubenswrapper[4813]: I1201 09:03:56.793151 4813 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage16-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage16-crc\") pod \"glance-default-external-api-0\" (UID: \"86363010-586b-4ee9-89e6-0549789e4f0a\") device mount path \"/mnt/openstack/pv16\"" pod="glance-kuttl-tests/glance-default-external-api-0" Dec 01 09:03:56 crc kubenswrapper[4813]: I1201 09:03:56.793626 4813 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage20-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage20-crc\") pod \"glance-default-external-api-0\" (UID: \"86363010-586b-4ee9-89e6-0549789e4f0a\") device mount path \"/mnt/openstack/pv20\"" pod="glance-kuttl-tests/glance-default-external-api-0" Dec 01 09:03:56 crc kubenswrapper[4813]: I1201 09:03:56.793719 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/86363010-586b-4ee9-89e6-0549789e4f0a-etc-nvme\") pod \"glance-default-external-api-0\" (UID: \"86363010-586b-4ee9-89e6-0549789e4f0a\") " pod="glance-kuttl-tests/glance-default-external-api-0" Dec 01 09:03:56 crc kubenswrapper[4813]: I1201 09:03:56.793158 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-internal-api-1\" (UID: \"83f41607-b074-4b35-b9d3-ec3199ae2ede\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Dec 01 09:03:56 crc kubenswrapper[4813]: I1201 09:03:56.793794 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9402b27d-1b2c-4336-b2d7-e400772dc110-config-data\") pod \"glance-default-internal-api-0\" (UID: \"9402b27d-1b2c-4336-b2d7-e400772dc110\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:03:56 crc kubenswrapper[4813]: I1201 09:03:56.793831 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9402b27d-1b2c-4336-b2d7-e400772dc110-scripts\") pod \"glance-default-internal-api-0\" (UID: \"9402b27d-1b2c-4336-b2d7-e400772dc110\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:03:56 crc kubenswrapper[4813]: I1201 09:03:56.793860 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/83f41607-b074-4b35-b9d3-ec3199ae2ede-lib-modules\") pod \"glance-default-internal-api-1\" (UID: \"83f41607-b074-4b35-b9d3-ec3199ae2ede\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Dec 01 09:03:56 crc kubenswrapper[4813]: I1201 09:03:56.793885 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rd8wb\" (UniqueName: \"kubernetes.io/projected/86363010-586b-4ee9-89e6-0549789e4f0a-kube-api-access-rd8wb\") pod \"glance-default-external-api-0\" (UID: \"86363010-586b-4ee9-89e6-0549789e4f0a\") " pod="glance-kuttl-tests/glance-default-external-api-0" Dec 01 09:03:56 crc kubenswrapper[4813]: I1201 09:03:56.793915 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/86363010-586b-4ee9-89e6-0549789e4f0a-sys\") pod \"glance-default-external-api-0\" (UID: \"86363010-586b-4ee9-89e6-0549789e4f0a\") " pod="glance-kuttl-tests/glance-default-external-api-0" Dec 01 09:03:56 crc kubenswrapper[4813]: I1201 09:03:56.793941 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/83f41607-b074-4b35-b9d3-ec3199ae2ede-scripts\") pod \"glance-default-internal-api-1\" (UID: \"83f41607-b074-4b35-b9d3-ec3199ae2ede\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Dec 01 09:03:56 crc kubenswrapper[4813]: I1201 09:03:56.794008 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zgxvz\" (UniqueName: \"kubernetes.io/projected/9402b27d-1b2c-4336-b2d7-e400772dc110-kube-api-access-zgxvz\") pod \"glance-default-internal-api-0\" (UID: \"9402b27d-1b2c-4336-b2d7-e400772dc110\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:03:56 crc kubenswrapper[4813]: I1201 09:03:56.794043 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/86363010-586b-4ee9-89e6-0549789e4f0a-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"86363010-586b-4ee9-89e6-0549789e4f0a\") " pod="glance-kuttl-tests/glance-default-external-api-0" Dec 01 09:03:56 crc kubenswrapper[4813]: I1201 09:03:56.794067 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-internal-api-1\" (UID: \"83f41607-b074-4b35-b9d3-ec3199ae2ede\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Dec 01 09:03:56 crc kubenswrapper[4813]: I1201 09:03:56.794101 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-internal-api-0\" (UID: \"9402b27d-1b2c-4336-b2d7-e400772dc110\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:03:56 crc kubenswrapper[4813]: I1201 09:03:56.794574 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/86363010-586b-4ee9-89e6-0549789e4f0a-sys\") pod \"glance-default-external-api-0\" (UID: \"86363010-586b-4ee9-89e6-0549789e4f0a\") " pod="glance-kuttl-tests/glance-default-external-api-0" Dec 01 09:03:56 crc kubenswrapper[4813]: I1201 09:03:56.794957 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/86363010-586b-4ee9-89e6-0549789e4f0a-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"86363010-586b-4ee9-89e6-0549789e4f0a\") " pod="glance-kuttl-tests/glance-default-external-api-0" Dec 01 09:03:56 crc kubenswrapper[4813]: I1201 09:03:56.815510 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/86363010-586b-4ee9-89e6-0549789e4f0a-config-data\") pod \"glance-default-external-api-0\" (UID: \"86363010-586b-4ee9-89e6-0549789e4f0a\") " pod="glance-kuttl-tests/glance-default-external-api-0" Dec 01 09:03:56 crc kubenswrapper[4813]: I1201 09:03:56.833833 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/86363010-586b-4ee9-89e6-0549789e4f0a-scripts\") pod \"glance-default-external-api-0\" (UID: \"86363010-586b-4ee9-89e6-0549789e4f0a\") " pod="glance-kuttl-tests/glance-default-external-api-0" Dec 01 09:03:56 crc kubenswrapper[4813]: I1201 09:03:56.836308 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rd8wb\" (UniqueName: \"kubernetes.io/projected/86363010-586b-4ee9-89e6-0549789e4f0a-kube-api-access-rd8wb\") pod \"glance-default-external-api-0\" (UID: \"86363010-586b-4ee9-89e6-0549789e4f0a\") " pod="glance-kuttl-tests/glance-default-external-api-0" Dec 01 09:03:56 crc kubenswrapper[4813]: I1201 09:03:56.854985 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage16-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage16-crc\") pod \"glance-default-external-api-0\" (UID: \"86363010-586b-4ee9-89e6-0549789e4f0a\") " pod="glance-kuttl-tests/glance-default-external-api-0" Dec 01 09:03:56 crc kubenswrapper[4813]: I1201 09:03:56.868682 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage20-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage20-crc\") pod \"glance-default-external-api-0\" (UID: \"86363010-586b-4ee9-89e6-0549789e4f0a\") " pod="glance-kuttl-tests/glance-default-external-api-0" Dec 01 09:03:56 crc kubenswrapper[4813]: I1201 09:03:56.897619 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/83f41607-b074-4b35-b9d3-ec3199ae2ede-sys\") pod \"glance-default-internal-api-1\" (UID: \"83f41607-b074-4b35-b9d3-ec3199ae2ede\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Dec 01 09:03:56 crc kubenswrapper[4813]: I1201 09:03:56.897683 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/9402b27d-1b2c-4336-b2d7-e400772dc110-etc-iscsi\") pod \"glance-default-internal-api-0\" (UID: \"9402b27d-1b2c-4336-b2d7-e400772dc110\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:03:56 crc kubenswrapper[4813]: I1201 09:03:56.897710 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/83f41607-b074-4b35-b9d3-ec3199ae2ede-config-data\") pod \"glance-default-internal-api-1\" (UID: \"83f41607-b074-4b35-b9d3-ec3199ae2ede\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Dec 01 09:03:56 crc kubenswrapper[4813]: I1201 09:03:56.897735 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/83f41607-b074-4b35-b9d3-ec3199ae2ede-dev\") pod \"glance-default-internal-api-1\" (UID: \"83f41607-b074-4b35-b9d3-ec3199ae2ede\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Dec 01 09:03:56 crc kubenswrapper[4813]: I1201 09:03:56.897779 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/83f41607-b074-4b35-b9d3-ec3199ae2ede-httpd-run\") pod \"glance-default-internal-api-1\" (UID: \"83f41607-b074-4b35-b9d3-ec3199ae2ede\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Dec 01 09:03:56 crc kubenswrapper[4813]: I1201 09:03:56.897823 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9402b27d-1b2c-4336-b2d7-e400772dc110-logs\") pod \"glance-default-internal-api-0\" (UID: \"9402b27d-1b2c-4336-b2d7-e400772dc110\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:03:56 crc kubenswrapper[4813]: I1201 09:03:56.897883 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/9402b27d-1b2c-4336-b2d7-e400772dc110-var-locks-brick\") pod \"glance-default-internal-api-0\" (UID: \"9402b27d-1b2c-4336-b2d7-e400772dc110\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:03:56 crc kubenswrapper[4813]: I1201 09:03:56.897937 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-internal-api-1\" (UID: \"83f41607-b074-4b35-b9d3-ec3199ae2ede\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Dec 01 09:03:56 crc kubenswrapper[4813]: I1201 09:03:56.897979 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9402b27d-1b2c-4336-b2d7-e400772dc110-config-data\") pod \"glance-default-internal-api-0\" (UID: \"9402b27d-1b2c-4336-b2d7-e400772dc110\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:03:56 crc kubenswrapper[4813]: I1201 09:03:56.898005 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9402b27d-1b2c-4336-b2d7-e400772dc110-scripts\") pod \"glance-default-internal-api-0\" (UID: \"9402b27d-1b2c-4336-b2d7-e400772dc110\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:03:56 crc kubenswrapper[4813]: I1201 09:03:56.898028 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/83f41607-b074-4b35-b9d3-ec3199ae2ede-lib-modules\") pod \"glance-default-internal-api-1\" (UID: \"83f41607-b074-4b35-b9d3-ec3199ae2ede\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Dec 01 09:03:56 crc kubenswrapper[4813]: I1201 09:03:56.898057 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/83f41607-b074-4b35-b9d3-ec3199ae2ede-scripts\") pod \"glance-default-internal-api-1\" (UID: \"83f41607-b074-4b35-b9d3-ec3199ae2ede\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Dec 01 09:03:56 crc kubenswrapper[4813]: I1201 09:03:56.898097 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zgxvz\" (UniqueName: \"kubernetes.io/projected/9402b27d-1b2c-4336-b2d7-e400772dc110-kube-api-access-zgxvz\") pod \"glance-default-internal-api-0\" (UID: \"9402b27d-1b2c-4336-b2d7-e400772dc110\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:03:56 crc kubenswrapper[4813]: I1201 09:03:56.898868 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-internal-api-1\" (UID: \"83f41607-b074-4b35-b9d3-ec3199ae2ede\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Dec 01 09:03:56 crc kubenswrapper[4813]: I1201 09:03:56.898916 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-internal-api-0\" (UID: \"9402b27d-1b2c-4336-b2d7-e400772dc110\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:03:56 crc kubenswrapper[4813]: I1201 09:03:56.898946 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-internal-api-0\" (UID: \"9402b27d-1b2c-4336-b2d7-e400772dc110\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:03:56 crc kubenswrapper[4813]: I1201 09:03:56.898993 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/83f41607-b074-4b35-b9d3-ec3199ae2ede-var-locks-brick\") pod \"glance-default-internal-api-1\" (UID: \"83f41607-b074-4b35-b9d3-ec3199ae2ede\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Dec 01 09:03:56 crc kubenswrapper[4813]: I1201 09:03:56.899035 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fwtdm\" (UniqueName: \"kubernetes.io/projected/83f41607-b074-4b35-b9d3-ec3199ae2ede-kube-api-access-fwtdm\") pod \"glance-default-internal-api-1\" (UID: \"83f41607-b074-4b35-b9d3-ec3199ae2ede\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Dec 01 09:03:56 crc kubenswrapper[4813]: I1201 09:03:56.899065 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/9402b27d-1b2c-4336-b2d7-e400772dc110-run\") pod \"glance-default-internal-api-0\" (UID: \"9402b27d-1b2c-4336-b2d7-e400772dc110\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:03:56 crc kubenswrapper[4813]: I1201 09:03:56.899101 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/9402b27d-1b2c-4336-b2d7-e400772dc110-dev\") pod \"glance-default-internal-api-0\" (UID: \"9402b27d-1b2c-4336-b2d7-e400772dc110\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:03:56 crc kubenswrapper[4813]: I1201 09:03:56.899127 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/9402b27d-1b2c-4336-b2d7-e400772dc110-etc-nvme\") pod \"glance-default-internal-api-0\" (UID: \"9402b27d-1b2c-4336-b2d7-e400772dc110\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:03:56 crc kubenswrapper[4813]: I1201 09:03:56.899155 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/9402b27d-1b2c-4336-b2d7-e400772dc110-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"9402b27d-1b2c-4336-b2d7-e400772dc110\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:03:56 crc kubenswrapper[4813]: I1201 09:03:56.899188 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/9402b27d-1b2c-4336-b2d7-e400772dc110-sys\") pod \"glance-default-internal-api-0\" (UID: \"9402b27d-1b2c-4336-b2d7-e400772dc110\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:03:56 crc kubenswrapper[4813]: I1201 09:03:56.899271 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/83f41607-b074-4b35-b9d3-ec3199ae2ede-etc-iscsi\") pod \"glance-default-internal-api-1\" (UID: \"83f41607-b074-4b35-b9d3-ec3199ae2ede\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Dec 01 09:03:56 crc kubenswrapper[4813]: I1201 09:03:56.899301 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/83f41607-b074-4b35-b9d3-ec3199ae2ede-logs\") pod \"glance-default-internal-api-1\" (UID: \"83f41607-b074-4b35-b9d3-ec3199ae2ede\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Dec 01 09:03:56 crc kubenswrapper[4813]: I1201 09:03:56.899354 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/9402b27d-1b2c-4336-b2d7-e400772dc110-lib-modules\") pod \"glance-default-internal-api-0\" (UID: \"9402b27d-1b2c-4336-b2d7-e400772dc110\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:03:56 crc kubenswrapper[4813]: I1201 09:03:56.899422 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/83f41607-b074-4b35-b9d3-ec3199ae2ede-run\") pod \"glance-default-internal-api-1\" (UID: \"83f41607-b074-4b35-b9d3-ec3199ae2ede\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Dec 01 09:03:56 crc kubenswrapper[4813]: I1201 09:03:56.899456 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/83f41607-b074-4b35-b9d3-ec3199ae2ede-etc-nvme\") pod \"glance-default-internal-api-1\" (UID: \"83f41607-b074-4b35-b9d3-ec3199ae2ede\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Dec 01 09:03:56 crc kubenswrapper[4813]: I1201 09:03:56.899596 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/83f41607-b074-4b35-b9d3-ec3199ae2ede-etc-nvme\") pod \"glance-default-internal-api-1\" (UID: \"83f41607-b074-4b35-b9d3-ec3199ae2ede\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Dec 01 09:03:56 crc kubenswrapper[4813]: I1201 09:03:56.899650 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/83f41607-b074-4b35-b9d3-ec3199ae2ede-sys\") pod \"glance-default-internal-api-1\" (UID: \"83f41607-b074-4b35-b9d3-ec3199ae2ede\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Dec 01 09:03:56 crc kubenswrapper[4813]: I1201 09:03:56.899682 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/9402b27d-1b2c-4336-b2d7-e400772dc110-etc-iscsi\") pod \"glance-default-internal-api-0\" (UID: \"9402b27d-1b2c-4336-b2d7-e400772dc110\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:03:56 crc kubenswrapper[4813]: I1201 09:03:56.900771 4813 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-internal-api-0\" (UID: \"9402b27d-1b2c-4336-b2d7-e400772dc110\") device mount path \"/mnt/openstack/pv07\"" pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:03:56 crc kubenswrapper[4813]: I1201 09:03:56.903352 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/9402b27d-1b2c-4336-b2d7-e400772dc110-etc-nvme\") pod \"glance-default-internal-api-0\" (UID: \"9402b27d-1b2c-4336-b2d7-e400772dc110\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:03:56 crc kubenswrapper[4813]: I1201 09:03:56.903344 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/83f41607-b074-4b35-b9d3-ec3199ae2ede-etc-iscsi\") pod \"glance-default-internal-api-1\" (UID: \"83f41607-b074-4b35-b9d3-ec3199ae2ede\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Dec 01 09:03:56 crc kubenswrapper[4813]: I1201 09:03:56.903484 4813 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-internal-api-0\" (UID: \"9402b27d-1b2c-4336-b2d7-e400772dc110\") device mount path \"/mnt/openstack/pv03\"" pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:03:56 crc kubenswrapper[4813]: I1201 09:03:56.903737 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/9402b27d-1b2c-4336-b2d7-e400772dc110-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"9402b27d-1b2c-4336-b2d7-e400772dc110\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:03:56 crc kubenswrapper[4813]: I1201 09:03:56.903781 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/9402b27d-1b2c-4336-b2d7-e400772dc110-sys\") pod \"glance-default-internal-api-0\" (UID: \"9402b27d-1b2c-4336-b2d7-e400772dc110\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:03:56 crc kubenswrapper[4813]: I1201 09:03:56.903808 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/9402b27d-1b2c-4336-b2d7-e400772dc110-lib-modules\") pod \"glance-default-internal-api-0\" (UID: \"9402b27d-1b2c-4336-b2d7-e400772dc110\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:03:56 crc kubenswrapper[4813]: I1201 09:03:56.904058 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/83f41607-b074-4b35-b9d3-ec3199ae2ede-logs\") pod \"glance-default-internal-api-1\" (UID: \"83f41607-b074-4b35-b9d3-ec3199ae2ede\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Dec 01 09:03:56 crc kubenswrapper[4813]: I1201 09:03:56.904423 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/83f41607-b074-4b35-b9d3-ec3199ae2ede-var-locks-brick\") pod \"glance-default-internal-api-1\" (UID: \"83f41607-b074-4b35-b9d3-ec3199ae2ede\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Dec 01 09:03:56 crc kubenswrapper[4813]: I1201 09:03:56.906149 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/83f41607-b074-4b35-b9d3-ec3199ae2ede-run\") pod \"glance-default-internal-api-1\" (UID: \"83f41607-b074-4b35-b9d3-ec3199ae2ede\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Dec 01 09:03:56 crc kubenswrapper[4813]: I1201 09:03:56.906495 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/9402b27d-1b2c-4336-b2d7-e400772dc110-run\") pod \"glance-default-internal-api-0\" (UID: \"9402b27d-1b2c-4336-b2d7-e400772dc110\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:03:56 crc kubenswrapper[4813]: I1201 09:03:56.906528 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9402b27d-1b2c-4336-b2d7-e400772dc110-logs\") pod \"glance-default-internal-api-0\" (UID: \"9402b27d-1b2c-4336-b2d7-e400772dc110\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:03:56 crc kubenswrapper[4813]: I1201 09:03:56.906563 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/83f41607-b074-4b35-b9d3-ec3199ae2ede-dev\") pod \"glance-default-internal-api-1\" (UID: \"83f41607-b074-4b35-b9d3-ec3199ae2ede\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Dec 01 09:03:56 crc kubenswrapper[4813]: I1201 09:03:56.906559 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/83f41607-b074-4b35-b9d3-ec3199ae2ede-config-data\") pod \"glance-default-internal-api-1\" (UID: \"83f41607-b074-4b35-b9d3-ec3199ae2ede\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Dec 01 09:03:56 crc kubenswrapper[4813]: I1201 09:03:56.906581 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/9402b27d-1b2c-4336-b2d7-e400772dc110-var-locks-brick\") pod \"glance-default-internal-api-0\" (UID: \"9402b27d-1b2c-4336-b2d7-e400772dc110\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:03:56 crc kubenswrapper[4813]: I1201 09:03:56.906736 4813 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-internal-api-1\" (UID: \"83f41607-b074-4b35-b9d3-ec3199ae2ede\") device mount path \"/mnt/openstack/pv05\"" pod="glance-kuttl-tests/glance-default-internal-api-1" Dec 01 09:03:56 crc kubenswrapper[4813]: I1201 09:03:56.906775 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/83f41607-b074-4b35-b9d3-ec3199ae2ede-httpd-run\") pod \"glance-default-internal-api-1\" (UID: \"83f41607-b074-4b35-b9d3-ec3199ae2ede\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Dec 01 09:03:56 crc kubenswrapper[4813]: I1201 09:03:56.906843 4813 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-internal-api-1\" (UID: \"83f41607-b074-4b35-b9d3-ec3199ae2ede\") device mount path \"/mnt/openstack/pv01\"" pod="glance-kuttl-tests/glance-default-internal-api-1" Dec 01 09:03:56 crc kubenswrapper[4813]: I1201 09:03:56.906120 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/9402b27d-1b2c-4336-b2d7-e400772dc110-dev\") pod \"glance-default-internal-api-0\" (UID: \"9402b27d-1b2c-4336-b2d7-e400772dc110\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:03:56 crc kubenswrapper[4813]: I1201 09:03:56.908063 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/83f41607-b074-4b35-b9d3-ec3199ae2ede-lib-modules\") pod \"glance-default-internal-api-1\" (UID: \"83f41607-b074-4b35-b9d3-ec3199ae2ede\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Dec 01 09:03:56 crc kubenswrapper[4813]: I1201 09:03:56.908769 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/83f41607-b074-4b35-b9d3-ec3199ae2ede-scripts\") pod \"glance-default-internal-api-1\" (UID: \"83f41607-b074-4b35-b9d3-ec3199ae2ede\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Dec 01 09:03:56 crc kubenswrapper[4813]: I1201 09:03:56.917091 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9402b27d-1b2c-4336-b2d7-e400772dc110-config-data\") pod \"glance-default-internal-api-0\" (UID: \"9402b27d-1b2c-4336-b2d7-e400772dc110\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:03:56 crc kubenswrapper[4813]: I1201 09:03:56.921112 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fwtdm\" (UniqueName: \"kubernetes.io/projected/83f41607-b074-4b35-b9d3-ec3199ae2ede-kube-api-access-fwtdm\") pod \"glance-default-internal-api-1\" (UID: \"83f41607-b074-4b35-b9d3-ec3199ae2ede\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Dec 01 09:03:56 crc kubenswrapper[4813]: I1201 09:03:56.922845 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-internal-api-0\" (UID: \"9402b27d-1b2c-4336-b2d7-e400772dc110\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:03:56 crc kubenswrapper[4813]: I1201 09:03:56.923740 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9402b27d-1b2c-4336-b2d7-e400772dc110-scripts\") pod \"glance-default-internal-api-0\" (UID: \"9402b27d-1b2c-4336-b2d7-e400772dc110\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:03:56 crc kubenswrapper[4813]: I1201 09:03:56.926170 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zgxvz\" (UniqueName: \"kubernetes.io/projected/9402b27d-1b2c-4336-b2d7-e400772dc110-kube-api-access-zgxvz\") pod \"glance-default-internal-api-0\" (UID: \"9402b27d-1b2c-4336-b2d7-e400772dc110\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:03:56 crc kubenswrapper[4813]: I1201 09:03:56.931632 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-internal-api-1\" (UID: \"83f41607-b074-4b35-b9d3-ec3199ae2ede\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Dec 01 09:03:56 crc kubenswrapper[4813]: I1201 09:03:56.936061 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-internal-api-1\" (UID: \"83f41607-b074-4b35-b9d3-ec3199ae2ede\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Dec 01 09:03:56 crc kubenswrapper[4813]: I1201 09:03:56.956429 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-internal-api-0\" (UID: \"9402b27d-1b2c-4336-b2d7-e400772dc110\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:03:56 crc kubenswrapper[4813]: I1201 09:03:56.977710 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:03:56 crc kubenswrapper[4813]: I1201 09:03:56.994817 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-1" Dec 01 09:03:57 crc kubenswrapper[4813]: I1201 09:03:57.036759 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-1"] Dec 01 09:03:57 crc kubenswrapper[4813]: W1201 09:03:57.046683 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb4c39a03_2938_4b10_b430_369066c79ce4.slice/crio-f5144a488ccb46ecee1341b65655217027f5b5efc12f927118b96e06368e6f8e WatchSource:0}: Error finding container f5144a488ccb46ecee1341b65655217027f5b5efc12f927118b96e06368e6f8e: Status 404 returned error can't find the container with id f5144a488ccb46ecee1341b65655217027f5b5efc12f927118b96e06368e6f8e Dec 01 09:03:57 crc kubenswrapper[4813]: I1201 09:03:57.140896 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-0" Dec 01 09:03:57 crc kubenswrapper[4813]: I1201 09:03:57.261824 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-1" event={"ID":"b4c39a03-2938-4b10-b430-369066c79ce4","Type":"ContainerStarted","Data":"86514afa63c5089fdc14f9b082809d402f7390c825756c51d1a5199647657b3d"} Dec 01 09:03:57 crc kubenswrapper[4813]: I1201 09:03:57.261889 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-1" event={"ID":"b4c39a03-2938-4b10-b430-369066c79ce4","Type":"ContainerStarted","Data":"f5144a488ccb46ecee1341b65655217027f5b5efc12f927118b96e06368e6f8e"} Dec 01 09:03:57 crc kubenswrapper[4813]: I1201 09:03:57.286854 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Dec 01 09:03:57 crc kubenswrapper[4813]: I1201 09:03:57.513657 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-0"] Dec 01 09:03:57 crc kubenswrapper[4813]: W1201 09:03:57.524586 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod86363010_586b_4ee9_89e6_0549789e4f0a.slice/crio-daf104d88c1c187e3531e2cfd38aa8d0b377662713baed5c4bc193cf319b57d4 WatchSource:0}: Error finding container daf104d88c1c187e3531e2cfd38aa8d0b377662713baed5c4bc193cf319b57d4: Status 404 returned error can't find the container with id daf104d88c1c187e3531e2cfd38aa8d0b377662713baed5c4bc193cf319b57d4 Dec 01 09:03:57 crc kubenswrapper[4813]: I1201 09:03:57.571666 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-1"] Dec 01 09:03:58 crc kubenswrapper[4813]: I1201 09:03:58.059277 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-1"] Dec 01 09:03:58 crc kubenswrapper[4813]: I1201 09:03:58.271727 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-1" event={"ID":"83f41607-b074-4b35-b9d3-ec3199ae2ede","Type":"ContainerStarted","Data":"615f6e1b9decf443fef1b527ed095bc7f319daf23a04a467e1a0ea7f6250ff5a"} Dec 01 09:03:58 crc kubenswrapper[4813]: I1201 09:03:58.271947 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-1" event={"ID":"83f41607-b074-4b35-b9d3-ec3199ae2ede","Type":"ContainerStarted","Data":"b7f8d981dda1692ad1a5cb2ff292ca2837357644d17cba9508a1a78a4d8bb79e"} Dec 01 09:03:58 crc kubenswrapper[4813]: I1201 09:03:58.271973 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-1" event={"ID":"83f41607-b074-4b35-b9d3-ec3199ae2ede","Type":"ContainerStarted","Data":"694570d3b2906d400888aaeb73c49c9f333baee6a147aa4dfb512aa8951a116f"} Dec 01 09:03:58 crc kubenswrapper[4813]: I1201 09:03:58.275444 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-1" event={"ID":"b4c39a03-2938-4b10-b430-369066c79ce4","Type":"ContainerStarted","Data":"bdc319325263c4f53d0f49b476c0d63f087082dd8eda5e1e4f5f291bdaaffad5"} Dec 01 09:03:58 crc kubenswrapper[4813]: I1201 09:03:58.277883 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"9402b27d-1b2c-4336-b2d7-e400772dc110","Type":"ContainerStarted","Data":"bcefcda3036d2b3710278d3a31141318f7fcff5e3c0317be8a09c5d9dab936e7"} Dec 01 09:03:58 crc kubenswrapper[4813]: I1201 09:03:58.277906 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"9402b27d-1b2c-4336-b2d7-e400772dc110","Type":"ContainerStarted","Data":"565f50a4ba9ebae84a0c5064170efc08f5c3b6a0e5dfb2a87873962b319f2134"} Dec 01 09:03:58 crc kubenswrapper[4813]: I1201 09:03:58.277916 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"9402b27d-1b2c-4336-b2d7-e400772dc110","Type":"ContainerStarted","Data":"3405104a949e10ce494451973e586f506d85620fc5ad6c96af27d70760db16b7"} Dec 01 09:03:58 crc kubenswrapper[4813]: I1201 09:03:58.280040 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-0" event={"ID":"86363010-586b-4ee9-89e6-0549789e4f0a","Type":"ContainerStarted","Data":"251f770953b244d0689cf3f4f5507594545fba685c12cfb8b522e7004dc4edd3"} Dec 01 09:03:58 crc kubenswrapper[4813]: I1201 09:03:58.280078 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-0" event={"ID":"86363010-586b-4ee9-89e6-0549789e4f0a","Type":"ContainerStarted","Data":"d36d636ba6a08f873897903ab8eace8f40250273d37efbf7a92741f6fb28b059"} Dec 01 09:03:58 crc kubenswrapper[4813]: I1201 09:03:58.280088 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-0" event={"ID":"86363010-586b-4ee9-89e6-0549789e4f0a","Type":"ContainerStarted","Data":"daf104d88c1c187e3531e2cfd38aa8d0b377662713baed5c4bc193cf319b57d4"} Dec 01 09:03:58 crc kubenswrapper[4813]: I1201 09:03:58.298881 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-default-external-api-1" podStartSLOduration=2.298857888 podStartE2EDuration="2.298857888s" podCreationTimestamp="2025-12-01 09:03:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 09:03:58.2968149 +0000 UTC m=+1422.919736486" watchObservedRunningTime="2025-12-01 09:03:58.298857888 +0000 UTC m=+1422.921779464" Dec 01 09:03:58 crc kubenswrapper[4813]: I1201 09:03:58.329737 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-default-internal-api-0" podStartSLOduration=3.3297193800000002 podStartE2EDuration="3.32971938s" podCreationTimestamp="2025-12-01 09:03:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 09:03:58.323557746 +0000 UTC m=+1422.946479342" watchObservedRunningTime="2025-12-01 09:03:58.32971938 +0000 UTC m=+1422.952640966" Dec 01 09:03:58 crc kubenswrapper[4813]: I1201 09:03:58.350673 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-default-external-api-0" podStartSLOduration=3.350647101 podStartE2EDuration="3.350647101s" podCreationTimestamp="2025-12-01 09:03:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 09:03:58.347373818 +0000 UTC m=+1422.970295404" watchObservedRunningTime="2025-12-01 09:03:58.350647101 +0000 UTC m=+1422.973568687" Dec 01 09:03:59 crc kubenswrapper[4813]: I1201 09:03:59.293390 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-internal-api-1" podUID="83f41607-b074-4b35-b9d3-ec3199ae2ede" containerName="glance-log" containerID="cri-o://b7f8d981dda1692ad1a5cb2ff292ca2837357644d17cba9508a1a78a4d8bb79e" gracePeriod=30 Dec 01 09:03:59 crc kubenswrapper[4813]: I1201 09:03:59.293783 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-internal-api-1" podUID="83f41607-b074-4b35-b9d3-ec3199ae2ede" containerName="glance-httpd" containerID="cri-o://615f6e1b9decf443fef1b527ed095bc7f319daf23a04a467e1a0ea7f6250ff5a" gracePeriod=30 Dec 01 09:03:59 crc kubenswrapper[4813]: I1201 09:03:59.326982 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-default-internal-api-1" podStartSLOduration=4.326942391 podStartE2EDuration="4.326942391s" podCreationTimestamp="2025-12-01 09:03:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 09:03:59.32551529 +0000 UTC m=+1423.948436886" watchObservedRunningTime="2025-12-01 09:03:59.326942391 +0000 UTC m=+1423.949863977" Dec 01 09:03:59 crc kubenswrapper[4813]: I1201 09:03:59.836723 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-1" Dec 01 09:03:59 crc kubenswrapper[4813]: I1201 09:03:59.943630 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"83f41607-b074-4b35-b9d3-ec3199ae2ede\" (UID: \"83f41607-b074-4b35-b9d3-ec3199ae2ede\") " Dec 01 09:03:59 crc kubenswrapper[4813]: I1201 09:03:59.943678 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/83f41607-b074-4b35-b9d3-ec3199ae2ede-sys\") pod \"83f41607-b074-4b35-b9d3-ec3199ae2ede\" (UID: \"83f41607-b074-4b35-b9d3-ec3199ae2ede\") " Dec 01 09:03:59 crc kubenswrapper[4813]: I1201 09:03:59.943700 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance-cache\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"83f41607-b074-4b35-b9d3-ec3199ae2ede\" (UID: \"83f41607-b074-4b35-b9d3-ec3199ae2ede\") " Dec 01 09:03:59 crc kubenswrapper[4813]: I1201 09:03:59.943722 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/83f41607-b074-4b35-b9d3-ec3199ae2ede-var-locks-brick\") pod \"83f41607-b074-4b35-b9d3-ec3199ae2ede\" (UID: \"83f41607-b074-4b35-b9d3-ec3199ae2ede\") " Dec 01 09:03:59 crc kubenswrapper[4813]: I1201 09:03:59.943760 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/83f41607-b074-4b35-b9d3-ec3199ae2ede-config-data\") pod \"83f41607-b074-4b35-b9d3-ec3199ae2ede\" (UID: \"83f41607-b074-4b35-b9d3-ec3199ae2ede\") " Dec 01 09:03:59 crc kubenswrapper[4813]: I1201 09:03:59.943784 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/83f41607-b074-4b35-b9d3-ec3199ae2ede-scripts\") pod \"83f41607-b074-4b35-b9d3-ec3199ae2ede\" (UID: \"83f41607-b074-4b35-b9d3-ec3199ae2ede\") " Dec 01 09:03:59 crc kubenswrapper[4813]: I1201 09:03:59.943819 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/83f41607-b074-4b35-b9d3-ec3199ae2ede-logs\") pod \"83f41607-b074-4b35-b9d3-ec3199ae2ede\" (UID: \"83f41607-b074-4b35-b9d3-ec3199ae2ede\") " Dec 01 09:03:59 crc kubenswrapper[4813]: I1201 09:03:59.943841 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/83f41607-b074-4b35-b9d3-ec3199ae2ede-dev\") pod \"83f41607-b074-4b35-b9d3-ec3199ae2ede\" (UID: \"83f41607-b074-4b35-b9d3-ec3199ae2ede\") " Dec 01 09:03:59 crc kubenswrapper[4813]: I1201 09:03:59.943858 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/83f41607-b074-4b35-b9d3-ec3199ae2ede-run\") pod \"83f41607-b074-4b35-b9d3-ec3199ae2ede\" (UID: \"83f41607-b074-4b35-b9d3-ec3199ae2ede\") " Dec 01 09:03:59 crc kubenswrapper[4813]: I1201 09:03:59.943818 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/83f41607-b074-4b35-b9d3-ec3199ae2ede-sys" (OuterVolumeSpecName: "sys") pod "83f41607-b074-4b35-b9d3-ec3199ae2ede" (UID: "83f41607-b074-4b35-b9d3-ec3199ae2ede"). InnerVolumeSpecName "sys". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 09:03:59 crc kubenswrapper[4813]: I1201 09:03:59.943902 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/83f41607-b074-4b35-b9d3-ec3199ae2ede-run" (OuterVolumeSpecName: "run") pod "83f41607-b074-4b35-b9d3-ec3199ae2ede" (UID: "83f41607-b074-4b35-b9d3-ec3199ae2ede"). InnerVolumeSpecName "run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 09:03:59 crc kubenswrapper[4813]: I1201 09:03:59.943908 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/83f41607-b074-4b35-b9d3-ec3199ae2ede-etc-nvme" (OuterVolumeSpecName: "etc-nvme") pod "83f41607-b074-4b35-b9d3-ec3199ae2ede" (UID: "83f41607-b074-4b35-b9d3-ec3199ae2ede"). InnerVolumeSpecName "etc-nvme". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 09:03:59 crc kubenswrapper[4813]: I1201 09:03:59.943879 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/83f41607-b074-4b35-b9d3-ec3199ae2ede-etc-nvme\") pod \"83f41607-b074-4b35-b9d3-ec3199ae2ede\" (UID: \"83f41607-b074-4b35-b9d3-ec3199ae2ede\") " Dec 01 09:03:59 crc kubenswrapper[4813]: I1201 09:03:59.943844 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/83f41607-b074-4b35-b9d3-ec3199ae2ede-var-locks-brick" (OuterVolumeSpecName: "var-locks-brick") pod "83f41607-b074-4b35-b9d3-ec3199ae2ede" (UID: "83f41607-b074-4b35-b9d3-ec3199ae2ede"). InnerVolumeSpecName "var-locks-brick". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 09:03:59 crc kubenswrapper[4813]: I1201 09:03:59.943877 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/83f41607-b074-4b35-b9d3-ec3199ae2ede-dev" (OuterVolumeSpecName: "dev") pod "83f41607-b074-4b35-b9d3-ec3199ae2ede" (UID: "83f41607-b074-4b35-b9d3-ec3199ae2ede"). InnerVolumeSpecName "dev". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 09:03:59 crc kubenswrapper[4813]: I1201 09:03:59.944002 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/83f41607-b074-4b35-b9d3-ec3199ae2ede-lib-modules\") pod \"83f41607-b074-4b35-b9d3-ec3199ae2ede\" (UID: \"83f41607-b074-4b35-b9d3-ec3199ae2ede\") " Dec 01 09:03:59 crc kubenswrapper[4813]: I1201 09:03:59.944052 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/83f41607-b074-4b35-b9d3-ec3199ae2ede-httpd-run\") pod \"83f41607-b074-4b35-b9d3-ec3199ae2ede\" (UID: \"83f41607-b074-4b35-b9d3-ec3199ae2ede\") " Dec 01 09:03:59 crc kubenswrapper[4813]: I1201 09:03:59.944079 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fwtdm\" (UniqueName: \"kubernetes.io/projected/83f41607-b074-4b35-b9d3-ec3199ae2ede-kube-api-access-fwtdm\") pod \"83f41607-b074-4b35-b9d3-ec3199ae2ede\" (UID: \"83f41607-b074-4b35-b9d3-ec3199ae2ede\") " Dec 01 09:03:59 crc kubenswrapper[4813]: I1201 09:03:59.944079 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/83f41607-b074-4b35-b9d3-ec3199ae2ede-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "83f41607-b074-4b35-b9d3-ec3199ae2ede" (UID: "83f41607-b074-4b35-b9d3-ec3199ae2ede"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 09:03:59 crc kubenswrapper[4813]: I1201 09:03:59.944114 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/83f41607-b074-4b35-b9d3-ec3199ae2ede-etc-iscsi\") pod \"83f41607-b074-4b35-b9d3-ec3199ae2ede\" (UID: \"83f41607-b074-4b35-b9d3-ec3199ae2ede\") " Dec 01 09:03:59 crc kubenswrapper[4813]: I1201 09:03:59.944237 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/83f41607-b074-4b35-b9d3-ec3199ae2ede-logs" (OuterVolumeSpecName: "logs") pod "83f41607-b074-4b35-b9d3-ec3199ae2ede" (UID: "83f41607-b074-4b35-b9d3-ec3199ae2ede"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 09:03:59 crc kubenswrapper[4813]: I1201 09:03:59.944273 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/83f41607-b074-4b35-b9d3-ec3199ae2ede-etc-iscsi" (OuterVolumeSpecName: "etc-iscsi") pod "83f41607-b074-4b35-b9d3-ec3199ae2ede" (UID: "83f41607-b074-4b35-b9d3-ec3199ae2ede"). InnerVolumeSpecName "etc-iscsi". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 09:03:59 crc kubenswrapper[4813]: I1201 09:03:59.944388 4813 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/83f41607-b074-4b35-b9d3-ec3199ae2ede-logs\") on node \"crc\" DevicePath \"\"" Dec 01 09:03:59 crc kubenswrapper[4813]: I1201 09:03:59.944419 4813 reconciler_common.go:293] "Volume detached for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/83f41607-b074-4b35-b9d3-ec3199ae2ede-dev\") on node \"crc\" DevicePath \"\"" Dec 01 09:03:59 crc kubenswrapper[4813]: I1201 09:03:59.944430 4813 reconciler_common.go:293] "Volume detached for volume \"run\" (UniqueName: \"kubernetes.io/host-path/83f41607-b074-4b35-b9d3-ec3199ae2ede-run\") on node \"crc\" DevicePath \"\"" Dec 01 09:03:59 crc kubenswrapper[4813]: I1201 09:03:59.944440 4813 reconciler_common.go:293] "Volume detached for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/83f41607-b074-4b35-b9d3-ec3199ae2ede-etc-nvme\") on node \"crc\" DevicePath \"\"" Dec 01 09:03:59 crc kubenswrapper[4813]: I1201 09:03:59.944448 4813 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/83f41607-b074-4b35-b9d3-ec3199ae2ede-lib-modules\") on node \"crc\" DevicePath \"\"" Dec 01 09:03:59 crc kubenswrapper[4813]: I1201 09:03:59.944455 4813 reconciler_common.go:293] "Volume detached for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/83f41607-b074-4b35-b9d3-ec3199ae2ede-etc-iscsi\") on node \"crc\" DevicePath \"\"" Dec 01 09:03:59 crc kubenswrapper[4813]: I1201 09:03:59.944464 4813 reconciler_common.go:293] "Volume detached for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/83f41607-b074-4b35-b9d3-ec3199ae2ede-sys\") on node \"crc\" DevicePath \"\"" Dec 01 09:03:59 crc kubenswrapper[4813]: I1201 09:03:59.944472 4813 reconciler_common.go:293] "Volume detached for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/83f41607-b074-4b35-b9d3-ec3199ae2ede-var-locks-brick\") on node \"crc\" DevicePath \"\"" Dec 01 09:03:59 crc kubenswrapper[4813]: I1201 09:03:59.944385 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/83f41607-b074-4b35-b9d3-ec3199ae2ede-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "83f41607-b074-4b35-b9d3-ec3199ae2ede" (UID: "83f41607-b074-4b35-b9d3-ec3199ae2ede"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 09:03:59 crc kubenswrapper[4813]: I1201 09:03:59.949231 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage05-crc" (OuterVolumeSpecName: "glance") pod "83f41607-b074-4b35-b9d3-ec3199ae2ede" (UID: "83f41607-b074-4b35-b9d3-ec3199ae2ede"). InnerVolumeSpecName "local-storage05-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 01 09:03:59 crc kubenswrapper[4813]: I1201 09:03:59.949398 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/83f41607-b074-4b35-b9d3-ec3199ae2ede-kube-api-access-fwtdm" (OuterVolumeSpecName: "kube-api-access-fwtdm") pod "83f41607-b074-4b35-b9d3-ec3199ae2ede" (UID: "83f41607-b074-4b35-b9d3-ec3199ae2ede"). InnerVolumeSpecName "kube-api-access-fwtdm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:03:59 crc kubenswrapper[4813]: I1201 09:03:59.950122 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/83f41607-b074-4b35-b9d3-ec3199ae2ede-scripts" (OuterVolumeSpecName: "scripts") pod "83f41607-b074-4b35-b9d3-ec3199ae2ede" (UID: "83f41607-b074-4b35-b9d3-ec3199ae2ede"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:03:59 crc kubenswrapper[4813]: I1201 09:03:59.950163 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage01-crc" (OuterVolumeSpecName: "glance-cache") pod "83f41607-b074-4b35-b9d3-ec3199ae2ede" (UID: "83f41607-b074-4b35-b9d3-ec3199ae2ede"). InnerVolumeSpecName "local-storage01-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 01 09:03:59 crc kubenswrapper[4813]: I1201 09:03:59.999005 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/83f41607-b074-4b35-b9d3-ec3199ae2ede-config-data" (OuterVolumeSpecName: "config-data") pod "83f41607-b074-4b35-b9d3-ec3199ae2ede" (UID: "83f41607-b074-4b35-b9d3-ec3199ae2ede"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:04:00 crc kubenswrapper[4813]: I1201 09:04:00.045504 4813 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/83f41607-b074-4b35-b9d3-ec3199ae2ede-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 01 09:04:00 crc kubenswrapper[4813]: I1201 09:04:00.045539 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fwtdm\" (UniqueName: \"kubernetes.io/projected/83f41607-b074-4b35-b9d3-ec3199ae2ede-kube-api-access-fwtdm\") on node \"crc\" DevicePath \"\"" Dec 01 09:04:00 crc kubenswrapper[4813]: I1201 09:04:00.045593 4813 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" " Dec 01 09:04:00 crc kubenswrapper[4813]: I1201 09:04:00.045613 4813 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" " Dec 01 09:04:00 crc kubenswrapper[4813]: I1201 09:04:00.045625 4813 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/83f41607-b074-4b35-b9d3-ec3199ae2ede-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 09:04:00 crc kubenswrapper[4813]: I1201 09:04:00.045639 4813 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/83f41607-b074-4b35-b9d3-ec3199ae2ede-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 09:04:00 crc kubenswrapper[4813]: I1201 09:04:00.062052 4813 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage05-crc" (UniqueName: "kubernetes.io/local-volume/local-storage05-crc") on node "crc" Dec 01 09:04:00 crc kubenswrapper[4813]: I1201 09:04:00.064895 4813 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage01-crc" (UniqueName: "kubernetes.io/local-volume/local-storage01-crc") on node "crc" Dec 01 09:04:00 crc kubenswrapper[4813]: I1201 09:04:00.147169 4813 reconciler_common.go:293] "Volume detached for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" DevicePath \"\"" Dec 01 09:04:00 crc kubenswrapper[4813]: I1201 09:04:00.147202 4813 reconciler_common.go:293] "Volume detached for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" DevicePath \"\"" Dec 01 09:04:00 crc kubenswrapper[4813]: I1201 09:04:00.304697 4813 generic.go:334] "Generic (PLEG): container finished" podID="83f41607-b074-4b35-b9d3-ec3199ae2ede" containerID="615f6e1b9decf443fef1b527ed095bc7f319daf23a04a467e1a0ea7f6250ff5a" exitCode=0 Dec 01 09:04:00 crc kubenswrapper[4813]: I1201 09:04:00.304767 4813 generic.go:334] "Generic (PLEG): container finished" podID="83f41607-b074-4b35-b9d3-ec3199ae2ede" containerID="b7f8d981dda1692ad1a5cb2ff292ca2837357644d17cba9508a1a78a4d8bb79e" exitCode=143 Dec 01 09:04:00 crc kubenswrapper[4813]: I1201 09:04:00.304789 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-1" Dec 01 09:04:00 crc kubenswrapper[4813]: I1201 09:04:00.304821 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-1" event={"ID":"83f41607-b074-4b35-b9d3-ec3199ae2ede","Type":"ContainerDied","Data":"615f6e1b9decf443fef1b527ed095bc7f319daf23a04a467e1a0ea7f6250ff5a"} Dec 01 09:04:00 crc kubenswrapper[4813]: I1201 09:04:00.305106 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-1" event={"ID":"83f41607-b074-4b35-b9d3-ec3199ae2ede","Type":"ContainerDied","Data":"b7f8d981dda1692ad1a5cb2ff292ca2837357644d17cba9508a1a78a4d8bb79e"} Dec 01 09:04:00 crc kubenswrapper[4813]: I1201 09:04:00.305131 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-1" event={"ID":"83f41607-b074-4b35-b9d3-ec3199ae2ede","Type":"ContainerDied","Data":"694570d3b2906d400888aaeb73c49c9f333baee6a147aa4dfb512aa8951a116f"} Dec 01 09:04:00 crc kubenswrapper[4813]: I1201 09:04:00.305184 4813 scope.go:117] "RemoveContainer" containerID="615f6e1b9decf443fef1b527ed095bc7f319daf23a04a467e1a0ea7f6250ff5a" Dec 01 09:04:00 crc kubenswrapper[4813]: I1201 09:04:00.329424 4813 scope.go:117] "RemoveContainer" containerID="b7f8d981dda1692ad1a5cb2ff292ca2837357644d17cba9508a1a78a4d8bb79e" Dec 01 09:04:00 crc kubenswrapper[4813]: I1201 09:04:00.349746 4813 scope.go:117] "RemoveContainer" containerID="615f6e1b9decf443fef1b527ed095bc7f319daf23a04a467e1a0ea7f6250ff5a" Dec 01 09:04:00 crc kubenswrapper[4813]: E1201 09:04:00.350435 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"615f6e1b9decf443fef1b527ed095bc7f319daf23a04a467e1a0ea7f6250ff5a\": container with ID starting with 615f6e1b9decf443fef1b527ed095bc7f319daf23a04a467e1a0ea7f6250ff5a not found: ID does not exist" containerID="615f6e1b9decf443fef1b527ed095bc7f319daf23a04a467e1a0ea7f6250ff5a" Dec 01 09:04:00 crc kubenswrapper[4813]: I1201 09:04:00.350489 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"615f6e1b9decf443fef1b527ed095bc7f319daf23a04a467e1a0ea7f6250ff5a"} err="failed to get container status \"615f6e1b9decf443fef1b527ed095bc7f319daf23a04a467e1a0ea7f6250ff5a\": rpc error: code = NotFound desc = could not find container \"615f6e1b9decf443fef1b527ed095bc7f319daf23a04a467e1a0ea7f6250ff5a\": container with ID starting with 615f6e1b9decf443fef1b527ed095bc7f319daf23a04a467e1a0ea7f6250ff5a not found: ID does not exist" Dec 01 09:04:00 crc kubenswrapper[4813]: I1201 09:04:00.350520 4813 scope.go:117] "RemoveContainer" containerID="b7f8d981dda1692ad1a5cb2ff292ca2837357644d17cba9508a1a78a4d8bb79e" Dec 01 09:04:00 crc kubenswrapper[4813]: E1201 09:04:00.350941 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b7f8d981dda1692ad1a5cb2ff292ca2837357644d17cba9508a1a78a4d8bb79e\": container with ID starting with b7f8d981dda1692ad1a5cb2ff292ca2837357644d17cba9508a1a78a4d8bb79e not found: ID does not exist" containerID="b7f8d981dda1692ad1a5cb2ff292ca2837357644d17cba9508a1a78a4d8bb79e" Dec 01 09:04:00 crc kubenswrapper[4813]: I1201 09:04:00.351011 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b7f8d981dda1692ad1a5cb2ff292ca2837357644d17cba9508a1a78a4d8bb79e"} err="failed to get container status \"b7f8d981dda1692ad1a5cb2ff292ca2837357644d17cba9508a1a78a4d8bb79e\": rpc error: code = NotFound desc = could not find container \"b7f8d981dda1692ad1a5cb2ff292ca2837357644d17cba9508a1a78a4d8bb79e\": container with ID starting with b7f8d981dda1692ad1a5cb2ff292ca2837357644d17cba9508a1a78a4d8bb79e not found: ID does not exist" Dec 01 09:04:00 crc kubenswrapper[4813]: I1201 09:04:00.351055 4813 scope.go:117] "RemoveContainer" containerID="615f6e1b9decf443fef1b527ed095bc7f319daf23a04a467e1a0ea7f6250ff5a" Dec 01 09:04:00 crc kubenswrapper[4813]: I1201 09:04:00.351156 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-1"] Dec 01 09:04:00 crc kubenswrapper[4813]: I1201 09:04:00.351371 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"615f6e1b9decf443fef1b527ed095bc7f319daf23a04a467e1a0ea7f6250ff5a"} err="failed to get container status \"615f6e1b9decf443fef1b527ed095bc7f319daf23a04a467e1a0ea7f6250ff5a\": rpc error: code = NotFound desc = could not find container \"615f6e1b9decf443fef1b527ed095bc7f319daf23a04a467e1a0ea7f6250ff5a\": container with ID starting with 615f6e1b9decf443fef1b527ed095bc7f319daf23a04a467e1a0ea7f6250ff5a not found: ID does not exist" Dec 01 09:04:00 crc kubenswrapper[4813]: I1201 09:04:00.351399 4813 scope.go:117] "RemoveContainer" containerID="b7f8d981dda1692ad1a5cb2ff292ca2837357644d17cba9508a1a78a4d8bb79e" Dec 01 09:04:00 crc kubenswrapper[4813]: I1201 09:04:00.351783 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b7f8d981dda1692ad1a5cb2ff292ca2837357644d17cba9508a1a78a4d8bb79e"} err="failed to get container status \"b7f8d981dda1692ad1a5cb2ff292ca2837357644d17cba9508a1a78a4d8bb79e\": rpc error: code = NotFound desc = could not find container \"b7f8d981dda1692ad1a5cb2ff292ca2837357644d17cba9508a1a78a4d8bb79e\": container with ID starting with b7f8d981dda1692ad1a5cb2ff292ca2837357644d17cba9508a1a78a4d8bb79e not found: ID does not exist" Dec 01 09:04:00 crc kubenswrapper[4813]: I1201 09:04:00.376837 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-1"] Dec 01 09:04:00 crc kubenswrapper[4813]: I1201 09:04:00.387347 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-1"] Dec 01 09:04:00 crc kubenswrapper[4813]: E1201 09:04:00.387763 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="83f41607-b074-4b35-b9d3-ec3199ae2ede" containerName="glance-httpd" Dec 01 09:04:00 crc kubenswrapper[4813]: I1201 09:04:00.387784 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="83f41607-b074-4b35-b9d3-ec3199ae2ede" containerName="glance-httpd" Dec 01 09:04:00 crc kubenswrapper[4813]: E1201 09:04:00.387800 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="83f41607-b074-4b35-b9d3-ec3199ae2ede" containerName="glance-log" Dec 01 09:04:00 crc kubenswrapper[4813]: I1201 09:04:00.387808 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="83f41607-b074-4b35-b9d3-ec3199ae2ede" containerName="glance-log" Dec 01 09:04:00 crc kubenswrapper[4813]: I1201 09:04:00.388004 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="83f41607-b074-4b35-b9d3-ec3199ae2ede" containerName="glance-log" Dec 01 09:04:00 crc kubenswrapper[4813]: I1201 09:04:00.388032 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="83f41607-b074-4b35-b9d3-ec3199ae2ede" containerName="glance-httpd" Dec 01 09:04:00 crc kubenswrapper[4813]: I1201 09:04:00.388952 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-1" Dec 01 09:04:00 crc kubenswrapper[4813]: I1201 09:04:00.405458 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="83f41607-b074-4b35-b9d3-ec3199ae2ede" path="/var/lib/kubelet/pods/83f41607-b074-4b35-b9d3-ec3199ae2ede/volumes" Dec 01 09:04:00 crc kubenswrapper[4813]: I1201 09:04:00.406039 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-1"] Dec 01 09:04:00 crc kubenswrapper[4813]: I1201 09:04:00.450903 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/c079746c-0f0e-403a-bb6e-b1866a063050-etc-iscsi\") pod \"glance-default-internal-api-1\" (UID: \"c079746c-0f0e-403a-bb6e-b1866a063050\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Dec 01 09:04:00 crc kubenswrapper[4813]: I1201 09:04:00.451014 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/c079746c-0f0e-403a-bb6e-b1866a063050-lib-modules\") pod \"glance-default-internal-api-1\" (UID: \"c079746c-0f0e-403a-bb6e-b1866a063050\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Dec 01 09:04:00 crc kubenswrapper[4813]: I1201 09:04:00.451038 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/c079746c-0f0e-403a-bb6e-b1866a063050-etc-nvme\") pod \"glance-default-internal-api-1\" (UID: \"c079746c-0f0e-403a-bb6e-b1866a063050\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Dec 01 09:04:00 crc kubenswrapper[4813]: I1201 09:04:00.451056 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c079746c-0f0e-403a-bb6e-b1866a063050-logs\") pod \"glance-default-internal-api-1\" (UID: \"c079746c-0f0e-403a-bb6e-b1866a063050\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Dec 01 09:04:00 crc kubenswrapper[4813]: I1201 09:04:00.451072 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/c079746c-0f0e-403a-bb6e-b1866a063050-httpd-run\") pod \"glance-default-internal-api-1\" (UID: \"c079746c-0f0e-403a-bb6e-b1866a063050\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Dec 01 09:04:00 crc kubenswrapper[4813]: I1201 09:04:00.451091 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/c079746c-0f0e-403a-bb6e-b1866a063050-run\") pod \"glance-default-internal-api-1\" (UID: \"c079746c-0f0e-403a-bb6e-b1866a063050\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Dec 01 09:04:00 crc kubenswrapper[4813]: I1201 09:04:00.451109 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/c079746c-0f0e-403a-bb6e-b1866a063050-var-locks-brick\") pod \"glance-default-internal-api-1\" (UID: \"c079746c-0f0e-403a-bb6e-b1866a063050\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Dec 01 09:04:00 crc kubenswrapper[4813]: I1201 09:04:00.451125 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/c079746c-0f0e-403a-bb6e-b1866a063050-sys\") pod \"glance-default-internal-api-1\" (UID: \"c079746c-0f0e-403a-bb6e-b1866a063050\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Dec 01 09:04:00 crc kubenswrapper[4813]: I1201 09:04:00.451158 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s6459\" (UniqueName: \"kubernetes.io/projected/c079746c-0f0e-403a-bb6e-b1866a063050-kube-api-access-s6459\") pod \"glance-default-internal-api-1\" (UID: \"c079746c-0f0e-403a-bb6e-b1866a063050\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Dec 01 09:04:00 crc kubenswrapper[4813]: I1201 09:04:00.451195 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-internal-api-1\" (UID: \"c079746c-0f0e-403a-bb6e-b1866a063050\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Dec 01 09:04:00 crc kubenswrapper[4813]: I1201 09:04:00.451221 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c079746c-0f0e-403a-bb6e-b1866a063050-config-data\") pod \"glance-default-internal-api-1\" (UID: \"c079746c-0f0e-403a-bb6e-b1866a063050\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Dec 01 09:04:00 crc kubenswrapper[4813]: I1201 09:04:00.451244 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/c079746c-0f0e-403a-bb6e-b1866a063050-dev\") pod \"glance-default-internal-api-1\" (UID: \"c079746c-0f0e-403a-bb6e-b1866a063050\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Dec 01 09:04:00 crc kubenswrapper[4813]: I1201 09:04:00.451269 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-internal-api-1\" (UID: \"c079746c-0f0e-403a-bb6e-b1866a063050\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Dec 01 09:04:00 crc kubenswrapper[4813]: I1201 09:04:00.451305 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c079746c-0f0e-403a-bb6e-b1866a063050-scripts\") pod \"glance-default-internal-api-1\" (UID: \"c079746c-0f0e-403a-bb6e-b1866a063050\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Dec 01 09:04:00 crc kubenswrapper[4813]: I1201 09:04:00.552189 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/c079746c-0f0e-403a-bb6e-b1866a063050-etc-nvme\") pod \"glance-default-internal-api-1\" (UID: \"c079746c-0f0e-403a-bb6e-b1866a063050\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Dec 01 09:04:00 crc kubenswrapper[4813]: I1201 09:04:00.552235 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c079746c-0f0e-403a-bb6e-b1866a063050-logs\") pod \"glance-default-internal-api-1\" (UID: \"c079746c-0f0e-403a-bb6e-b1866a063050\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Dec 01 09:04:00 crc kubenswrapper[4813]: I1201 09:04:00.552252 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/c079746c-0f0e-403a-bb6e-b1866a063050-httpd-run\") pod \"glance-default-internal-api-1\" (UID: \"c079746c-0f0e-403a-bb6e-b1866a063050\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Dec 01 09:04:00 crc kubenswrapper[4813]: I1201 09:04:00.552277 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/c079746c-0f0e-403a-bb6e-b1866a063050-run\") pod \"glance-default-internal-api-1\" (UID: \"c079746c-0f0e-403a-bb6e-b1866a063050\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Dec 01 09:04:00 crc kubenswrapper[4813]: I1201 09:04:00.552301 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/c079746c-0f0e-403a-bb6e-b1866a063050-var-locks-brick\") pod \"glance-default-internal-api-1\" (UID: \"c079746c-0f0e-403a-bb6e-b1866a063050\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Dec 01 09:04:00 crc kubenswrapper[4813]: I1201 09:04:00.552320 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/c079746c-0f0e-403a-bb6e-b1866a063050-sys\") pod \"glance-default-internal-api-1\" (UID: \"c079746c-0f0e-403a-bb6e-b1866a063050\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Dec 01 09:04:00 crc kubenswrapper[4813]: I1201 09:04:00.552357 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s6459\" (UniqueName: \"kubernetes.io/projected/c079746c-0f0e-403a-bb6e-b1866a063050-kube-api-access-s6459\") pod \"glance-default-internal-api-1\" (UID: \"c079746c-0f0e-403a-bb6e-b1866a063050\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Dec 01 09:04:00 crc kubenswrapper[4813]: I1201 09:04:00.552381 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-internal-api-1\" (UID: \"c079746c-0f0e-403a-bb6e-b1866a063050\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Dec 01 09:04:00 crc kubenswrapper[4813]: I1201 09:04:00.552406 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c079746c-0f0e-403a-bb6e-b1866a063050-config-data\") pod \"glance-default-internal-api-1\" (UID: \"c079746c-0f0e-403a-bb6e-b1866a063050\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Dec 01 09:04:00 crc kubenswrapper[4813]: I1201 09:04:00.552409 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/c079746c-0f0e-403a-bb6e-b1866a063050-sys\") pod \"glance-default-internal-api-1\" (UID: \"c079746c-0f0e-403a-bb6e-b1866a063050\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Dec 01 09:04:00 crc kubenswrapper[4813]: I1201 09:04:00.552449 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/c079746c-0f0e-403a-bb6e-b1866a063050-dev\") pod \"glance-default-internal-api-1\" (UID: \"c079746c-0f0e-403a-bb6e-b1866a063050\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Dec 01 09:04:00 crc kubenswrapper[4813]: I1201 09:04:00.552422 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/c079746c-0f0e-403a-bb6e-b1866a063050-dev\") pod \"glance-default-internal-api-1\" (UID: \"c079746c-0f0e-403a-bb6e-b1866a063050\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Dec 01 09:04:00 crc kubenswrapper[4813]: I1201 09:04:00.552446 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/c079746c-0f0e-403a-bb6e-b1866a063050-run\") pod \"glance-default-internal-api-1\" (UID: \"c079746c-0f0e-403a-bb6e-b1866a063050\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Dec 01 09:04:00 crc kubenswrapper[4813]: I1201 09:04:00.552501 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-internal-api-1\" (UID: \"c079746c-0f0e-403a-bb6e-b1866a063050\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Dec 01 09:04:00 crc kubenswrapper[4813]: I1201 09:04:00.552500 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/c079746c-0f0e-403a-bb6e-b1866a063050-var-locks-brick\") pod \"glance-default-internal-api-1\" (UID: \"c079746c-0f0e-403a-bb6e-b1866a063050\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Dec 01 09:04:00 crc kubenswrapper[4813]: I1201 09:04:00.552376 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/c079746c-0f0e-403a-bb6e-b1866a063050-etc-nvme\") pod \"glance-default-internal-api-1\" (UID: \"c079746c-0f0e-403a-bb6e-b1866a063050\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Dec 01 09:04:00 crc kubenswrapper[4813]: I1201 09:04:00.552541 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c079746c-0f0e-403a-bb6e-b1866a063050-scripts\") pod \"glance-default-internal-api-1\" (UID: \"c079746c-0f0e-403a-bb6e-b1866a063050\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Dec 01 09:04:00 crc kubenswrapper[4813]: I1201 09:04:00.552573 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/c079746c-0f0e-403a-bb6e-b1866a063050-etc-iscsi\") pod \"glance-default-internal-api-1\" (UID: \"c079746c-0f0e-403a-bb6e-b1866a063050\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Dec 01 09:04:00 crc kubenswrapper[4813]: I1201 09:04:00.552618 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/c079746c-0f0e-403a-bb6e-b1866a063050-lib-modules\") pod \"glance-default-internal-api-1\" (UID: \"c079746c-0f0e-403a-bb6e-b1866a063050\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Dec 01 09:04:00 crc kubenswrapper[4813]: I1201 09:04:00.552683 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/c079746c-0f0e-403a-bb6e-b1866a063050-lib-modules\") pod \"glance-default-internal-api-1\" (UID: \"c079746c-0f0e-403a-bb6e-b1866a063050\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Dec 01 09:04:00 crc kubenswrapper[4813]: I1201 09:04:00.552696 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c079746c-0f0e-403a-bb6e-b1866a063050-logs\") pod \"glance-default-internal-api-1\" (UID: \"c079746c-0f0e-403a-bb6e-b1866a063050\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Dec 01 09:04:00 crc kubenswrapper[4813]: I1201 09:04:00.552909 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/c079746c-0f0e-403a-bb6e-b1866a063050-httpd-run\") pod \"glance-default-internal-api-1\" (UID: \"c079746c-0f0e-403a-bb6e-b1866a063050\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Dec 01 09:04:00 crc kubenswrapper[4813]: I1201 09:04:00.553009 4813 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-internal-api-1\" (UID: \"c079746c-0f0e-403a-bb6e-b1866a063050\") device mount path \"/mnt/openstack/pv01\"" pod="glance-kuttl-tests/glance-default-internal-api-1" Dec 01 09:04:00 crc kubenswrapper[4813]: I1201 09:04:00.553071 4813 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-internal-api-1\" (UID: \"c079746c-0f0e-403a-bb6e-b1866a063050\") device mount path \"/mnt/openstack/pv05\"" pod="glance-kuttl-tests/glance-default-internal-api-1" Dec 01 09:04:00 crc kubenswrapper[4813]: I1201 09:04:00.553117 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/c079746c-0f0e-403a-bb6e-b1866a063050-etc-iscsi\") pod \"glance-default-internal-api-1\" (UID: \"c079746c-0f0e-403a-bb6e-b1866a063050\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Dec 01 09:04:00 crc kubenswrapper[4813]: I1201 09:04:00.557241 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c079746c-0f0e-403a-bb6e-b1866a063050-scripts\") pod \"glance-default-internal-api-1\" (UID: \"c079746c-0f0e-403a-bb6e-b1866a063050\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Dec 01 09:04:00 crc kubenswrapper[4813]: I1201 09:04:00.568644 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c079746c-0f0e-403a-bb6e-b1866a063050-config-data\") pod \"glance-default-internal-api-1\" (UID: \"c079746c-0f0e-403a-bb6e-b1866a063050\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Dec 01 09:04:00 crc kubenswrapper[4813]: I1201 09:04:00.573827 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s6459\" (UniqueName: \"kubernetes.io/projected/c079746c-0f0e-403a-bb6e-b1866a063050-kube-api-access-s6459\") pod \"glance-default-internal-api-1\" (UID: \"c079746c-0f0e-403a-bb6e-b1866a063050\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Dec 01 09:04:00 crc kubenswrapper[4813]: I1201 09:04:00.579888 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-internal-api-1\" (UID: \"c079746c-0f0e-403a-bb6e-b1866a063050\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Dec 01 09:04:00 crc kubenswrapper[4813]: I1201 09:04:00.580924 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-internal-api-1\" (UID: \"c079746c-0f0e-403a-bb6e-b1866a063050\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Dec 01 09:04:00 crc kubenswrapper[4813]: I1201 09:04:00.708885 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-1" Dec 01 09:04:01 crc kubenswrapper[4813]: I1201 09:04:01.137404 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-1"] Dec 01 09:04:01 crc kubenswrapper[4813]: I1201 09:04:01.330756 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-1" event={"ID":"c079746c-0f0e-403a-bb6e-b1866a063050","Type":"ContainerStarted","Data":"597e9606975353626d3b79d9dd36fc9fcfb029ca4c8d3da2851ae76fe368b589"} Dec 01 09:04:02 crc kubenswrapper[4813]: I1201 09:04:02.342742 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-1" event={"ID":"c079746c-0f0e-403a-bb6e-b1866a063050","Type":"ContainerStarted","Data":"d191ad43cbe03d01275077b6d3666ca27b942e2e476948af348780daa2824a34"} Dec 01 09:04:02 crc kubenswrapper[4813]: I1201 09:04:02.343296 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-1" event={"ID":"c079746c-0f0e-403a-bb6e-b1866a063050","Type":"ContainerStarted","Data":"d0e434dc9af434eac9e19163ad506f93f21a1b45f5335294ea9b9a8d72872b0f"} Dec 01 09:04:02 crc kubenswrapper[4813]: I1201 09:04:02.362060 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-default-internal-api-1" podStartSLOduration=2.36203395 podStartE2EDuration="2.36203395s" podCreationTimestamp="2025-12-01 09:04:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 09:04:02.361540496 +0000 UTC m=+1426.984462082" watchObservedRunningTime="2025-12-01 09:04:02.36203395 +0000 UTC m=+1426.984955546" Dec 01 09:04:06 crc kubenswrapper[4813]: I1201 09:04:06.678897 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-external-api-1" Dec 01 09:04:06 crc kubenswrapper[4813]: I1201 09:04:06.679240 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-external-api-1" Dec 01 09:04:06 crc kubenswrapper[4813]: I1201 09:04:06.710051 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-external-api-1" Dec 01 09:04:06 crc kubenswrapper[4813]: I1201 09:04:06.723550 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-external-api-1" Dec 01 09:04:06 crc kubenswrapper[4813]: I1201 09:04:06.979299 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:04:06 crc kubenswrapper[4813]: I1201 09:04:06.979409 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:04:07 crc kubenswrapper[4813]: I1201 09:04:07.022855 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:04:07 crc kubenswrapper[4813]: I1201 09:04:07.038171 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:04:07 crc kubenswrapper[4813]: I1201 09:04:07.142015 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-external-api-0" Dec 01 09:04:07 crc kubenswrapper[4813]: I1201 09:04:07.142079 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-external-api-0" Dec 01 09:04:07 crc kubenswrapper[4813]: I1201 09:04:07.184343 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-external-api-0" Dec 01 09:04:07 crc kubenswrapper[4813]: I1201 09:04:07.203204 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-external-api-0" Dec 01 09:04:07 crc kubenswrapper[4813]: I1201 09:04:07.386064 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-external-api-1" Dec 01 09:04:07 crc kubenswrapper[4813]: I1201 09:04:07.386111 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-external-api-1" Dec 01 09:04:07 crc kubenswrapper[4813]: I1201 09:04:07.386123 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-external-api-0" Dec 01 09:04:07 crc kubenswrapper[4813]: I1201 09:04:07.386132 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-external-api-0" Dec 01 09:04:07 crc kubenswrapper[4813]: I1201 09:04:07.386140 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:04:07 crc kubenswrapper[4813]: I1201 09:04:07.386150 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:04:09 crc kubenswrapper[4813]: I1201 09:04:09.382373 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-external-api-1" Dec 01 09:04:09 crc kubenswrapper[4813]: I1201 09:04:09.400904 4813 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 01 09:04:09 crc kubenswrapper[4813]: I1201 09:04:09.415540 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-external-api-1" Dec 01 09:04:09 crc kubenswrapper[4813]: I1201 09:04:09.457988 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-0"] Dec 01 09:04:09 crc kubenswrapper[4813]: I1201 09:04:09.458667 4813 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 01 09:04:09 crc kubenswrapper[4813]: I1201 09:04:09.458758 4813 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 01 09:04:09 crc kubenswrapper[4813]: I1201 09:04:09.459055 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-external-api-0" podUID="86363010-586b-4ee9-89e6-0549789e4f0a" containerName="glance-log" containerID="cri-o://d36d636ba6a08f873897903ab8eace8f40250273d37efbf7a92741f6fb28b059" gracePeriod=30 Dec 01 09:04:09 crc kubenswrapper[4813]: I1201 09:04:09.459254 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-external-api-0" podUID="86363010-586b-4ee9-89e6-0549789e4f0a" containerName="glance-httpd" containerID="cri-o://251f770953b244d0689cf3f4f5507594545fba685c12cfb8b522e7004dc4edd3" gracePeriod=30 Dec 01 09:04:09 crc kubenswrapper[4813]: I1201 09:04:09.475732 4813 prober.go:107] "Probe failed" probeType="Readiness" pod="glance-kuttl-tests/glance-default-external-api-0" podUID="86363010-586b-4ee9-89e6-0549789e4f0a" containerName="glance-httpd" probeResult="failure" output="Get \"http://10.217.0.146:9292/healthcheck\": EOF" Dec 01 09:04:09 crc kubenswrapper[4813]: I1201 09:04:09.485131 4813 prober.go:107] "Probe failed" probeType="Readiness" pod="glance-kuttl-tests/glance-default-external-api-0" podUID="86363010-586b-4ee9-89e6-0549789e4f0a" containerName="glance-httpd" probeResult="failure" output="Get \"http://10.217.0.146:9292/healthcheck\": EOF" Dec 01 09:04:09 crc kubenswrapper[4813]: I1201 09:04:09.485233 4813 prober.go:107] "Probe failed" probeType="Readiness" pod="glance-kuttl-tests/glance-default-external-api-0" podUID="86363010-586b-4ee9-89e6-0549789e4f0a" containerName="glance-log" probeResult="failure" output="Get \"http://10.217.0.146:9292/healthcheck\": EOF" Dec 01 09:04:09 crc kubenswrapper[4813]: I1201 09:04:09.544849 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:04:09 crc kubenswrapper[4813]: I1201 09:04:09.545404 4813 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 01 09:04:09 crc kubenswrapper[4813]: I1201 09:04:09.550048 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:04:10 crc kubenswrapper[4813]: I1201 09:04:10.411191 4813 generic.go:334] "Generic (PLEG): container finished" podID="86363010-586b-4ee9-89e6-0549789e4f0a" containerID="d36d636ba6a08f873897903ab8eace8f40250273d37efbf7a92741f6fb28b059" exitCode=143 Dec 01 09:04:10 crc kubenswrapper[4813]: I1201 09:04:10.412279 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-0" event={"ID":"86363010-586b-4ee9-89e6-0549789e4f0a","Type":"ContainerDied","Data":"d36d636ba6a08f873897903ab8eace8f40250273d37efbf7a92741f6fb28b059"} Dec 01 09:04:10 crc kubenswrapper[4813]: I1201 09:04:10.709585 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-internal-api-1" Dec 01 09:04:10 crc kubenswrapper[4813]: I1201 09:04:10.710099 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-internal-api-1" Dec 01 09:04:10 crc kubenswrapper[4813]: I1201 09:04:10.735018 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-internal-api-1" Dec 01 09:04:10 crc kubenswrapper[4813]: I1201 09:04:10.755575 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-internal-api-1" Dec 01 09:04:11 crc kubenswrapper[4813]: I1201 09:04:11.418210 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-internal-api-1" Dec 01 09:04:11 crc kubenswrapper[4813]: I1201 09:04:11.418523 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-internal-api-1" Dec 01 09:04:13 crc kubenswrapper[4813]: I1201 09:04:13.303083 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-internal-api-1" Dec 01 09:04:13 crc kubenswrapper[4813]: I1201 09:04:13.368179 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-internal-api-1" Dec 01 09:04:13 crc kubenswrapper[4813]: I1201 09:04:13.412346 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Dec 01 09:04:13 crc kubenswrapper[4813]: I1201 09:04:13.412774 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-internal-api-0" podUID="9402b27d-1b2c-4336-b2d7-e400772dc110" containerName="glance-log" containerID="cri-o://565f50a4ba9ebae84a0c5064170efc08f5c3b6a0e5dfb2a87873962b319f2134" gracePeriod=30 Dec 01 09:04:13 crc kubenswrapper[4813]: I1201 09:04:13.412863 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-internal-api-0" podUID="9402b27d-1b2c-4336-b2d7-e400772dc110" containerName="glance-httpd" containerID="cri-o://bcefcda3036d2b3710278d3a31141318f7fcff5e3c0317be8a09c5d9dab936e7" gracePeriod=30 Dec 01 09:04:14 crc kubenswrapper[4813]: I1201 09:04:14.474927 4813 generic.go:334] "Generic (PLEG): container finished" podID="9402b27d-1b2c-4336-b2d7-e400772dc110" containerID="565f50a4ba9ebae84a0c5064170efc08f5c3b6a0e5dfb2a87873962b319f2134" exitCode=143 Dec 01 09:04:14 crc kubenswrapper[4813]: I1201 09:04:14.475473 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"9402b27d-1b2c-4336-b2d7-e400772dc110","Type":"ContainerDied","Data":"565f50a4ba9ebae84a0c5064170efc08f5c3b6a0e5dfb2a87873962b319f2134"} Dec 01 09:04:14 crc kubenswrapper[4813]: I1201 09:04:14.924735 4813 prober.go:107] "Probe failed" probeType="Readiness" pod="glance-kuttl-tests/glance-default-external-api-0" podUID="86363010-586b-4ee9-89e6-0549789e4f0a" containerName="glance-log" probeResult="failure" output="Get \"http://10.217.0.146:9292/healthcheck\": read tcp 10.217.0.2:45848->10.217.0.146:9292: read: connection reset by peer" Dec 01 09:04:15 crc kubenswrapper[4813]: I1201 09:04:15.347815 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-0" Dec 01 09:04:15 crc kubenswrapper[4813]: I1201 09:04:15.488537 4813 generic.go:334] "Generic (PLEG): container finished" podID="86363010-586b-4ee9-89e6-0549789e4f0a" containerID="251f770953b244d0689cf3f4f5507594545fba685c12cfb8b522e7004dc4edd3" exitCode=0 Dec 01 09:04:15 crc kubenswrapper[4813]: I1201 09:04:15.488598 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-0" Dec 01 09:04:15 crc kubenswrapper[4813]: I1201 09:04:15.488625 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-0" event={"ID":"86363010-586b-4ee9-89e6-0549789e4f0a","Type":"ContainerDied","Data":"251f770953b244d0689cf3f4f5507594545fba685c12cfb8b522e7004dc4edd3"} Dec 01 09:04:15 crc kubenswrapper[4813]: I1201 09:04:15.488682 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-0" event={"ID":"86363010-586b-4ee9-89e6-0549789e4f0a","Type":"ContainerDied","Data":"daf104d88c1c187e3531e2cfd38aa8d0b377662713baed5c4bc193cf319b57d4"} Dec 01 09:04:15 crc kubenswrapper[4813]: I1201 09:04:15.488706 4813 scope.go:117] "RemoveContainer" containerID="251f770953b244d0689cf3f4f5507594545fba685c12cfb8b522e7004dc4edd3" Dec 01 09:04:15 crc kubenswrapper[4813]: I1201 09:04:15.518284 4813 scope.go:117] "RemoveContainer" containerID="d36d636ba6a08f873897903ab8eace8f40250273d37efbf7a92741f6fb28b059" Dec 01 09:04:15 crc kubenswrapper[4813]: I1201 09:04:15.536042 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rd8wb\" (UniqueName: \"kubernetes.io/projected/86363010-586b-4ee9-89e6-0549789e4f0a-kube-api-access-rd8wb\") pod \"86363010-586b-4ee9-89e6-0549789e4f0a\" (UID: \"86363010-586b-4ee9-89e6-0549789e4f0a\") " Dec 01 09:04:15 crc kubenswrapper[4813]: I1201 09:04:15.536158 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/86363010-586b-4ee9-89e6-0549789e4f0a-config-data\") pod \"86363010-586b-4ee9-89e6-0549789e4f0a\" (UID: \"86363010-586b-4ee9-89e6-0549789e4f0a\") " Dec 01 09:04:15 crc kubenswrapper[4813]: I1201 09:04:15.536197 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage20-crc\") pod \"86363010-586b-4ee9-89e6-0549789e4f0a\" (UID: \"86363010-586b-4ee9-89e6-0549789e4f0a\") " Dec 01 09:04:15 crc kubenswrapper[4813]: I1201 09:04:15.536227 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/86363010-586b-4ee9-89e6-0549789e4f0a-etc-iscsi\") pod \"86363010-586b-4ee9-89e6-0549789e4f0a\" (UID: \"86363010-586b-4ee9-89e6-0549789e4f0a\") " Dec 01 09:04:15 crc kubenswrapper[4813]: I1201 09:04:15.536259 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/86363010-586b-4ee9-89e6-0549789e4f0a-lib-modules\") pod \"86363010-586b-4ee9-89e6-0549789e4f0a\" (UID: \"86363010-586b-4ee9-89e6-0549789e4f0a\") " Dec 01 09:04:15 crc kubenswrapper[4813]: I1201 09:04:15.536278 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/86363010-586b-4ee9-89e6-0549789e4f0a-dev\") pod \"86363010-586b-4ee9-89e6-0549789e4f0a\" (UID: \"86363010-586b-4ee9-89e6-0549789e4f0a\") " Dec 01 09:04:15 crc kubenswrapper[4813]: I1201 09:04:15.536314 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/86363010-586b-4ee9-89e6-0549789e4f0a-sys\") pod \"86363010-586b-4ee9-89e6-0549789e4f0a\" (UID: \"86363010-586b-4ee9-89e6-0549789e4f0a\") " Dec 01 09:04:15 crc kubenswrapper[4813]: I1201 09:04:15.536368 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/86363010-586b-4ee9-89e6-0549789e4f0a-logs\") pod \"86363010-586b-4ee9-89e6-0549789e4f0a\" (UID: \"86363010-586b-4ee9-89e6-0549789e4f0a\") " Dec 01 09:04:15 crc kubenswrapper[4813]: I1201 09:04:15.536373 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/86363010-586b-4ee9-89e6-0549789e4f0a-etc-iscsi" (OuterVolumeSpecName: "etc-iscsi") pod "86363010-586b-4ee9-89e6-0549789e4f0a" (UID: "86363010-586b-4ee9-89e6-0549789e4f0a"). InnerVolumeSpecName "etc-iscsi". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 09:04:15 crc kubenswrapper[4813]: I1201 09:04:15.536397 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/86363010-586b-4ee9-89e6-0549789e4f0a-httpd-run\") pod \"86363010-586b-4ee9-89e6-0549789e4f0a\" (UID: \"86363010-586b-4ee9-89e6-0549789e4f0a\") " Dec 01 09:04:15 crc kubenswrapper[4813]: I1201 09:04:15.536474 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/86363010-586b-4ee9-89e6-0549789e4f0a-var-locks-brick\") pod \"86363010-586b-4ee9-89e6-0549789e4f0a\" (UID: \"86363010-586b-4ee9-89e6-0549789e4f0a\") " Dec 01 09:04:15 crc kubenswrapper[4813]: I1201 09:04:15.536511 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/86363010-586b-4ee9-89e6-0549789e4f0a-dev" (OuterVolumeSpecName: "dev") pod "86363010-586b-4ee9-89e6-0549789e4f0a" (UID: "86363010-586b-4ee9-89e6-0549789e4f0a"). InnerVolumeSpecName "dev". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 09:04:15 crc kubenswrapper[4813]: I1201 09:04:15.536559 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/86363010-586b-4ee9-89e6-0549789e4f0a-run\") pod \"86363010-586b-4ee9-89e6-0549789e4f0a\" (UID: \"86363010-586b-4ee9-89e6-0549789e4f0a\") " Dec 01 09:04:15 crc kubenswrapper[4813]: I1201 09:04:15.536606 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance-cache\" (UniqueName: \"kubernetes.io/local-volume/local-storage16-crc\") pod \"86363010-586b-4ee9-89e6-0549789e4f0a\" (UID: \"86363010-586b-4ee9-89e6-0549789e4f0a\") " Dec 01 09:04:15 crc kubenswrapper[4813]: I1201 09:04:15.536558 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/86363010-586b-4ee9-89e6-0549789e4f0a-sys" (OuterVolumeSpecName: "sys") pod "86363010-586b-4ee9-89e6-0549789e4f0a" (UID: "86363010-586b-4ee9-89e6-0549789e4f0a"). InnerVolumeSpecName "sys". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 09:04:15 crc kubenswrapper[4813]: I1201 09:04:15.536638 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/86363010-586b-4ee9-89e6-0549789e4f0a-etc-nvme\") pod \"86363010-586b-4ee9-89e6-0549789e4f0a\" (UID: \"86363010-586b-4ee9-89e6-0549789e4f0a\") " Dec 01 09:04:15 crc kubenswrapper[4813]: I1201 09:04:15.536559 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/86363010-586b-4ee9-89e6-0549789e4f0a-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "86363010-586b-4ee9-89e6-0549789e4f0a" (UID: "86363010-586b-4ee9-89e6-0549789e4f0a"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 09:04:15 crc kubenswrapper[4813]: I1201 09:04:15.536585 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/86363010-586b-4ee9-89e6-0549789e4f0a-var-locks-brick" (OuterVolumeSpecName: "var-locks-brick") pod "86363010-586b-4ee9-89e6-0549789e4f0a" (UID: "86363010-586b-4ee9-89e6-0549789e4f0a"). InnerVolumeSpecName "var-locks-brick". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 09:04:15 crc kubenswrapper[4813]: I1201 09:04:15.536607 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/86363010-586b-4ee9-89e6-0549789e4f0a-run" (OuterVolumeSpecName: "run") pod "86363010-586b-4ee9-89e6-0549789e4f0a" (UID: "86363010-586b-4ee9-89e6-0549789e4f0a"). InnerVolumeSpecName "run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 09:04:15 crc kubenswrapper[4813]: I1201 09:04:15.536683 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/86363010-586b-4ee9-89e6-0549789e4f0a-scripts\") pod \"86363010-586b-4ee9-89e6-0549789e4f0a\" (UID: \"86363010-586b-4ee9-89e6-0549789e4f0a\") " Dec 01 09:04:15 crc kubenswrapper[4813]: I1201 09:04:15.537046 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/86363010-586b-4ee9-89e6-0549789e4f0a-etc-nvme" (OuterVolumeSpecName: "etc-nvme") pod "86363010-586b-4ee9-89e6-0549789e4f0a" (UID: "86363010-586b-4ee9-89e6-0549789e4f0a"). InnerVolumeSpecName "etc-nvme". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 09:04:15 crc kubenswrapper[4813]: I1201 09:04:15.537212 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/86363010-586b-4ee9-89e6-0549789e4f0a-logs" (OuterVolumeSpecName: "logs") pod "86363010-586b-4ee9-89e6-0549789e4f0a" (UID: "86363010-586b-4ee9-89e6-0549789e4f0a"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 09:04:15 crc kubenswrapper[4813]: I1201 09:04:15.537324 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/86363010-586b-4ee9-89e6-0549789e4f0a-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "86363010-586b-4ee9-89e6-0549789e4f0a" (UID: "86363010-586b-4ee9-89e6-0549789e4f0a"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 09:04:15 crc kubenswrapper[4813]: I1201 09:04:15.537656 4813 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/86363010-586b-4ee9-89e6-0549789e4f0a-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 01 09:04:15 crc kubenswrapper[4813]: I1201 09:04:15.537673 4813 reconciler_common.go:293] "Volume detached for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/86363010-586b-4ee9-89e6-0549789e4f0a-var-locks-brick\") on node \"crc\" DevicePath \"\"" Dec 01 09:04:15 crc kubenswrapper[4813]: I1201 09:04:15.537685 4813 reconciler_common.go:293] "Volume detached for volume \"run\" (UniqueName: \"kubernetes.io/host-path/86363010-586b-4ee9-89e6-0549789e4f0a-run\") on node \"crc\" DevicePath \"\"" Dec 01 09:04:15 crc kubenswrapper[4813]: I1201 09:04:15.537694 4813 reconciler_common.go:293] "Volume detached for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/86363010-586b-4ee9-89e6-0549789e4f0a-etc-nvme\") on node \"crc\" DevicePath \"\"" Dec 01 09:04:15 crc kubenswrapper[4813]: I1201 09:04:15.537702 4813 reconciler_common.go:293] "Volume detached for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/86363010-586b-4ee9-89e6-0549789e4f0a-etc-iscsi\") on node \"crc\" DevicePath \"\"" Dec 01 09:04:15 crc kubenswrapper[4813]: I1201 09:04:15.537710 4813 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/86363010-586b-4ee9-89e6-0549789e4f0a-lib-modules\") on node \"crc\" DevicePath \"\"" Dec 01 09:04:15 crc kubenswrapper[4813]: I1201 09:04:15.537718 4813 reconciler_common.go:293] "Volume detached for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/86363010-586b-4ee9-89e6-0549789e4f0a-dev\") on node \"crc\" DevicePath \"\"" Dec 01 09:04:15 crc kubenswrapper[4813]: I1201 09:04:15.537727 4813 reconciler_common.go:293] "Volume detached for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/86363010-586b-4ee9-89e6-0549789e4f0a-sys\") on node \"crc\" DevicePath \"\"" Dec 01 09:04:15 crc kubenswrapper[4813]: I1201 09:04:15.537735 4813 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/86363010-586b-4ee9-89e6-0549789e4f0a-logs\") on node \"crc\" DevicePath \"\"" Dec 01 09:04:15 crc kubenswrapper[4813]: I1201 09:04:15.543695 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage16-crc" (OuterVolumeSpecName: "glance-cache") pod "86363010-586b-4ee9-89e6-0549789e4f0a" (UID: "86363010-586b-4ee9-89e6-0549789e4f0a"). InnerVolumeSpecName "local-storage16-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 01 09:04:15 crc kubenswrapper[4813]: I1201 09:04:15.544602 4813 scope.go:117] "RemoveContainer" containerID="251f770953b244d0689cf3f4f5507594545fba685c12cfb8b522e7004dc4edd3" Dec 01 09:04:15 crc kubenswrapper[4813]: I1201 09:04:15.544637 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage20-crc" (OuterVolumeSpecName: "glance") pod "86363010-586b-4ee9-89e6-0549789e4f0a" (UID: "86363010-586b-4ee9-89e6-0549789e4f0a"). InnerVolumeSpecName "local-storage20-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 01 09:04:15 crc kubenswrapper[4813]: I1201 09:04:15.545083 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/86363010-586b-4ee9-89e6-0549789e4f0a-scripts" (OuterVolumeSpecName: "scripts") pod "86363010-586b-4ee9-89e6-0549789e4f0a" (UID: "86363010-586b-4ee9-89e6-0549789e4f0a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:04:15 crc kubenswrapper[4813]: I1201 09:04:15.547225 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/86363010-586b-4ee9-89e6-0549789e4f0a-kube-api-access-rd8wb" (OuterVolumeSpecName: "kube-api-access-rd8wb") pod "86363010-586b-4ee9-89e6-0549789e4f0a" (UID: "86363010-586b-4ee9-89e6-0549789e4f0a"). InnerVolumeSpecName "kube-api-access-rd8wb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:04:15 crc kubenswrapper[4813]: E1201 09:04:15.547285 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"251f770953b244d0689cf3f4f5507594545fba685c12cfb8b522e7004dc4edd3\": container with ID starting with 251f770953b244d0689cf3f4f5507594545fba685c12cfb8b522e7004dc4edd3 not found: ID does not exist" containerID="251f770953b244d0689cf3f4f5507594545fba685c12cfb8b522e7004dc4edd3" Dec 01 09:04:15 crc kubenswrapper[4813]: I1201 09:04:15.547336 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"251f770953b244d0689cf3f4f5507594545fba685c12cfb8b522e7004dc4edd3"} err="failed to get container status \"251f770953b244d0689cf3f4f5507594545fba685c12cfb8b522e7004dc4edd3\": rpc error: code = NotFound desc = could not find container \"251f770953b244d0689cf3f4f5507594545fba685c12cfb8b522e7004dc4edd3\": container with ID starting with 251f770953b244d0689cf3f4f5507594545fba685c12cfb8b522e7004dc4edd3 not found: ID does not exist" Dec 01 09:04:15 crc kubenswrapper[4813]: I1201 09:04:15.547372 4813 scope.go:117] "RemoveContainer" containerID="d36d636ba6a08f873897903ab8eace8f40250273d37efbf7a92741f6fb28b059" Dec 01 09:04:15 crc kubenswrapper[4813]: E1201 09:04:15.547772 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d36d636ba6a08f873897903ab8eace8f40250273d37efbf7a92741f6fb28b059\": container with ID starting with d36d636ba6a08f873897903ab8eace8f40250273d37efbf7a92741f6fb28b059 not found: ID does not exist" containerID="d36d636ba6a08f873897903ab8eace8f40250273d37efbf7a92741f6fb28b059" Dec 01 09:04:15 crc kubenswrapper[4813]: I1201 09:04:15.547812 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d36d636ba6a08f873897903ab8eace8f40250273d37efbf7a92741f6fb28b059"} err="failed to get container status \"d36d636ba6a08f873897903ab8eace8f40250273d37efbf7a92741f6fb28b059\": rpc error: code = NotFound desc = could not find container \"d36d636ba6a08f873897903ab8eace8f40250273d37efbf7a92741f6fb28b059\": container with ID starting with d36d636ba6a08f873897903ab8eace8f40250273d37efbf7a92741f6fb28b059 not found: ID does not exist" Dec 01 09:04:15 crc kubenswrapper[4813]: I1201 09:04:15.584834 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/86363010-586b-4ee9-89e6-0549789e4f0a-config-data" (OuterVolumeSpecName: "config-data") pod "86363010-586b-4ee9-89e6-0549789e4f0a" (UID: "86363010-586b-4ee9-89e6-0549789e4f0a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:04:15 crc kubenswrapper[4813]: I1201 09:04:15.639301 4813 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage16-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage16-crc\") on node \"crc\" " Dec 01 09:04:15 crc kubenswrapper[4813]: I1201 09:04:15.639345 4813 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/86363010-586b-4ee9-89e6-0549789e4f0a-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 09:04:15 crc kubenswrapper[4813]: I1201 09:04:15.639360 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rd8wb\" (UniqueName: \"kubernetes.io/projected/86363010-586b-4ee9-89e6-0549789e4f0a-kube-api-access-rd8wb\") on node \"crc\" DevicePath \"\"" Dec 01 09:04:15 crc kubenswrapper[4813]: I1201 09:04:15.639374 4813 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/86363010-586b-4ee9-89e6-0549789e4f0a-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 09:04:15 crc kubenswrapper[4813]: I1201 09:04:15.639394 4813 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage20-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage20-crc\") on node \"crc\" " Dec 01 09:04:15 crc kubenswrapper[4813]: I1201 09:04:15.654233 4813 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage16-crc" (UniqueName: "kubernetes.io/local-volume/local-storage16-crc") on node "crc" Dec 01 09:04:15 crc kubenswrapper[4813]: I1201 09:04:15.656491 4813 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage20-crc" (UniqueName: "kubernetes.io/local-volume/local-storage20-crc") on node "crc" Dec 01 09:04:15 crc kubenswrapper[4813]: I1201 09:04:15.741730 4813 reconciler_common.go:293] "Volume detached for volume \"local-storage20-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage20-crc\") on node \"crc\" DevicePath \"\"" Dec 01 09:04:15 crc kubenswrapper[4813]: I1201 09:04:15.741780 4813 reconciler_common.go:293] "Volume detached for volume \"local-storage16-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage16-crc\") on node \"crc\" DevicePath \"\"" Dec 01 09:04:15 crc kubenswrapper[4813]: I1201 09:04:15.821386 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-0"] Dec 01 09:04:15 crc kubenswrapper[4813]: I1201 09:04:15.826521 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-0"] Dec 01 09:04:15 crc kubenswrapper[4813]: I1201 09:04:15.847925 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-external-api-0"] Dec 01 09:04:15 crc kubenswrapper[4813]: E1201 09:04:15.848280 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="86363010-586b-4ee9-89e6-0549789e4f0a" containerName="glance-httpd" Dec 01 09:04:15 crc kubenswrapper[4813]: I1201 09:04:15.848293 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="86363010-586b-4ee9-89e6-0549789e4f0a" containerName="glance-httpd" Dec 01 09:04:15 crc kubenswrapper[4813]: E1201 09:04:15.848315 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="86363010-586b-4ee9-89e6-0549789e4f0a" containerName="glance-log" Dec 01 09:04:15 crc kubenswrapper[4813]: I1201 09:04:15.848321 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="86363010-586b-4ee9-89e6-0549789e4f0a" containerName="glance-log" Dec 01 09:04:15 crc kubenswrapper[4813]: I1201 09:04:15.848444 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="86363010-586b-4ee9-89e6-0549789e4f0a" containerName="glance-httpd" Dec 01 09:04:15 crc kubenswrapper[4813]: I1201 09:04:15.848453 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="86363010-586b-4ee9-89e6-0549789e4f0a" containerName="glance-log" Dec 01 09:04:15 crc kubenswrapper[4813]: I1201 09:04:15.849257 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-0" Dec 01 09:04:15 crc kubenswrapper[4813]: I1201 09:04:15.912265 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-0"] Dec 01 09:04:15 crc kubenswrapper[4813]: I1201 09:04:15.944176 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d6f4323d-561b-4725-a1fe-140db5f6a33d-scripts\") pod \"glance-default-external-api-0\" (UID: \"d6f4323d-561b-4725-a1fe-140db5f6a33d\") " pod="glance-kuttl-tests/glance-default-external-api-0" Dec 01 09:04:15 crc kubenswrapper[4813]: I1201 09:04:15.944231 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/d6f4323d-561b-4725-a1fe-140db5f6a33d-var-locks-brick\") pod \"glance-default-external-api-0\" (UID: \"d6f4323d-561b-4725-a1fe-140db5f6a33d\") " pod="glance-kuttl-tests/glance-default-external-api-0" Dec 01 09:04:15 crc kubenswrapper[4813]: I1201 09:04:15.944265 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/d6f4323d-561b-4725-a1fe-140db5f6a33d-run\") pod \"glance-default-external-api-0\" (UID: \"d6f4323d-561b-4725-a1fe-140db5f6a33d\") " pod="glance-kuttl-tests/glance-default-external-api-0" Dec 01 09:04:15 crc kubenswrapper[4813]: I1201 09:04:15.944281 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/d6f4323d-561b-4725-a1fe-140db5f6a33d-sys\") pod \"glance-default-external-api-0\" (UID: \"d6f4323d-561b-4725-a1fe-140db5f6a33d\") " pod="glance-kuttl-tests/glance-default-external-api-0" Dec 01 09:04:15 crc kubenswrapper[4813]: I1201 09:04:15.944319 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jtfbh\" (UniqueName: \"kubernetes.io/projected/d6f4323d-561b-4725-a1fe-140db5f6a33d-kube-api-access-jtfbh\") pod \"glance-default-external-api-0\" (UID: \"d6f4323d-561b-4725-a1fe-140db5f6a33d\") " pod="glance-kuttl-tests/glance-default-external-api-0" Dec 01 09:04:15 crc kubenswrapper[4813]: I1201 09:04:15.944345 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage16-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage16-crc\") pod \"glance-default-external-api-0\" (UID: \"d6f4323d-561b-4725-a1fe-140db5f6a33d\") " pod="glance-kuttl-tests/glance-default-external-api-0" Dec 01 09:04:15 crc kubenswrapper[4813]: I1201 09:04:15.944377 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage20-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage20-crc\") pod \"glance-default-external-api-0\" (UID: \"d6f4323d-561b-4725-a1fe-140db5f6a33d\") " pod="glance-kuttl-tests/glance-default-external-api-0" Dec 01 09:04:15 crc kubenswrapper[4813]: I1201 09:04:15.944633 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/d6f4323d-561b-4725-a1fe-140db5f6a33d-etc-iscsi\") pod \"glance-default-external-api-0\" (UID: \"d6f4323d-561b-4725-a1fe-140db5f6a33d\") " pod="glance-kuttl-tests/glance-default-external-api-0" Dec 01 09:04:15 crc kubenswrapper[4813]: I1201 09:04:15.944657 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/d6f4323d-561b-4725-a1fe-140db5f6a33d-dev\") pod \"glance-default-external-api-0\" (UID: \"d6f4323d-561b-4725-a1fe-140db5f6a33d\") " pod="glance-kuttl-tests/glance-default-external-api-0" Dec 01 09:04:15 crc kubenswrapper[4813]: I1201 09:04:15.944684 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/d6f4323d-561b-4725-a1fe-140db5f6a33d-etc-nvme\") pod \"glance-default-external-api-0\" (UID: \"d6f4323d-561b-4725-a1fe-140db5f6a33d\") " pod="glance-kuttl-tests/glance-default-external-api-0" Dec 01 09:04:15 crc kubenswrapper[4813]: I1201 09:04:15.944752 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d6f4323d-561b-4725-a1fe-140db5f6a33d-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"d6f4323d-561b-4725-a1fe-140db5f6a33d\") " pod="glance-kuttl-tests/glance-default-external-api-0" Dec 01 09:04:15 crc kubenswrapper[4813]: I1201 09:04:15.944787 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d6f4323d-561b-4725-a1fe-140db5f6a33d-logs\") pod \"glance-default-external-api-0\" (UID: \"d6f4323d-561b-4725-a1fe-140db5f6a33d\") " pod="glance-kuttl-tests/glance-default-external-api-0" Dec 01 09:04:15 crc kubenswrapper[4813]: I1201 09:04:15.944806 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d6f4323d-561b-4725-a1fe-140db5f6a33d-config-data\") pod \"glance-default-external-api-0\" (UID: \"d6f4323d-561b-4725-a1fe-140db5f6a33d\") " pod="glance-kuttl-tests/glance-default-external-api-0" Dec 01 09:04:15 crc kubenswrapper[4813]: I1201 09:04:15.944833 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/d6f4323d-561b-4725-a1fe-140db5f6a33d-lib-modules\") pod \"glance-default-external-api-0\" (UID: \"d6f4323d-561b-4725-a1fe-140db5f6a33d\") " pod="glance-kuttl-tests/glance-default-external-api-0" Dec 01 09:04:16 crc kubenswrapper[4813]: I1201 09:04:16.046259 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d6f4323d-561b-4725-a1fe-140db5f6a33d-scripts\") pod \"glance-default-external-api-0\" (UID: \"d6f4323d-561b-4725-a1fe-140db5f6a33d\") " pod="glance-kuttl-tests/glance-default-external-api-0" Dec 01 09:04:16 crc kubenswrapper[4813]: I1201 09:04:16.046347 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/d6f4323d-561b-4725-a1fe-140db5f6a33d-var-locks-brick\") pod \"glance-default-external-api-0\" (UID: \"d6f4323d-561b-4725-a1fe-140db5f6a33d\") " pod="glance-kuttl-tests/glance-default-external-api-0" Dec 01 09:04:16 crc kubenswrapper[4813]: I1201 09:04:16.046384 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/d6f4323d-561b-4725-a1fe-140db5f6a33d-run\") pod \"glance-default-external-api-0\" (UID: \"d6f4323d-561b-4725-a1fe-140db5f6a33d\") " pod="glance-kuttl-tests/glance-default-external-api-0" Dec 01 09:04:16 crc kubenswrapper[4813]: I1201 09:04:16.046407 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/d6f4323d-561b-4725-a1fe-140db5f6a33d-sys\") pod \"glance-default-external-api-0\" (UID: \"d6f4323d-561b-4725-a1fe-140db5f6a33d\") " pod="glance-kuttl-tests/glance-default-external-api-0" Dec 01 09:04:16 crc kubenswrapper[4813]: I1201 09:04:16.046428 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jtfbh\" (UniqueName: \"kubernetes.io/projected/d6f4323d-561b-4725-a1fe-140db5f6a33d-kube-api-access-jtfbh\") pod \"glance-default-external-api-0\" (UID: \"d6f4323d-561b-4725-a1fe-140db5f6a33d\") " pod="glance-kuttl-tests/glance-default-external-api-0" Dec 01 09:04:16 crc kubenswrapper[4813]: I1201 09:04:16.046462 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage16-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage16-crc\") pod \"glance-default-external-api-0\" (UID: \"d6f4323d-561b-4725-a1fe-140db5f6a33d\") " pod="glance-kuttl-tests/glance-default-external-api-0" Dec 01 09:04:16 crc kubenswrapper[4813]: I1201 09:04:16.046500 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage20-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage20-crc\") pod \"glance-default-external-api-0\" (UID: \"d6f4323d-561b-4725-a1fe-140db5f6a33d\") " pod="glance-kuttl-tests/glance-default-external-api-0" Dec 01 09:04:16 crc kubenswrapper[4813]: I1201 09:04:16.046508 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/d6f4323d-561b-4725-a1fe-140db5f6a33d-var-locks-brick\") pod \"glance-default-external-api-0\" (UID: \"d6f4323d-561b-4725-a1fe-140db5f6a33d\") " pod="glance-kuttl-tests/glance-default-external-api-0" Dec 01 09:04:16 crc kubenswrapper[4813]: I1201 09:04:16.046579 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/d6f4323d-561b-4725-a1fe-140db5f6a33d-etc-iscsi\") pod \"glance-default-external-api-0\" (UID: \"d6f4323d-561b-4725-a1fe-140db5f6a33d\") " pod="glance-kuttl-tests/glance-default-external-api-0" Dec 01 09:04:16 crc kubenswrapper[4813]: I1201 09:04:16.046547 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/d6f4323d-561b-4725-a1fe-140db5f6a33d-etc-iscsi\") pod \"glance-default-external-api-0\" (UID: \"d6f4323d-561b-4725-a1fe-140db5f6a33d\") " pod="glance-kuttl-tests/glance-default-external-api-0" Dec 01 09:04:16 crc kubenswrapper[4813]: I1201 09:04:16.046612 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/d6f4323d-561b-4725-a1fe-140db5f6a33d-sys\") pod \"glance-default-external-api-0\" (UID: \"d6f4323d-561b-4725-a1fe-140db5f6a33d\") " pod="glance-kuttl-tests/glance-default-external-api-0" Dec 01 09:04:16 crc kubenswrapper[4813]: I1201 09:04:16.046645 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/d6f4323d-561b-4725-a1fe-140db5f6a33d-dev\") pod \"glance-default-external-api-0\" (UID: \"d6f4323d-561b-4725-a1fe-140db5f6a33d\") " pod="glance-kuttl-tests/glance-default-external-api-0" Dec 01 09:04:16 crc kubenswrapper[4813]: I1201 09:04:16.046552 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/d6f4323d-561b-4725-a1fe-140db5f6a33d-run\") pod \"glance-default-external-api-0\" (UID: \"d6f4323d-561b-4725-a1fe-140db5f6a33d\") " pod="glance-kuttl-tests/glance-default-external-api-0" Dec 01 09:04:16 crc kubenswrapper[4813]: I1201 09:04:16.046714 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/d6f4323d-561b-4725-a1fe-140db5f6a33d-etc-nvme\") pod \"glance-default-external-api-0\" (UID: \"d6f4323d-561b-4725-a1fe-140db5f6a33d\") " pod="glance-kuttl-tests/glance-default-external-api-0" Dec 01 09:04:16 crc kubenswrapper[4813]: I1201 09:04:16.046717 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/d6f4323d-561b-4725-a1fe-140db5f6a33d-dev\") pod \"glance-default-external-api-0\" (UID: \"d6f4323d-561b-4725-a1fe-140db5f6a33d\") " pod="glance-kuttl-tests/glance-default-external-api-0" Dec 01 09:04:16 crc kubenswrapper[4813]: I1201 09:04:16.046804 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/d6f4323d-561b-4725-a1fe-140db5f6a33d-etc-nvme\") pod \"glance-default-external-api-0\" (UID: \"d6f4323d-561b-4725-a1fe-140db5f6a33d\") " pod="glance-kuttl-tests/glance-default-external-api-0" Dec 01 09:04:16 crc kubenswrapper[4813]: I1201 09:04:16.046824 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d6f4323d-561b-4725-a1fe-140db5f6a33d-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"d6f4323d-561b-4725-a1fe-140db5f6a33d\") " pod="glance-kuttl-tests/glance-default-external-api-0" Dec 01 09:04:16 crc kubenswrapper[4813]: I1201 09:04:16.046853 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d6f4323d-561b-4725-a1fe-140db5f6a33d-logs\") pod \"glance-default-external-api-0\" (UID: \"d6f4323d-561b-4725-a1fe-140db5f6a33d\") " pod="glance-kuttl-tests/glance-default-external-api-0" Dec 01 09:04:16 crc kubenswrapper[4813]: I1201 09:04:16.046880 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d6f4323d-561b-4725-a1fe-140db5f6a33d-config-data\") pod \"glance-default-external-api-0\" (UID: \"d6f4323d-561b-4725-a1fe-140db5f6a33d\") " pod="glance-kuttl-tests/glance-default-external-api-0" Dec 01 09:04:16 crc kubenswrapper[4813]: I1201 09:04:16.046941 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/d6f4323d-561b-4725-a1fe-140db5f6a33d-lib-modules\") pod \"glance-default-external-api-0\" (UID: \"d6f4323d-561b-4725-a1fe-140db5f6a33d\") " pod="glance-kuttl-tests/glance-default-external-api-0" Dec 01 09:04:16 crc kubenswrapper[4813]: I1201 09:04:16.046938 4813 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage20-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage20-crc\") pod \"glance-default-external-api-0\" (UID: \"d6f4323d-561b-4725-a1fe-140db5f6a33d\") device mount path \"/mnt/openstack/pv20\"" pod="glance-kuttl-tests/glance-default-external-api-0" Dec 01 09:04:16 crc kubenswrapper[4813]: I1201 09:04:16.047110 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/d6f4323d-561b-4725-a1fe-140db5f6a33d-lib-modules\") pod \"glance-default-external-api-0\" (UID: \"d6f4323d-561b-4725-a1fe-140db5f6a33d\") " pod="glance-kuttl-tests/glance-default-external-api-0" Dec 01 09:04:16 crc kubenswrapper[4813]: I1201 09:04:16.047823 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d6f4323d-561b-4725-a1fe-140db5f6a33d-logs\") pod \"glance-default-external-api-0\" (UID: \"d6f4323d-561b-4725-a1fe-140db5f6a33d\") " pod="glance-kuttl-tests/glance-default-external-api-0" Dec 01 09:04:16 crc kubenswrapper[4813]: I1201 09:04:16.047888 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d6f4323d-561b-4725-a1fe-140db5f6a33d-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"d6f4323d-561b-4725-a1fe-140db5f6a33d\") " pod="glance-kuttl-tests/glance-default-external-api-0" Dec 01 09:04:16 crc kubenswrapper[4813]: I1201 09:04:16.048183 4813 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage16-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage16-crc\") pod \"glance-default-external-api-0\" (UID: \"d6f4323d-561b-4725-a1fe-140db5f6a33d\") device mount path \"/mnt/openstack/pv16\"" pod="glance-kuttl-tests/glance-default-external-api-0" Dec 01 09:04:16 crc kubenswrapper[4813]: I1201 09:04:16.053039 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d6f4323d-561b-4725-a1fe-140db5f6a33d-config-data\") pod \"glance-default-external-api-0\" (UID: \"d6f4323d-561b-4725-a1fe-140db5f6a33d\") " pod="glance-kuttl-tests/glance-default-external-api-0" Dec 01 09:04:16 crc kubenswrapper[4813]: I1201 09:04:16.058522 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d6f4323d-561b-4725-a1fe-140db5f6a33d-scripts\") pod \"glance-default-external-api-0\" (UID: \"d6f4323d-561b-4725-a1fe-140db5f6a33d\") " pod="glance-kuttl-tests/glance-default-external-api-0" Dec 01 09:04:16 crc kubenswrapper[4813]: I1201 09:04:16.067461 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jtfbh\" (UniqueName: \"kubernetes.io/projected/d6f4323d-561b-4725-a1fe-140db5f6a33d-kube-api-access-jtfbh\") pod \"glance-default-external-api-0\" (UID: \"d6f4323d-561b-4725-a1fe-140db5f6a33d\") " pod="glance-kuttl-tests/glance-default-external-api-0" Dec 01 09:04:16 crc kubenswrapper[4813]: I1201 09:04:16.073557 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage20-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage20-crc\") pod \"glance-default-external-api-0\" (UID: \"d6f4323d-561b-4725-a1fe-140db5f6a33d\") " pod="glance-kuttl-tests/glance-default-external-api-0" Dec 01 09:04:16 crc kubenswrapper[4813]: I1201 09:04:16.078162 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage16-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage16-crc\") pod \"glance-default-external-api-0\" (UID: \"d6f4323d-561b-4725-a1fe-140db5f6a33d\") " pod="glance-kuttl-tests/glance-default-external-api-0" Dec 01 09:04:16 crc kubenswrapper[4813]: I1201 09:04:16.216523 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-0" Dec 01 09:04:16 crc kubenswrapper[4813]: I1201 09:04:16.413598 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="86363010-586b-4ee9-89e6-0549789e4f0a" path="/var/lib/kubelet/pods/86363010-586b-4ee9-89e6-0549789e4f0a/volumes" Dec 01 09:04:16 crc kubenswrapper[4813]: I1201 09:04:16.662673 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-0"] Dec 01 09:04:16 crc kubenswrapper[4813]: W1201 09:04:16.668389 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd6f4323d_561b_4725_a1fe_140db5f6a33d.slice/crio-10b31d489e5914c49ad6b4a668c2fa1494515f79057c148aeb333ecc1b27a574 WatchSource:0}: Error finding container 10b31d489e5914c49ad6b4a668c2fa1494515f79057c148aeb333ecc1b27a574: Status 404 returned error can't find the container with id 10b31d489e5914c49ad6b4a668c2fa1494515f79057c148aeb333ecc1b27a574 Dec 01 09:04:16 crc kubenswrapper[4813]: I1201 09:04:16.911799 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:04:16 crc kubenswrapper[4813]: I1201 09:04:16.967356 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/9402b27d-1b2c-4336-b2d7-e400772dc110-lib-modules\") pod \"9402b27d-1b2c-4336-b2d7-e400772dc110\" (UID: \"9402b27d-1b2c-4336-b2d7-e400772dc110\") " Dec 01 09:04:16 crc kubenswrapper[4813]: I1201 09:04:16.967424 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9402b27d-1b2c-4336-b2d7-e400772dc110-config-data\") pod \"9402b27d-1b2c-4336-b2d7-e400772dc110\" (UID: \"9402b27d-1b2c-4336-b2d7-e400772dc110\") " Dec 01 09:04:16 crc kubenswrapper[4813]: I1201 09:04:16.967467 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgxvz\" (UniqueName: \"kubernetes.io/projected/9402b27d-1b2c-4336-b2d7-e400772dc110-kube-api-access-zgxvz\") pod \"9402b27d-1b2c-4336-b2d7-e400772dc110\" (UID: \"9402b27d-1b2c-4336-b2d7-e400772dc110\") " Dec 01 09:04:16 crc kubenswrapper[4813]: I1201 09:04:16.967503 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/9402b27d-1b2c-4336-b2d7-e400772dc110-sys\") pod \"9402b27d-1b2c-4336-b2d7-e400772dc110\" (UID: \"9402b27d-1b2c-4336-b2d7-e400772dc110\") " Dec 01 09:04:16 crc kubenswrapper[4813]: I1201 09:04:16.967528 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9402b27d-1b2c-4336-b2d7-e400772dc110-scripts\") pod \"9402b27d-1b2c-4336-b2d7-e400772dc110\" (UID: \"9402b27d-1b2c-4336-b2d7-e400772dc110\") " Dec 01 09:04:16 crc kubenswrapper[4813]: I1201 09:04:16.967562 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9402b27d-1b2c-4336-b2d7-e400772dc110-logs\") pod \"9402b27d-1b2c-4336-b2d7-e400772dc110\" (UID: \"9402b27d-1b2c-4336-b2d7-e400772dc110\") " Dec 01 09:04:16 crc kubenswrapper[4813]: I1201 09:04:16.967695 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance-cache\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"9402b27d-1b2c-4336-b2d7-e400772dc110\" (UID: \"9402b27d-1b2c-4336-b2d7-e400772dc110\") " Dec 01 09:04:16 crc kubenswrapper[4813]: I1201 09:04:16.967729 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/9402b27d-1b2c-4336-b2d7-e400772dc110-etc-iscsi\") pod \"9402b27d-1b2c-4336-b2d7-e400772dc110\" (UID: \"9402b27d-1b2c-4336-b2d7-e400772dc110\") " Dec 01 09:04:16 crc kubenswrapper[4813]: I1201 09:04:16.967758 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/9402b27d-1b2c-4336-b2d7-e400772dc110-run\") pod \"9402b27d-1b2c-4336-b2d7-e400772dc110\" (UID: \"9402b27d-1b2c-4336-b2d7-e400772dc110\") " Dec 01 09:04:16 crc kubenswrapper[4813]: I1201 09:04:16.967834 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/9402b27d-1b2c-4336-b2d7-e400772dc110-var-locks-brick\") pod \"9402b27d-1b2c-4336-b2d7-e400772dc110\" (UID: \"9402b27d-1b2c-4336-b2d7-e400772dc110\") " Dec 01 09:04:16 crc kubenswrapper[4813]: I1201 09:04:16.967857 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/9402b27d-1b2c-4336-b2d7-e400772dc110-dev\") pod \"9402b27d-1b2c-4336-b2d7-e400772dc110\" (UID: \"9402b27d-1b2c-4336-b2d7-e400772dc110\") " Dec 01 09:04:16 crc kubenswrapper[4813]: I1201 09:04:16.967922 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/9402b27d-1b2c-4336-b2d7-e400772dc110-etc-nvme\") pod \"9402b27d-1b2c-4336-b2d7-e400772dc110\" (UID: \"9402b27d-1b2c-4336-b2d7-e400772dc110\") " Dec 01 09:04:16 crc kubenswrapper[4813]: I1201 09:04:16.967958 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/9402b27d-1b2c-4336-b2d7-e400772dc110-httpd-run\") pod \"9402b27d-1b2c-4336-b2d7-e400772dc110\" (UID: \"9402b27d-1b2c-4336-b2d7-e400772dc110\") " Dec 01 09:04:16 crc kubenswrapper[4813]: I1201 09:04:16.968891 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9402b27d-1b2c-4336-b2d7-e400772dc110-logs" (OuterVolumeSpecName: "logs") pod "9402b27d-1b2c-4336-b2d7-e400772dc110" (UID: "9402b27d-1b2c-4336-b2d7-e400772dc110"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 09:04:16 crc kubenswrapper[4813]: I1201 09:04:16.969003 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9402b27d-1b2c-4336-b2d7-e400772dc110-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "9402b27d-1b2c-4336-b2d7-e400772dc110" (UID: "9402b27d-1b2c-4336-b2d7-e400772dc110"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 09:04:16 crc kubenswrapper[4813]: I1201 09:04:16.969134 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9402b27d-1b2c-4336-b2d7-e400772dc110-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "9402b27d-1b2c-4336-b2d7-e400772dc110" (UID: "9402b27d-1b2c-4336-b2d7-e400772dc110"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 09:04:16 crc kubenswrapper[4813]: I1201 09:04:16.969362 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9402b27d-1b2c-4336-b2d7-e400772dc110-sys" (OuterVolumeSpecName: "sys") pod "9402b27d-1b2c-4336-b2d7-e400772dc110" (UID: "9402b27d-1b2c-4336-b2d7-e400772dc110"). InnerVolumeSpecName "sys". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 09:04:16 crc kubenswrapper[4813]: I1201 09:04:16.969581 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9402b27d-1b2c-4336-b2d7-e400772dc110-etc-iscsi" (OuterVolumeSpecName: "etc-iscsi") pod "9402b27d-1b2c-4336-b2d7-e400772dc110" (UID: "9402b27d-1b2c-4336-b2d7-e400772dc110"). InnerVolumeSpecName "etc-iscsi". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 09:04:16 crc kubenswrapper[4813]: I1201 09:04:16.969613 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9402b27d-1b2c-4336-b2d7-e400772dc110-run" (OuterVolumeSpecName: "run") pod "9402b27d-1b2c-4336-b2d7-e400772dc110" (UID: "9402b27d-1b2c-4336-b2d7-e400772dc110"). InnerVolumeSpecName "run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 09:04:16 crc kubenswrapper[4813]: I1201 09:04:16.969593 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9402b27d-1b2c-4336-b2d7-e400772dc110-var-locks-brick" (OuterVolumeSpecName: "var-locks-brick") pod "9402b27d-1b2c-4336-b2d7-e400772dc110" (UID: "9402b27d-1b2c-4336-b2d7-e400772dc110"). InnerVolumeSpecName "var-locks-brick". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 09:04:16 crc kubenswrapper[4813]: I1201 09:04:16.969691 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9402b27d-1b2c-4336-b2d7-e400772dc110-etc-nvme" (OuterVolumeSpecName: "etc-nvme") pod "9402b27d-1b2c-4336-b2d7-e400772dc110" (UID: "9402b27d-1b2c-4336-b2d7-e400772dc110"). InnerVolumeSpecName "etc-nvme". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 09:04:16 crc kubenswrapper[4813]: I1201 09:04:16.969829 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9402b27d-1b2c-4336-b2d7-e400772dc110-dev" (OuterVolumeSpecName: "dev") pod "9402b27d-1b2c-4336-b2d7-e400772dc110" (UID: "9402b27d-1b2c-4336-b2d7-e400772dc110"). InnerVolumeSpecName "dev". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 09:04:16 crc kubenswrapper[4813]: I1201 09:04:16.974009 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9402b27d-1b2c-4336-b2d7-e400772dc110-kube-api-access-zgxvz" (OuterVolumeSpecName: "kube-api-access-zgxvz") pod "9402b27d-1b2c-4336-b2d7-e400772dc110" (UID: "9402b27d-1b2c-4336-b2d7-e400772dc110"). InnerVolumeSpecName "kube-api-access-zgxvz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:04:16 crc kubenswrapper[4813]: I1201 09:04:16.979155 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage03-crc" (OuterVolumeSpecName: "glance-cache") pod "9402b27d-1b2c-4336-b2d7-e400772dc110" (UID: "9402b27d-1b2c-4336-b2d7-e400772dc110"). InnerVolumeSpecName "local-storage03-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 01 09:04:16 crc kubenswrapper[4813]: I1201 09:04:16.990625 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9402b27d-1b2c-4336-b2d7-e400772dc110-scripts" (OuterVolumeSpecName: "scripts") pod "9402b27d-1b2c-4336-b2d7-e400772dc110" (UID: "9402b27d-1b2c-4336-b2d7-e400772dc110"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:04:17 crc kubenswrapper[4813]: I1201 09:04:17.018840 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9402b27d-1b2c-4336-b2d7-e400772dc110-config-data" (OuterVolumeSpecName: "config-data") pod "9402b27d-1b2c-4336-b2d7-e400772dc110" (UID: "9402b27d-1b2c-4336-b2d7-e400772dc110"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:04:17 crc kubenswrapper[4813]: I1201 09:04:17.069085 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"9402b27d-1b2c-4336-b2d7-e400772dc110\" (UID: \"9402b27d-1b2c-4336-b2d7-e400772dc110\") " Dec 01 09:04:17 crc kubenswrapper[4813]: I1201 09:04:17.069452 4813 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/9402b27d-1b2c-4336-b2d7-e400772dc110-lib-modules\") on node \"crc\" DevicePath \"\"" Dec 01 09:04:17 crc kubenswrapper[4813]: I1201 09:04:17.069475 4813 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9402b27d-1b2c-4336-b2d7-e400772dc110-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 09:04:17 crc kubenswrapper[4813]: I1201 09:04:17.069489 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgxvz\" (UniqueName: \"kubernetes.io/projected/9402b27d-1b2c-4336-b2d7-e400772dc110-kube-api-access-zgxvz\") on node \"crc\" DevicePath \"\"" Dec 01 09:04:17 crc kubenswrapper[4813]: I1201 09:04:17.069504 4813 reconciler_common.go:293] "Volume detached for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/9402b27d-1b2c-4336-b2d7-e400772dc110-sys\") on node \"crc\" DevicePath \"\"" Dec 01 09:04:17 crc kubenswrapper[4813]: I1201 09:04:17.069517 4813 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9402b27d-1b2c-4336-b2d7-e400772dc110-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 09:04:17 crc kubenswrapper[4813]: I1201 09:04:17.069528 4813 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9402b27d-1b2c-4336-b2d7-e400772dc110-logs\") on node \"crc\" DevicePath \"\"" Dec 01 09:04:17 crc kubenswrapper[4813]: I1201 09:04:17.069553 4813 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" " Dec 01 09:04:17 crc kubenswrapper[4813]: I1201 09:04:17.069564 4813 reconciler_common.go:293] "Volume detached for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/9402b27d-1b2c-4336-b2d7-e400772dc110-etc-iscsi\") on node \"crc\" DevicePath \"\"" Dec 01 09:04:17 crc kubenswrapper[4813]: I1201 09:04:17.069575 4813 reconciler_common.go:293] "Volume detached for volume \"run\" (UniqueName: \"kubernetes.io/host-path/9402b27d-1b2c-4336-b2d7-e400772dc110-run\") on node \"crc\" DevicePath \"\"" Dec 01 09:04:17 crc kubenswrapper[4813]: I1201 09:04:17.069584 4813 reconciler_common.go:293] "Volume detached for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/9402b27d-1b2c-4336-b2d7-e400772dc110-var-locks-brick\") on node \"crc\" DevicePath \"\"" Dec 01 09:04:17 crc kubenswrapper[4813]: I1201 09:04:17.069595 4813 reconciler_common.go:293] "Volume detached for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/9402b27d-1b2c-4336-b2d7-e400772dc110-dev\") on node \"crc\" DevicePath \"\"" Dec 01 09:04:17 crc kubenswrapper[4813]: I1201 09:04:17.069602 4813 reconciler_common.go:293] "Volume detached for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/9402b27d-1b2c-4336-b2d7-e400772dc110-etc-nvme\") on node \"crc\" DevicePath \"\"" Dec 01 09:04:17 crc kubenswrapper[4813]: I1201 09:04:17.069611 4813 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/9402b27d-1b2c-4336-b2d7-e400772dc110-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 01 09:04:17 crc kubenswrapper[4813]: I1201 09:04:17.073552 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage07-crc" (OuterVolumeSpecName: "glance") pod "9402b27d-1b2c-4336-b2d7-e400772dc110" (UID: "9402b27d-1b2c-4336-b2d7-e400772dc110"). InnerVolumeSpecName "local-storage07-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 01 09:04:17 crc kubenswrapper[4813]: I1201 09:04:17.083465 4813 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage03-crc" (UniqueName: "kubernetes.io/local-volume/local-storage03-crc") on node "crc" Dec 01 09:04:17 crc kubenswrapper[4813]: I1201 09:04:17.171749 4813 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" " Dec 01 09:04:17 crc kubenswrapper[4813]: I1201 09:04:17.176095 4813 reconciler_common.go:293] "Volume detached for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" DevicePath \"\"" Dec 01 09:04:17 crc kubenswrapper[4813]: I1201 09:04:17.189270 4813 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage07-crc" (UniqueName: "kubernetes.io/local-volume/local-storage07-crc") on node "crc" Dec 01 09:04:17 crc kubenswrapper[4813]: I1201 09:04:17.222516 4813 patch_prober.go:28] interesting pod/machine-config-daemon-9wn4q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 09:04:17 crc kubenswrapper[4813]: I1201 09:04:17.222642 4813 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9wn4q" podUID="3d423402-e08f-42e9-b60c-6eb9bf2f6d48" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 09:04:17 crc kubenswrapper[4813]: I1201 09:04:17.276738 4813 reconciler_common.go:293] "Volume detached for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" DevicePath \"\"" Dec 01 09:04:17 crc kubenswrapper[4813]: I1201 09:04:17.511536 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-0" event={"ID":"d6f4323d-561b-4725-a1fe-140db5f6a33d","Type":"ContainerStarted","Data":"9ed932164815080b18b0a321c5504fcf333b242a507ed815a0fcb1e071c97b5c"} Dec 01 09:04:17 crc kubenswrapper[4813]: I1201 09:04:17.511581 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-0" event={"ID":"d6f4323d-561b-4725-a1fe-140db5f6a33d","Type":"ContainerStarted","Data":"83cba63dc3ec859be23efccb9eb8d4d3a8dbe205a4e16dbcad3a7a0ffd16a54e"} Dec 01 09:04:17 crc kubenswrapper[4813]: I1201 09:04:17.511592 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-0" event={"ID":"d6f4323d-561b-4725-a1fe-140db5f6a33d","Type":"ContainerStarted","Data":"10b31d489e5914c49ad6b4a668c2fa1494515f79057c148aeb333ecc1b27a574"} Dec 01 09:04:17 crc kubenswrapper[4813]: I1201 09:04:17.521643 4813 generic.go:334] "Generic (PLEG): container finished" podID="9402b27d-1b2c-4336-b2d7-e400772dc110" containerID="bcefcda3036d2b3710278d3a31141318f7fcff5e3c0317be8a09c5d9dab936e7" exitCode=0 Dec 01 09:04:17 crc kubenswrapper[4813]: I1201 09:04:17.521700 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"9402b27d-1b2c-4336-b2d7-e400772dc110","Type":"ContainerDied","Data":"bcefcda3036d2b3710278d3a31141318f7fcff5e3c0317be8a09c5d9dab936e7"} Dec 01 09:04:17 crc kubenswrapper[4813]: I1201 09:04:17.521732 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"9402b27d-1b2c-4336-b2d7-e400772dc110","Type":"ContainerDied","Data":"3405104a949e10ce494451973e586f506d85620fc5ad6c96af27d70760db16b7"} Dec 01 09:04:17 crc kubenswrapper[4813]: I1201 09:04:17.521759 4813 scope.go:117] "RemoveContainer" containerID="bcefcda3036d2b3710278d3a31141318f7fcff5e3c0317be8a09c5d9dab936e7" Dec 01 09:04:17 crc kubenswrapper[4813]: I1201 09:04:17.521980 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:04:17 crc kubenswrapper[4813]: E1201 09:04:17.530225 4813 kuberuntime_gc.go:389] "Failed to remove container log dead symlink" err="remove /var/log/containers/glance-default-internal-api-0_glance-kuttl-tests_glance-httpd-bcefcda3036d2b3710278d3a31141318f7fcff5e3c0317be8a09c5d9dab936e7.log: no such file or directory" path="/var/log/containers/glance-default-internal-api-0_glance-kuttl-tests_glance-httpd-bcefcda3036d2b3710278d3a31141318f7fcff5e3c0317be8a09c5d9dab936e7.log" Dec 01 09:04:17 crc kubenswrapper[4813]: I1201 09:04:17.562890 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-default-external-api-0" podStartSLOduration=2.562860577 podStartE2EDuration="2.562860577s" podCreationTimestamp="2025-12-01 09:04:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 09:04:17.545479002 +0000 UTC m=+1442.168400588" watchObservedRunningTime="2025-12-01 09:04:17.562860577 +0000 UTC m=+1442.185782163" Dec 01 09:04:17 crc kubenswrapper[4813]: I1201 09:04:17.569316 4813 scope.go:117] "RemoveContainer" containerID="565f50a4ba9ebae84a0c5064170efc08f5c3b6a0e5dfb2a87873962b319f2134" Dec 01 09:04:17 crc kubenswrapper[4813]: I1201 09:04:17.579181 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Dec 01 09:04:17 crc kubenswrapper[4813]: I1201 09:04:17.590465 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Dec 01 09:04:17 crc kubenswrapper[4813]: I1201 09:04:17.606636 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Dec 01 09:04:17 crc kubenswrapper[4813]: E1201 09:04:17.606949 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9402b27d-1b2c-4336-b2d7-e400772dc110" containerName="glance-httpd" Dec 01 09:04:17 crc kubenswrapper[4813]: I1201 09:04:17.606982 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="9402b27d-1b2c-4336-b2d7-e400772dc110" containerName="glance-httpd" Dec 01 09:04:17 crc kubenswrapper[4813]: E1201 09:04:17.607006 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9402b27d-1b2c-4336-b2d7-e400772dc110" containerName="glance-log" Dec 01 09:04:17 crc kubenswrapper[4813]: I1201 09:04:17.607013 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="9402b27d-1b2c-4336-b2d7-e400772dc110" containerName="glance-log" Dec 01 09:04:17 crc kubenswrapper[4813]: I1201 09:04:17.607170 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="9402b27d-1b2c-4336-b2d7-e400772dc110" containerName="glance-log" Dec 01 09:04:17 crc kubenswrapper[4813]: I1201 09:04:17.607194 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="9402b27d-1b2c-4336-b2d7-e400772dc110" containerName="glance-httpd" Dec 01 09:04:17 crc kubenswrapper[4813]: I1201 09:04:17.607957 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:04:17 crc kubenswrapper[4813]: I1201 09:04:17.610067 4813 scope.go:117] "RemoveContainer" containerID="bcefcda3036d2b3710278d3a31141318f7fcff5e3c0317be8a09c5d9dab936e7" Dec 01 09:04:17 crc kubenswrapper[4813]: E1201 09:04:17.611398 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bcefcda3036d2b3710278d3a31141318f7fcff5e3c0317be8a09c5d9dab936e7\": container with ID starting with bcefcda3036d2b3710278d3a31141318f7fcff5e3c0317be8a09c5d9dab936e7 not found: ID does not exist" containerID="bcefcda3036d2b3710278d3a31141318f7fcff5e3c0317be8a09c5d9dab936e7" Dec 01 09:04:17 crc kubenswrapper[4813]: I1201 09:04:17.611438 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bcefcda3036d2b3710278d3a31141318f7fcff5e3c0317be8a09c5d9dab936e7"} err="failed to get container status \"bcefcda3036d2b3710278d3a31141318f7fcff5e3c0317be8a09c5d9dab936e7\": rpc error: code = NotFound desc = could not find container \"bcefcda3036d2b3710278d3a31141318f7fcff5e3c0317be8a09c5d9dab936e7\": container with ID starting with bcefcda3036d2b3710278d3a31141318f7fcff5e3c0317be8a09c5d9dab936e7 not found: ID does not exist" Dec 01 09:04:17 crc kubenswrapper[4813]: I1201 09:04:17.611468 4813 scope.go:117] "RemoveContainer" containerID="565f50a4ba9ebae84a0c5064170efc08f5c3b6a0e5dfb2a87873962b319f2134" Dec 01 09:04:17 crc kubenswrapper[4813]: E1201 09:04:17.611884 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"565f50a4ba9ebae84a0c5064170efc08f5c3b6a0e5dfb2a87873962b319f2134\": container with ID starting with 565f50a4ba9ebae84a0c5064170efc08f5c3b6a0e5dfb2a87873962b319f2134 not found: ID does not exist" containerID="565f50a4ba9ebae84a0c5064170efc08f5c3b6a0e5dfb2a87873962b319f2134" Dec 01 09:04:17 crc kubenswrapper[4813]: I1201 09:04:17.611910 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"565f50a4ba9ebae84a0c5064170efc08f5c3b6a0e5dfb2a87873962b319f2134"} err="failed to get container status \"565f50a4ba9ebae84a0c5064170efc08f5c3b6a0e5dfb2a87873962b319f2134\": rpc error: code = NotFound desc = could not find container \"565f50a4ba9ebae84a0c5064170efc08f5c3b6a0e5dfb2a87873962b319f2134\": container with ID starting with 565f50a4ba9ebae84a0c5064170efc08f5c3b6a0e5dfb2a87873962b319f2134 not found: ID does not exist" Dec 01 09:04:17 crc kubenswrapper[4813]: I1201 09:04:17.657073 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Dec 01 09:04:17 crc kubenswrapper[4813]: I1201 09:04:17.785731 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/522f36f6-9da6-46b0-8ca6-9753311b3dfc-etc-nvme\") pod \"glance-default-internal-api-0\" (UID: \"522f36f6-9da6-46b0-8ca6-9753311b3dfc\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:04:17 crc kubenswrapper[4813]: I1201 09:04:17.785787 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/522f36f6-9da6-46b0-8ca6-9753311b3dfc-var-locks-brick\") pod \"glance-default-internal-api-0\" (UID: \"522f36f6-9da6-46b0-8ca6-9753311b3dfc\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:04:17 crc kubenswrapper[4813]: I1201 09:04:17.785838 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/522f36f6-9da6-46b0-8ca6-9753311b3dfc-scripts\") pod \"glance-default-internal-api-0\" (UID: \"522f36f6-9da6-46b0-8ca6-9753311b3dfc\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:04:17 crc kubenswrapper[4813]: I1201 09:04:17.785864 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/522f36f6-9da6-46b0-8ca6-9753311b3dfc-dev\") pod \"glance-default-internal-api-0\" (UID: \"522f36f6-9da6-46b0-8ca6-9753311b3dfc\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:04:17 crc kubenswrapper[4813]: I1201 09:04:17.785897 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/522f36f6-9da6-46b0-8ca6-9753311b3dfc-etc-iscsi\") pod \"glance-default-internal-api-0\" (UID: \"522f36f6-9da6-46b0-8ca6-9753311b3dfc\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:04:17 crc kubenswrapper[4813]: I1201 09:04:17.785921 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/522f36f6-9da6-46b0-8ca6-9753311b3dfc-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"522f36f6-9da6-46b0-8ca6-9753311b3dfc\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:04:17 crc kubenswrapper[4813]: I1201 09:04:17.785939 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/522f36f6-9da6-46b0-8ca6-9753311b3dfc-logs\") pod \"glance-default-internal-api-0\" (UID: \"522f36f6-9da6-46b0-8ca6-9753311b3dfc\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:04:17 crc kubenswrapper[4813]: I1201 09:04:17.785959 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/522f36f6-9da6-46b0-8ca6-9753311b3dfc-lib-modules\") pod \"glance-default-internal-api-0\" (UID: \"522f36f6-9da6-46b0-8ca6-9753311b3dfc\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:04:17 crc kubenswrapper[4813]: I1201 09:04:17.786008 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/522f36f6-9da6-46b0-8ca6-9753311b3dfc-run\") pod \"glance-default-internal-api-0\" (UID: \"522f36f6-9da6-46b0-8ca6-9753311b3dfc\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:04:17 crc kubenswrapper[4813]: I1201 09:04:17.786025 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v7h7b\" (UniqueName: \"kubernetes.io/projected/522f36f6-9da6-46b0-8ca6-9753311b3dfc-kube-api-access-v7h7b\") pod \"glance-default-internal-api-0\" (UID: \"522f36f6-9da6-46b0-8ca6-9753311b3dfc\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:04:17 crc kubenswrapper[4813]: I1201 09:04:17.786041 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/522f36f6-9da6-46b0-8ca6-9753311b3dfc-config-data\") pod \"glance-default-internal-api-0\" (UID: \"522f36f6-9da6-46b0-8ca6-9753311b3dfc\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:04:17 crc kubenswrapper[4813]: I1201 09:04:17.786061 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/522f36f6-9da6-46b0-8ca6-9753311b3dfc-sys\") pod \"glance-default-internal-api-0\" (UID: \"522f36f6-9da6-46b0-8ca6-9753311b3dfc\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:04:17 crc kubenswrapper[4813]: I1201 09:04:17.786082 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-internal-api-0\" (UID: \"522f36f6-9da6-46b0-8ca6-9753311b3dfc\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:04:17 crc kubenswrapper[4813]: I1201 09:04:17.786098 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-internal-api-0\" (UID: \"522f36f6-9da6-46b0-8ca6-9753311b3dfc\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:04:17 crc kubenswrapper[4813]: I1201 09:04:17.888035 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/522f36f6-9da6-46b0-8ca6-9753311b3dfc-scripts\") pod \"glance-default-internal-api-0\" (UID: \"522f36f6-9da6-46b0-8ca6-9753311b3dfc\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:04:17 crc kubenswrapper[4813]: I1201 09:04:17.888109 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/522f36f6-9da6-46b0-8ca6-9753311b3dfc-dev\") pod \"glance-default-internal-api-0\" (UID: \"522f36f6-9da6-46b0-8ca6-9753311b3dfc\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:04:17 crc kubenswrapper[4813]: I1201 09:04:17.888162 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/522f36f6-9da6-46b0-8ca6-9753311b3dfc-etc-iscsi\") pod \"glance-default-internal-api-0\" (UID: \"522f36f6-9da6-46b0-8ca6-9753311b3dfc\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:04:17 crc kubenswrapper[4813]: I1201 09:04:17.888197 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/522f36f6-9da6-46b0-8ca6-9753311b3dfc-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"522f36f6-9da6-46b0-8ca6-9753311b3dfc\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:04:17 crc kubenswrapper[4813]: I1201 09:04:17.888225 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/522f36f6-9da6-46b0-8ca6-9753311b3dfc-logs\") pod \"glance-default-internal-api-0\" (UID: \"522f36f6-9da6-46b0-8ca6-9753311b3dfc\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:04:17 crc kubenswrapper[4813]: I1201 09:04:17.888251 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/522f36f6-9da6-46b0-8ca6-9753311b3dfc-lib-modules\") pod \"glance-default-internal-api-0\" (UID: \"522f36f6-9da6-46b0-8ca6-9753311b3dfc\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:04:17 crc kubenswrapper[4813]: I1201 09:04:17.888295 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/522f36f6-9da6-46b0-8ca6-9753311b3dfc-run\") pod \"glance-default-internal-api-0\" (UID: \"522f36f6-9da6-46b0-8ca6-9753311b3dfc\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:04:17 crc kubenswrapper[4813]: I1201 09:04:17.888319 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v7h7b\" (UniqueName: \"kubernetes.io/projected/522f36f6-9da6-46b0-8ca6-9753311b3dfc-kube-api-access-v7h7b\") pod \"glance-default-internal-api-0\" (UID: \"522f36f6-9da6-46b0-8ca6-9753311b3dfc\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:04:17 crc kubenswrapper[4813]: I1201 09:04:17.888341 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/522f36f6-9da6-46b0-8ca6-9753311b3dfc-config-data\") pod \"glance-default-internal-api-0\" (UID: \"522f36f6-9da6-46b0-8ca6-9753311b3dfc\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:04:17 crc kubenswrapper[4813]: I1201 09:04:17.888371 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/522f36f6-9da6-46b0-8ca6-9753311b3dfc-sys\") pod \"glance-default-internal-api-0\" (UID: \"522f36f6-9da6-46b0-8ca6-9753311b3dfc\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:04:17 crc kubenswrapper[4813]: I1201 09:04:17.888395 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-internal-api-0\" (UID: \"522f36f6-9da6-46b0-8ca6-9753311b3dfc\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:04:17 crc kubenswrapper[4813]: I1201 09:04:17.888416 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-internal-api-0\" (UID: \"522f36f6-9da6-46b0-8ca6-9753311b3dfc\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:04:17 crc kubenswrapper[4813]: I1201 09:04:17.888451 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/522f36f6-9da6-46b0-8ca6-9753311b3dfc-etc-nvme\") pod \"glance-default-internal-api-0\" (UID: \"522f36f6-9da6-46b0-8ca6-9753311b3dfc\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:04:17 crc kubenswrapper[4813]: I1201 09:04:17.888479 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/522f36f6-9da6-46b0-8ca6-9753311b3dfc-var-locks-brick\") pod \"glance-default-internal-api-0\" (UID: \"522f36f6-9da6-46b0-8ca6-9753311b3dfc\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:04:17 crc kubenswrapper[4813]: I1201 09:04:17.888604 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/522f36f6-9da6-46b0-8ca6-9753311b3dfc-var-locks-brick\") pod \"glance-default-internal-api-0\" (UID: \"522f36f6-9da6-46b0-8ca6-9753311b3dfc\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:04:17 crc kubenswrapper[4813]: I1201 09:04:17.888650 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/522f36f6-9da6-46b0-8ca6-9753311b3dfc-dev\") pod \"glance-default-internal-api-0\" (UID: \"522f36f6-9da6-46b0-8ca6-9753311b3dfc\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:04:17 crc kubenswrapper[4813]: I1201 09:04:17.888678 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/522f36f6-9da6-46b0-8ca6-9753311b3dfc-etc-iscsi\") pod \"glance-default-internal-api-0\" (UID: \"522f36f6-9da6-46b0-8ca6-9753311b3dfc\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:04:17 crc kubenswrapper[4813]: I1201 09:04:17.889074 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/522f36f6-9da6-46b0-8ca6-9753311b3dfc-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"522f36f6-9da6-46b0-8ca6-9753311b3dfc\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:04:17 crc kubenswrapper[4813]: I1201 09:04:17.889255 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/522f36f6-9da6-46b0-8ca6-9753311b3dfc-run\") pod \"glance-default-internal-api-0\" (UID: \"522f36f6-9da6-46b0-8ca6-9753311b3dfc\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:04:17 crc kubenswrapper[4813]: I1201 09:04:17.889300 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/522f36f6-9da6-46b0-8ca6-9753311b3dfc-logs\") pod \"glance-default-internal-api-0\" (UID: \"522f36f6-9da6-46b0-8ca6-9753311b3dfc\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:04:17 crc kubenswrapper[4813]: I1201 09:04:17.889349 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/522f36f6-9da6-46b0-8ca6-9753311b3dfc-lib-modules\") pod \"glance-default-internal-api-0\" (UID: \"522f36f6-9da6-46b0-8ca6-9753311b3dfc\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:04:17 crc kubenswrapper[4813]: I1201 09:04:17.889368 4813 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-internal-api-0\" (UID: \"522f36f6-9da6-46b0-8ca6-9753311b3dfc\") device mount path \"/mnt/openstack/pv07\"" pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:04:17 crc kubenswrapper[4813]: I1201 09:04:17.889409 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/522f36f6-9da6-46b0-8ca6-9753311b3dfc-etc-nvme\") pod \"glance-default-internal-api-0\" (UID: \"522f36f6-9da6-46b0-8ca6-9753311b3dfc\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:04:17 crc kubenswrapper[4813]: I1201 09:04:17.889467 4813 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-internal-api-0\" (UID: \"522f36f6-9da6-46b0-8ca6-9753311b3dfc\") device mount path \"/mnt/openstack/pv03\"" pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:04:17 crc kubenswrapper[4813]: I1201 09:04:17.889463 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/522f36f6-9da6-46b0-8ca6-9753311b3dfc-sys\") pod \"glance-default-internal-api-0\" (UID: \"522f36f6-9da6-46b0-8ca6-9753311b3dfc\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:04:17 crc kubenswrapper[4813]: I1201 09:04:17.895824 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/522f36f6-9da6-46b0-8ca6-9753311b3dfc-scripts\") pod \"glance-default-internal-api-0\" (UID: \"522f36f6-9da6-46b0-8ca6-9753311b3dfc\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:04:17 crc kubenswrapper[4813]: I1201 09:04:17.905751 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/522f36f6-9da6-46b0-8ca6-9753311b3dfc-config-data\") pod \"glance-default-internal-api-0\" (UID: \"522f36f6-9da6-46b0-8ca6-9753311b3dfc\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:04:17 crc kubenswrapper[4813]: I1201 09:04:17.908606 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v7h7b\" (UniqueName: \"kubernetes.io/projected/522f36f6-9da6-46b0-8ca6-9753311b3dfc-kube-api-access-v7h7b\") pod \"glance-default-internal-api-0\" (UID: \"522f36f6-9da6-46b0-8ca6-9753311b3dfc\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:04:17 crc kubenswrapper[4813]: I1201 09:04:17.911863 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-internal-api-0\" (UID: \"522f36f6-9da6-46b0-8ca6-9753311b3dfc\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:04:17 crc kubenswrapper[4813]: I1201 09:04:17.916554 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-internal-api-0\" (UID: \"522f36f6-9da6-46b0-8ca6-9753311b3dfc\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:04:17 crc kubenswrapper[4813]: I1201 09:04:17.944795 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:04:18 crc kubenswrapper[4813]: I1201 09:04:18.267307 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Dec 01 09:04:18 crc kubenswrapper[4813]: I1201 09:04:18.402555 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9402b27d-1b2c-4336-b2d7-e400772dc110" path="/var/lib/kubelet/pods/9402b27d-1b2c-4336-b2d7-e400772dc110/volumes" Dec 01 09:04:18 crc kubenswrapper[4813]: I1201 09:04:18.548501 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"522f36f6-9da6-46b0-8ca6-9753311b3dfc","Type":"ContainerStarted","Data":"954887d4b9b9439ab9fdfc4078054b7d9d86884474571c85ee8708f64ca6198a"} Dec 01 09:04:18 crc kubenswrapper[4813]: I1201 09:04:18.548821 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"522f36f6-9da6-46b0-8ca6-9753311b3dfc","Type":"ContainerStarted","Data":"011eaf2acf8bb7bc009d418aad121279c5ff5d494e6e7b6e142025263a66edeb"} Dec 01 09:04:19 crc kubenswrapper[4813]: I1201 09:04:19.564819 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"522f36f6-9da6-46b0-8ca6-9753311b3dfc","Type":"ContainerStarted","Data":"269455c89215c29899616183ae9d07b026d8fc34efad53048b988fc15a0f5889"} Dec 01 09:04:19 crc kubenswrapper[4813]: I1201 09:04:19.593826 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-default-internal-api-0" podStartSLOduration=2.5937943150000002 podStartE2EDuration="2.593794315s" podCreationTimestamp="2025-12-01 09:04:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 09:04:19.589142322 +0000 UTC m=+1444.212063908" watchObservedRunningTime="2025-12-01 09:04:19.593794315 +0000 UTC m=+1444.216715901" Dec 01 09:04:26 crc kubenswrapper[4813]: I1201 09:04:26.219028 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-external-api-0" Dec 01 09:04:26 crc kubenswrapper[4813]: I1201 09:04:26.219729 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-external-api-0" Dec 01 09:04:26 crc kubenswrapper[4813]: I1201 09:04:26.254059 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-external-api-0" Dec 01 09:04:26 crc kubenswrapper[4813]: I1201 09:04:26.271585 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-external-api-0" Dec 01 09:04:26 crc kubenswrapper[4813]: I1201 09:04:26.635903 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-external-api-0" Dec 01 09:04:26 crc kubenswrapper[4813]: I1201 09:04:26.635951 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-external-api-0" Dec 01 09:04:27 crc kubenswrapper[4813]: I1201 09:04:27.945514 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:04:27 crc kubenswrapper[4813]: I1201 09:04:27.946913 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:04:27 crc kubenswrapper[4813]: I1201 09:04:27.972487 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:04:27 crc kubenswrapper[4813]: I1201 09:04:27.994286 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:04:28 crc kubenswrapper[4813]: I1201 09:04:28.669316 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:04:28 crc kubenswrapper[4813]: I1201 09:04:28.669677 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:04:28 crc kubenswrapper[4813]: I1201 09:04:28.764763 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-external-api-0" Dec 01 09:04:28 crc kubenswrapper[4813]: I1201 09:04:28.765034 4813 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 01 09:04:28 crc kubenswrapper[4813]: I1201 09:04:28.767696 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-external-api-0" Dec 01 09:04:30 crc kubenswrapper[4813]: I1201 09:04:30.574635 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:04:30 crc kubenswrapper[4813]: I1201 09:04:30.575599 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:04:47 crc kubenswrapper[4813]: I1201 09:04:47.222398 4813 patch_prober.go:28] interesting pod/machine-config-daemon-9wn4q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 09:04:47 crc kubenswrapper[4813]: I1201 09:04:47.224290 4813 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9wn4q" podUID="3d423402-e08f-42e9-b60c-6eb9bf2f6d48" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 09:05:09 crc kubenswrapper[4813]: I1201 09:05:09.568864 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-1"] Dec 01 09:05:09 crc kubenswrapper[4813]: I1201 09:05:09.569759 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-external-api-1" podUID="b4c39a03-2938-4b10-b430-369066c79ce4" containerName="glance-log" containerID="cri-o://86514afa63c5089fdc14f9b082809d402f7390c825756c51d1a5199647657b3d" gracePeriod=30 Dec 01 09:05:09 crc kubenswrapper[4813]: I1201 09:05:09.569839 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-external-api-1" podUID="b4c39a03-2938-4b10-b430-369066c79ce4" containerName="glance-httpd" containerID="cri-o://bdc319325263c4f53d0f49b476c0d63f087082dd8eda5e1e4f5f291bdaaffad5" gracePeriod=30 Dec 01 09:05:09 crc kubenswrapper[4813]: I1201 09:05:09.729465 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-1"] Dec 01 09:05:09 crc kubenswrapper[4813]: I1201 09:05:09.730266 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-internal-api-1" podUID="c079746c-0f0e-403a-bb6e-b1866a063050" containerName="glance-httpd" containerID="cri-o://d191ad43cbe03d01275077b6d3666ca27b942e2e476948af348780daa2824a34" gracePeriod=30 Dec 01 09:05:09 crc kubenswrapper[4813]: I1201 09:05:09.730522 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-internal-api-1" podUID="c079746c-0f0e-403a-bb6e-b1866a063050" containerName="glance-log" containerID="cri-o://d0e434dc9af434eac9e19163ad506f93f21a1b45f5335294ea9b9a8d72872b0f" gracePeriod=30 Dec 01 09:05:10 crc kubenswrapper[4813]: I1201 09:05:10.080790 4813 generic.go:334] "Generic (PLEG): container finished" podID="c079746c-0f0e-403a-bb6e-b1866a063050" containerID="d0e434dc9af434eac9e19163ad506f93f21a1b45f5335294ea9b9a8d72872b0f" exitCode=143 Dec 01 09:05:10 crc kubenswrapper[4813]: I1201 09:05:10.080860 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-1" event={"ID":"c079746c-0f0e-403a-bb6e-b1866a063050","Type":"ContainerDied","Data":"d0e434dc9af434eac9e19163ad506f93f21a1b45f5335294ea9b9a8d72872b0f"} Dec 01 09:05:10 crc kubenswrapper[4813]: I1201 09:05:10.084580 4813 generic.go:334] "Generic (PLEG): container finished" podID="b4c39a03-2938-4b10-b430-369066c79ce4" containerID="86514afa63c5089fdc14f9b082809d402f7390c825756c51d1a5199647657b3d" exitCode=143 Dec 01 09:05:10 crc kubenswrapper[4813]: I1201 09:05:10.084621 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-1" event={"ID":"b4c39a03-2938-4b10-b430-369066c79ce4","Type":"ContainerDied","Data":"86514afa63c5089fdc14f9b082809d402f7390c825756c51d1a5199647657b3d"} Dec 01 09:05:11 crc kubenswrapper[4813]: I1201 09:05:11.017262 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-db-sync-78knl"] Dec 01 09:05:11 crc kubenswrapper[4813]: I1201 09:05:11.026476 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-db-sync-78knl"] Dec 01 09:05:11 crc kubenswrapper[4813]: I1201 09:05:11.034686 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glancea9c4-account-delete-wl47f"] Dec 01 09:05:11 crc kubenswrapper[4813]: I1201 09:05:11.037356 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glancea9c4-account-delete-wl47f" Dec 01 09:05:11 crc kubenswrapper[4813]: I1201 09:05:11.050687 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glancea9c4-account-delete-wl47f"] Dec 01 09:05:11 crc kubenswrapper[4813]: I1201 09:05:11.090587 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-0"] Dec 01 09:05:11 crc kubenswrapper[4813]: I1201 09:05:11.090854 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-external-api-0" podUID="d6f4323d-561b-4725-a1fe-140db5f6a33d" containerName="glance-log" containerID="cri-o://83cba63dc3ec859be23efccb9eb8d4d3a8dbe205a4e16dbcad3a7a0ffd16a54e" gracePeriod=30 Dec 01 09:05:11 crc kubenswrapper[4813]: I1201 09:05:11.091399 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-external-api-0" podUID="d6f4323d-561b-4725-a1fe-140db5f6a33d" containerName="glance-httpd" containerID="cri-o://9ed932164815080b18b0a321c5504fcf333b242a507ed815a0fcb1e071c97b5c" gracePeriod=30 Dec 01 09:05:11 crc kubenswrapper[4813]: I1201 09:05:11.198284 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/54c06e84-7478-4fbc-9e2c-bdc6e1be658b-operator-scripts\") pod \"glancea9c4-account-delete-wl47f\" (UID: \"54c06e84-7478-4fbc-9e2c-bdc6e1be658b\") " pod="glance-kuttl-tests/glancea9c4-account-delete-wl47f" Dec 01 09:05:11 crc kubenswrapper[4813]: I1201 09:05:11.198416 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4r5z5\" (UniqueName: \"kubernetes.io/projected/54c06e84-7478-4fbc-9e2c-bdc6e1be658b-kube-api-access-4r5z5\") pod \"glancea9c4-account-delete-wl47f\" (UID: \"54c06e84-7478-4fbc-9e2c-bdc6e1be658b\") " pod="glance-kuttl-tests/glancea9c4-account-delete-wl47f" Dec 01 09:05:11 crc kubenswrapper[4813]: I1201 09:05:11.223016 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Dec 01 09:05:11 crc kubenswrapper[4813]: I1201 09:05:11.231167 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-internal-api-0" podUID="522f36f6-9da6-46b0-8ca6-9753311b3dfc" containerName="glance-log" containerID="cri-o://954887d4b9b9439ab9fdfc4078054b7d9d86884474571c85ee8708f64ca6198a" gracePeriod=30 Dec 01 09:05:11 crc kubenswrapper[4813]: I1201 09:05:11.231779 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-internal-api-0" podUID="522f36f6-9da6-46b0-8ca6-9753311b3dfc" containerName="glance-httpd" containerID="cri-o://269455c89215c29899616183ae9d07b026d8fc34efad53048b988fc15a0f5889" gracePeriod=30 Dec 01 09:05:11 crc kubenswrapper[4813]: I1201 09:05:11.299708 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4r5z5\" (UniqueName: \"kubernetes.io/projected/54c06e84-7478-4fbc-9e2c-bdc6e1be658b-kube-api-access-4r5z5\") pod \"glancea9c4-account-delete-wl47f\" (UID: \"54c06e84-7478-4fbc-9e2c-bdc6e1be658b\") " pod="glance-kuttl-tests/glancea9c4-account-delete-wl47f" Dec 01 09:05:11 crc kubenswrapper[4813]: I1201 09:05:11.299829 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/54c06e84-7478-4fbc-9e2c-bdc6e1be658b-operator-scripts\") pod \"glancea9c4-account-delete-wl47f\" (UID: \"54c06e84-7478-4fbc-9e2c-bdc6e1be658b\") " pod="glance-kuttl-tests/glancea9c4-account-delete-wl47f" Dec 01 09:05:11 crc kubenswrapper[4813]: I1201 09:05:11.300691 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/54c06e84-7478-4fbc-9e2c-bdc6e1be658b-operator-scripts\") pod \"glancea9c4-account-delete-wl47f\" (UID: \"54c06e84-7478-4fbc-9e2c-bdc6e1be658b\") " pod="glance-kuttl-tests/glancea9c4-account-delete-wl47f" Dec 01 09:05:11 crc kubenswrapper[4813]: I1201 09:05:11.316645 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4r5z5\" (UniqueName: \"kubernetes.io/projected/54c06e84-7478-4fbc-9e2c-bdc6e1be658b-kube-api-access-4r5z5\") pod \"glancea9c4-account-delete-wl47f\" (UID: \"54c06e84-7478-4fbc-9e2c-bdc6e1be658b\") " pod="glance-kuttl-tests/glancea9c4-account-delete-wl47f" Dec 01 09:05:11 crc kubenswrapper[4813]: I1201 09:05:11.361080 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glancea9c4-account-delete-wl47f" Dec 01 09:05:11 crc kubenswrapper[4813]: I1201 09:05:11.803002 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glancea9c4-account-delete-wl47f"] Dec 01 09:05:12 crc kubenswrapper[4813]: I1201 09:05:12.104186 4813 generic.go:334] "Generic (PLEG): container finished" podID="54c06e84-7478-4fbc-9e2c-bdc6e1be658b" containerID="1e92e4c9dd814667919f1d0dd4ef78e7e090b2558af069309ed4e0894bcc82e4" exitCode=0 Dec 01 09:05:12 crc kubenswrapper[4813]: I1201 09:05:12.104298 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glancea9c4-account-delete-wl47f" event={"ID":"54c06e84-7478-4fbc-9e2c-bdc6e1be658b","Type":"ContainerDied","Data":"1e92e4c9dd814667919f1d0dd4ef78e7e090b2558af069309ed4e0894bcc82e4"} Dec 01 09:05:12 crc kubenswrapper[4813]: I1201 09:05:12.104615 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glancea9c4-account-delete-wl47f" event={"ID":"54c06e84-7478-4fbc-9e2c-bdc6e1be658b","Type":"ContainerStarted","Data":"17d938babbce43861ca8450a7ec26271d9ef23e64fde291843ab3616d7793eac"} Dec 01 09:05:12 crc kubenswrapper[4813]: I1201 09:05:12.108834 4813 generic.go:334] "Generic (PLEG): container finished" podID="d6f4323d-561b-4725-a1fe-140db5f6a33d" containerID="83cba63dc3ec859be23efccb9eb8d4d3a8dbe205a4e16dbcad3a7a0ffd16a54e" exitCode=143 Dec 01 09:05:12 crc kubenswrapper[4813]: I1201 09:05:12.108938 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-0" event={"ID":"d6f4323d-561b-4725-a1fe-140db5f6a33d","Type":"ContainerDied","Data":"83cba63dc3ec859be23efccb9eb8d4d3a8dbe205a4e16dbcad3a7a0ffd16a54e"} Dec 01 09:05:12 crc kubenswrapper[4813]: I1201 09:05:12.111921 4813 generic.go:334] "Generic (PLEG): container finished" podID="522f36f6-9da6-46b0-8ca6-9753311b3dfc" containerID="954887d4b9b9439ab9fdfc4078054b7d9d86884474571c85ee8708f64ca6198a" exitCode=143 Dec 01 09:05:12 crc kubenswrapper[4813]: I1201 09:05:12.112032 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"522f36f6-9da6-46b0-8ca6-9753311b3dfc","Type":"ContainerDied","Data":"954887d4b9b9439ab9fdfc4078054b7d9d86884474571c85ee8708f64ca6198a"} Dec 01 09:05:12 crc kubenswrapper[4813]: I1201 09:05:12.403089 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="40ca776e-7b84-4e84-b976-521ae258cf0e" path="/var/lib/kubelet/pods/40ca776e-7b84-4e84-b976-521ae258cf0e/volumes" Dec 01 09:05:12 crc kubenswrapper[4813]: E1201 09:05:12.845206 4813 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb4c39a03_2938_4b10_b430_369066c79ce4.slice/crio-bdc319325263c4f53d0f49b476c0d63f087082dd8eda5e1e4f5f291bdaaffad5.scope\": RecentStats: unable to find data in memory cache]" Dec 01 09:05:13 crc kubenswrapper[4813]: I1201 09:05:13.121922 4813 generic.go:334] "Generic (PLEG): container finished" podID="b4c39a03-2938-4b10-b430-369066c79ce4" containerID="bdc319325263c4f53d0f49b476c0d63f087082dd8eda5e1e4f5f291bdaaffad5" exitCode=0 Dec 01 09:05:13 crc kubenswrapper[4813]: I1201 09:05:13.122106 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-1" event={"ID":"b4c39a03-2938-4b10-b430-369066c79ce4","Type":"ContainerDied","Data":"bdc319325263c4f53d0f49b476c0d63f087082dd8eda5e1e4f5f291bdaaffad5"} Dec 01 09:05:13 crc kubenswrapper[4813]: I1201 09:05:13.122324 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-1" event={"ID":"b4c39a03-2938-4b10-b430-369066c79ce4","Type":"ContainerDied","Data":"f5144a488ccb46ecee1341b65655217027f5b5efc12f927118b96e06368e6f8e"} Dec 01 09:05:13 crc kubenswrapper[4813]: I1201 09:05:13.122346 4813 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f5144a488ccb46ecee1341b65655217027f5b5efc12f927118b96e06368e6f8e" Dec 01 09:05:13 crc kubenswrapper[4813]: I1201 09:05:13.124040 4813 generic.go:334] "Generic (PLEG): container finished" podID="c079746c-0f0e-403a-bb6e-b1866a063050" containerID="d191ad43cbe03d01275077b6d3666ca27b942e2e476948af348780daa2824a34" exitCode=0 Dec 01 09:05:13 crc kubenswrapper[4813]: I1201 09:05:13.124201 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-1" event={"ID":"c079746c-0f0e-403a-bb6e-b1866a063050","Type":"ContainerDied","Data":"d191ad43cbe03d01275077b6d3666ca27b942e2e476948af348780daa2824a34"} Dec 01 09:05:13 crc kubenswrapper[4813]: I1201 09:05:13.132914 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-1" Dec 01 09:05:13 crc kubenswrapper[4813]: I1201 09:05:13.257649 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/b4c39a03-2938-4b10-b430-369066c79ce4-sys\") pod \"b4c39a03-2938-4b10-b430-369066c79ce4\" (UID: \"b4c39a03-2938-4b10-b430-369066c79ce4\") " Dec 01 09:05:13 crc kubenswrapper[4813]: I1201 09:05:13.257699 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/b4c39a03-2938-4b10-b430-369066c79ce4-lib-modules\") pod \"b4c39a03-2938-4b10-b430-369066c79ce4\" (UID: \"b4c39a03-2938-4b10-b430-369066c79ce4\") " Dec 01 09:05:13 crc kubenswrapper[4813]: I1201 09:05:13.257784 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sgn7t\" (UniqueName: \"kubernetes.io/projected/b4c39a03-2938-4b10-b430-369066c79ce4-kube-api-access-sgn7t\") pod \"b4c39a03-2938-4b10-b430-369066c79ce4\" (UID: \"b4c39a03-2938-4b10-b430-369066c79ce4\") " Dec 01 09:05:13 crc kubenswrapper[4813]: I1201 09:05:13.257811 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/b4c39a03-2938-4b10-b430-369066c79ce4-etc-nvme\") pod \"b4c39a03-2938-4b10-b430-369066c79ce4\" (UID: \"b4c39a03-2938-4b10-b430-369066c79ce4\") " Dec 01 09:05:13 crc kubenswrapper[4813]: I1201 09:05:13.257980 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b4c39a03-2938-4b10-b430-369066c79ce4-sys" (OuterVolumeSpecName: "sys") pod "b4c39a03-2938-4b10-b430-369066c79ce4" (UID: "b4c39a03-2938-4b10-b430-369066c79ce4"). InnerVolumeSpecName "sys". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 09:05:13 crc kubenswrapper[4813]: I1201 09:05:13.258030 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b4c39a03-2938-4b10-b430-369066c79ce4-etc-nvme" (OuterVolumeSpecName: "etc-nvme") pod "b4c39a03-2938-4b10-b430-369066c79ce4" (UID: "b4c39a03-2938-4b10-b430-369066c79ce4"). InnerVolumeSpecName "etc-nvme". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 09:05:13 crc kubenswrapper[4813]: I1201 09:05:13.257978 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b4c39a03-2938-4b10-b430-369066c79ce4-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "b4c39a03-2938-4b10-b430-369066c79ce4" (UID: "b4c39a03-2938-4b10-b430-369066c79ce4"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 09:05:13 crc kubenswrapper[4813]: I1201 09:05:13.258838 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/b4c39a03-2938-4b10-b430-369066c79ce4-run\") pod \"b4c39a03-2938-4b10-b430-369066c79ce4\" (UID: \"b4c39a03-2938-4b10-b430-369066c79ce4\") " Dec 01 09:05:13 crc kubenswrapper[4813]: I1201 09:05:13.258895 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/b4c39a03-2938-4b10-b430-369066c79ce4-var-locks-brick\") pod \"b4c39a03-2938-4b10-b430-369066c79ce4\" (UID: \"b4c39a03-2938-4b10-b430-369066c79ce4\") " Dec 01 09:05:13 crc kubenswrapper[4813]: I1201 09:05:13.258946 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance-cache\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"b4c39a03-2938-4b10-b430-369066c79ce4\" (UID: \"b4c39a03-2938-4b10-b430-369066c79ce4\") " Dec 01 09:05:13 crc kubenswrapper[4813]: I1201 09:05:13.258982 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/b4c39a03-2938-4b10-b430-369066c79ce4-etc-iscsi\") pod \"b4c39a03-2938-4b10-b430-369066c79ce4\" (UID: \"b4c39a03-2938-4b10-b430-369066c79ce4\") " Dec 01 09:05:13 crc kubenswrapper[4813]: I1201 09:05:13.259018 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b4c39a03-2938-4b10-b430-369066c79ce4-config-data\") pod \"b4c39a03-2938-4b10-b430-369066c79ce4\" (UID: \"b4c39a03-2938-4b10-b430-369066c79ce4\") " Dec 01 09:05:13 crc kubenswrapper[4813]: I1201 09:05:13.259045 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/b4c39a03-2938-4b10-b430-369066c79ce4-dev\") pod \"b4c39a03-2938-4b10-b430-369066c79ce4\" (UID: \"b4c39a03-2938-4b10-b430-369066c79ce4\") " Dec 01 09:05:13 crc kubenswrapper[4813]: I1201 09:05:13.259076 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage14-crc\") pod \"b4c39a03-2938-4b10-b430-369066c79ce4\" (UID: \"b4c39a03-2938-4b10-b430-369066c79ce4\") " Dec 01 09:05:13 crc kubenswrapper[4813]: I1201 09:05:13.259100 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b4c39a03-2938-4b10-b430-369066c79ce4-logs\") pod \"b4c39a03-2938-4b10-b430-369066c79ce4\" (UID: \"b4c39a03-2938-4b10-b430-369066c79ce4\") " Dec 01 09:05:13 crc kubenswrapper[4813]: I1201 09:05:13.259129 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b4c39a03-2938-4b10-b430-369066c79ce4-scripts\") pod \"b4c39a03-2938-4b10-b430-369066c79ce4\" (UID: \"b4c39a03-2938-4b10-b430-369066c79ce4\") " Dec 01 09:05:13 crc kubenswrapper[4813]: I1201 09:05:13.259172 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b4c39a03-2938-4b10-b430-369066c79ce4-httpd-run\") pod \"b4c39a03-2938-4b10-b430-369066c79ce4\" (UID: \"b4c39a03-2938-4b10-b430-369066c79ce4\") " Dec 01 09:05:13 crc kubenswrapper[4813]: I1201 09:05:13.259545 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b4c39a03-2938-4b10-b430-369066c79ce4-var-locks-brick" (OuterVolumeSpecName: "var-locks-brick") pod "b4c39a03-2938-4b10-b430-369066c79ce4" (UID: "b4c39a03-2938-4b10-b430-369066c79ce4"). InnerVolumeSpecName "var-locks-brick". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 09:05:13 crc kubenswrapper[4813]: I1201 09:05:13.259603 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b4c39a03-2938-4b10-b430-369066c79ce4-run" (OuterVolumeSpecName: "run") pod "b4c39a03-2938-4b10-b430-369066c79ce4" (UID: "b4c39a03-2938-4b10-b430-369066c79ce4"). InnerVolumeSpecName "run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 09:05:13 crc kubenswrapper[4813]: I1201 09:05:13.259620 4813 reconciler_common.go:293] "Volume detached for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/b4c39a03-2938-4b10-b430-369066c79ce4-sys\") on node \"crc\" DevicePath \"\"" Dec 01 09:05:13 crc kubenswrapper[4813]: I1201 09:05:13.259636 4813 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/b4c39a03-2938-4b10-b430-369066c79ce4-lib-modules\") on node \"crc\" DevicePath \"\"" Dec 01 09:05:13 crc kubenswrapper[4813]: I1201 09:05:13.259646 4813 reconciler_common.go:293] "Volume detached for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/b4c39a03-2938-4b10-b430-369066c79ce4-etc-nvme\") on node \"crc\" DevicePath \"\"" Dec 01 09:05:13 crc kubenswrapper[4813]: I1201 09:05:13.259941 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b4c39a03-2938-4b10-b430-369066c79ce4-etc-iscsi" (OuterVolumeSpecName: "etc-iscsi") pod "b4c39a03-2938-4b10-b430-369066c79ce4" (UID: "b4c39a03-2938-4b10-b430-369066c79ce4"). InnerVolumeSpecName "etc-iscsi". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 09:05:13 crc kubenswrapper[4813]: I1201 09:05:13.260002 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b4c39a03-2938-4b10-b430-369066c79ce4-dev" (OuterVolumeSpecName: "dev") pod "b4c39a03-2938-4b10-b430-369066c79ce4" (UID: "b4c39a03-2938-4b10-b430-369066c79ce4"). InnerVolumeSpecName "dev". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 09:05:13 crc kubenswrapper[4813]: I1201 09:05:13.260197 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b4c39a03-2938-4b10-b430-369066c79ce4-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "b4c39a03-2938-4b10-b430-369066c79ce4" (UID: "b4c39a03-2938-4b10-b430-369066c79ce4"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 09:05:13 crc kubenswrapper[4813]: I1201 09:05:13.260357 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b4c39a03-2938-4b10-b430-369066c79ce4-logs" (OuterVolumeSpecName: "logs") pod "b4c39a03-2938-4b10-b430-369066c79ce4" (UID: "b4c39a03-2938-4b10-b430-369066c79ce4"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 09:05:13 crc kubenswrapper[4813]: I1201 09:05:13.264287 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage04-crc" (OuterVolumeSpecName: "glance-cache") pod "b4c39a03-2938-4b10-b430-369066c79ce4" (UID: "b4c39a03-2938-4b10-b430-369066c79ce4"). InnerVolumeSpecName "local-storage04-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 01 09:05:13 crc kubenswrapper[4813]: I1201 09:05:13.264671 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b4c39a03-2938-4b10-b430-369066c79ce4-scripts" (OuterVolumeSpecName: "scripts") pod "b4c39a03-2938-4b10-b430-369066c79ce4" (UID: "b4c39a03-2938-4b10-b430-369066c79ce4"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:05:13 crc kubenswrapper[4813]: I1201 09:05:13.268447 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b4c39a03-2938-4b10-b430-369066c79ce4-kube-api-access-sgn7t" (OuterVolumeSpecName: "kube-api-access-sgn7t") pod "b4c39a03-2938-4b10-b430-369066c79ce4" (UID: "b4c39a03-2938-4b10-b430-369066c79ce4"). InnerVolumeSpecName "kube-api-access-sgn7t". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:05:13 crc kubenswrapper[4813]: I1201 09:05:13.295229 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage14-crc" (OuterVolumeSpecName: "glance") pod "b4c39a03-2938-4b10-b430-369066c79ce4" (UID: "b4c39a03-2938-4b10-b430-369066c79ce4"). InnerVolumeSpecName "local-storage14-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 01 09:05:13 crc kubenswrapper[4813]: I1201 09:05:13.301935 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b4c39a03-2938-4b10-b430-369066c79ce4-config-data" (OuterVolumeSpecName: "config-data") pod "b4c39a03-2938-4b10-b430-369066c79ce4" (UID: "b4c39a03-2938-4b10-b430-369066c79ce4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:05:13 crc kubenswrapper[4813]: I1201 09:05:13.362210 4813 reconciler_common.go:293] "Volume detached for volume \"run\" (UniqueName: \"kubernetes.io/host-path/b4c39a03-2938-4b10-b430-369066c79ce4-run\") on node \"crc\" DevicePath \"\"" Dec 01 09:05:13 crc kubenswrapper[4813]: I1201 09:05:13.362247 4813 reconciler_common.go:293] "Volume detached for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/b4c39a03-2938-4b10-b430-369066c79ce4-var-locks-brick\") on node \"crc\" DevicePath \"\"" Dec 01 09:05:13 crc kubenswrapper[4813]: I1201 09:05:13.362337 4813 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" " Dec 01 09:05:13 crc kubenswrapper[4813]: I1201 09:05:13.362349 4813 reconciler_common.go:293] "Volume detached for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/b4c39a03-2938-4b10-b430-369066c79ce4-etc-iscsi\") on node \"crc\" DevicePath \"\"" Dec 01 09:05:13 crc kubenswrapper[4813]: I1201 09:05:13.362360 4813 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b4c39a03-2938-4b10-b430-369066c79ce4-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 09:05:13 crc kubenswrapper[4813]: I1201 09:05:13.362368 4813 reconciler_common.go:293] "Volume detached for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/b4c39a03-2938-4b10-b430-369066c79ce4-dev\") on node \"crc\" DevicePath \"\"" Dec 01 09:05:13 crc kubenswrapper[4813]: I1201 09:05:13.362383 4813 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage14-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage14-crc\") on node \"crc\" " Dec 01 09:05:13 crc kubenswrapper[4813]: I1201 09:05:13.362391 4813 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b4c39a03-2938-4b10-b430-369066c79ce4-logs\") on node \"crc\" DevicePath \"\"" Dec 01 09:05:13 crc kubenswrapper[4813]: I1201 09:05:13.362399 4813 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b4c39a03-2938-4b10-b430-369066c79ce4-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 09:05:13 crc kubenswrapper[4813]: I1201 09:05:13.362407 4813 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b4c39a03-2938-4b10-b430-369066c79ce4-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 01 09:05:13 crc kubenswrapper[4813]: I1201 09:05:13.362419 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sgn7t\" (UniqueName: \"kubernetes.io/projected/b4c39a03-2938-4b10-b430-369066c79ce4-kube-api-access-sgn7t\") on node \"crc\" DevicePath \"\"" Dec 01 09:05:13 crc kubenswrapper[4813]: I1201 09:05:13.378084 4813 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage14-crc" (UniqueName: "kubernetes.io/local-volume/local-storage14-crc") on node "crc" Dec 01 09:05:13 crc kubenswrapper[4813]: I1201 09:05:13.387245 4813 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage04-crc" (UniqueName: "kubernetes.io/local-volume/local-storage04-crc") on node "crc" Dec 01 09:05:13 crc kubenswrapper[4813]: I1201 09:05:13.464154 4813 reconciler_common.go:293] "Volume detached for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" DevicePath \"\"" Dec 01 09:05:13 crc kubenswrapper[4813]: I1201 09:05:13.464187 4813 reconciler_common.go:293] "Volume detached for volume \"local-storage14-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage14-crc\") on node \"crc\" DevicePath \"\"" Dec 01 09:05:13 crc kubenswrapper[4813]: I1201 09:05:13.519317 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glancea9c4-account-delete-wl47f" Dec 01 09:05:13 crc kubenswrapper[4813]: I1201 09:05:13.558856 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-1" Dec 01 09:05:13 crc kubenswrapper[4813]: I1201 09:05:13.667822 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c079746c-0f0e-403a-bb6e-b1866a063050-logs\") pod \"c079746c-0f0e-403a-bb6e-b1866a063050\" (UID: \"c079746c-0f0e-403a-bb6e-b1866a063050\") " Dec 01 09:05:13 crc kubenswrapper[4813]: I1201 09:05:13.667935 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/c079746c-0f0e-403a-bb6e-b1866a063050-etc-nvme\") pod \"c079746c-0f0e-403a-bb6e-b1866a063050\" (UID: \"c079746c-0f0e-403a-bb6e-b1866a063050\") " Dec 01 09:05:13 crc kubenswrapper[4813]: I1201 09:05:13.668036 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/c079746c-0f0e-403a-bb6e-b1866a063050-sys\") pod \"c079746c-0f0e-403a-bb6e-b1866a063050\" (UID: \"c079746c-0f0e-403a-bb6e-b1866a063050\") " Dec 01 09:05:13 crc kubenswrapper[4813]: I1201 09:05:13.668087 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"c079746c-0f0e-403a-bb6e-b1866a063050\" (UID: \"c079746c-0f0e-403a-bb6e-b1866a063050\") " Dec 01 09:05:13 crc kubenswrapper[4813]: I1201 09:05:13.668148 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance-cache\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"c079746c-0f0e-403a-bb6e-b1866a063050\" (UID: \"c079746c-0f0e-403a-bb6e-b1866a063050\") " Dec 01 09:05:13 crc kubenswrapper[4813]: I1201 09:05:13.668224 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4r5z5\" (UniqueName: \"kubernetes.io/projected/54c06e84-7478-4fbc-9e2c-bdc6e1be658b-kube-api-access-4r5z5\") pod \"54c06e84-7478-4fbc-9e2c-bdc6e1be658b\" (UID: \"54c06e84-7478-4fbc-9e2c-bdc6e1be658b\") " Dec 01 09:05:13 crc kubenswrapper[4813]: I1201 09:05:13.668275 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c079746c-0f0e-403a-bb6e-b1866a063050-config-data\") pod \"c079746c-0f0e-403a-bb6e-b1866a063050\" (UID: \"c079746c-0f0e-403a-bb6e-b1866a063050\") " Dec 01 09:05:13 crc kubenswrapper[4813]: I1201 09:05:13.668276 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c079746c-0f0e-403a-bb6e-b1866a063050-sys" (OuterVolumeSpecName: "sys") pod "c079746c-0f0e-403a-bb6e-b1866a063050" (UID: "c079746c-0f0e-403a-bb6e-b1866a063050"). InnerVolumeSpecName "sys". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 09:05:13 crc kubenswrapper[4813]: I1201 09:05:13.668306 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c079746c-0f0e-403a-bb6e-b1866a063050-scripts\") pod \"c079746c-0f0e-403a-bb6e-b1866a063050\" (UID: \"c079746c-0f0e-403a-bb6e-b1866a063050\") " Dec 01 09:05:13 crc kubenswrapper[4813]: I1201 09:05:13.668339 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/54c06e84-7478-4fbc-9e2c-bdc6e1be658b-operator-scripts\") pod \"54c06e84-7478-4fbc-9e2c-bdc6e1be658b\" (UID: \"54c06e84-7478-4fbc-9e2c-bdc6e1be658b\") " Dec 01 09:05:13 crc kubenswrapper[4813]: I1201 09:05:13.668393 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s6459\" (UniqueName: \"kubernetes.io/projected/c079746c-0f0e-403a-bb6e-b1866a063050-kube-api-access-s6459\") pod \"c079746c-0f0e-403a-bb6e-b1866a063050\" (UID: \"c079746c-0f0e-403a-bb6e-b1866a063050\") " Dec 01 09:05:13 crc kubenswrapper[4813]: I1201 09:05:13.668428 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/c079746c-0f0e-403a-bb6e-b1866a063050-etc-iscsi\") pod \"c079746c-0f0e-403a-bb6e-b1866a063050\" (UID: \"c079746c-0f0e-403a-bb6e-b1866a063050\") " Dec 01 09:05:13 crc kubenswrapper[4813]: I1201 09:05:13.668455 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c079746c-0f0e-403a-bb6e-b1866a063050-logs" (OuterVolumeSpecName: "logs") pod "c079746c-0f0e-403a-bb6e-b1866a063050" (UID: "c079746c-0f0e-403a-bb6e-b1866a063050"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 09:05:13 crc kubenswrapper[4813]: I1201 09:05:13.668475 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/c079746c-0f0e-403a-bb6e-b1866a063050-httpd-run\") pod \"c079746c-0f0e-403a-bb6e-b1866a063050\" (UID: \"c079746c-0f0e-403a-bb6e-b1866a063050\") " Dec 01 09:05:13 crc kubenswrapper[4813]: I1201 09:05:13.668433 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c079746c-0f0e-403a-bb6e-b1866a063050-etc-nvme" (OuterVolumeSpecName: "etc-nvme") pod "c079746c-0f0e-403a-bb6e-b1866a063050" (UID: "c079746c-0f0e-403a-bb6e-b1866a063050"). InnerVolumeSpecName "etc-nvme". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 09:05:13 crc kubenswrapper[4813]: I1201 09:05:13.668506 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/c079746c-0f0e-403a-bb6e-b1866a063050-run\") pod \"c079746c-0f0e-403a-bb6e-b1866a063050\" (UID: \"c079746c-0f0e-403a-bb6e-b1866a063050\") " Dec 01 09:05:13 crc kubenswrapper[4813]: I1201 09:05:13.668547 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/c079746c-0f0e-403a-bb6e-b1866a063050-lib-modules\") pod \"c079746c-0f0e-403a-bb6e-b1866a063050\" (UID: \"c079746c-0f0e-403a-bb6e-b1866a063050\") " Dec 01 09:05:13 crc kubenswrapper[4813]: I1201 09:05:13.668605 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/c079746c-0f0e-403a-bb6e-b1866a063050-var-locks-brick\") pod \"c079746c-0f0e-403a-bb6e-b1866a063050\" (UID: \"c079746c-0f0e-403a-bb6e-b1866a063050\") " Dec 01 09:05:13 crc kubenswrapper[4813]: I1201 09:05:13.668664 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/c079746c-0f0e-403a-bb6e-b1866a063050-dev\") pod \"c079746c-0f0e-403a-bb6e-b1866a063050\" (UID: \"c079746c-0f0e-403a-bb6e-b1866a063050\") " Dec 01 09:05:13 crc kubenswrapper[4813]: I1201 09:05:13.668811 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c079746c-0f0e-403a-bb6e-b1866a063050-run" (OuterVolumeSpecName: "run") pod "c079746c-0f0e-403a-bb6e-b1866a063050" (UID: "c079746c-0f0e-403a-bb6e-b1866a063050"). InnerVolumeSpecName "run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 09:05:13 crc kubenswrapper[4813]: I1201 09:05:13.668874 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c079746c-0f0e-403a-bb6e-b1866a063050-etc-iscsi" (OuterVolumeSpecName: "etc-iscsi") pod "c079746c-0f0e-403a-bb6e-b1866a063050" (UID: "c079746c-0f0e-403a-bb6e-b1866a063050"). InnerVolumeSpecName "etc-iscsi". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 09:05:13 crc kubenswrapper[4813]: I1201 09:05:13.668952 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c079746c-0f0e-403a-bb6e-b1866a063050-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "c079746c-0f0e-403a-bb6e-b1866a063050" (UID: "c079746c-0f0e-403a-bb6e-b1866a063050"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 09:05:13 crc kubenswrapper[4813]: I1201 09:05:13.669041 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c079746c-0f0e-403a-bb6e-b1866a063050-var-locks-brick" (OuterVolumeSpecName: "var-locks-brick") pod "c079746c-0f0e-403a-bb6e-b1866a063050" (UID: "c079746c-0f0e-403a-bb6e-b1866a063050"). InnerVolumeSpecName "var-locks-brick". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 09:05:13 crc kubenswrapper[4813]: I1201 09:05:13.669091 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c079746c-0f0e-403a-bb6e-b1866a063050-dev" (OuterVolumeSpecName: "dev") pod "c079746c-0f0e-403a-bb6e-b1866a063050" (UID: "c079746c-0f0e-403a-bb6e-b1866a063050"). InnerVolumeSpecName "dev". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 09:05:13 crc kubenswrapper[4813]: I1201 09:05:13.669597 4813 reconciler_common.go:293] "Volume detached for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/c079746c-0f0e-403a-bb6e-b1866a063050-etc-iscsi\") on node \"crc\" DevicePath \"\"" Dec 01 09:05:13 crc kubenswrapper[4813]: I1201 09:05:13.669641 4813 reconciler_common.go:293] "Volume detached for volume \"run\" (UniqueName: \"kubernetes.io/host-path/c079746c-0f0e-403a-bb6e-b1866a063050-run\") on node \"crc\" DevicePath \"\"" Dec 01 09:05:13 crc kubenswrapper[4813]: I1201 09:05:13.669660 4813 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/c079746c-0f0e-403a-bb6e-b1866a063050-lib-modules\") on node \"crc\" DevicePath \"\"" Dec 01 09:05:13 crc kubenswrapper[4813]: I1201 09:05:13.669678 4813 reconciler_common.go:293] "Volume detached for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/c079746c-0f0e-403a-bb6e-b1866a063050-var-locks-brick\") on node \"crc\" DevicePath \"\"" Dec 01 09:05:13 crc kubenswrapper[4813]: I1201 09:05:13.669696 4813 reconciler_common.go:293] "Volume detached for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/c079746c-0f0e-403a-bb6e-b1866a063050-dev\") on node \"crc\" DevicePath \"\"" Dec 01 09:05:13 crc kubenswrapper[4813]: I1201 09:05:13.669713 4813 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c079746c-0f0e-403a-bb6e-b1866a063050-logs\") on node \"crc\" DevicePath \"\"" Dec 01 09:05:13 crc kubenswrapper[4813]: I1201 09:05:13.669727 4813 reconciler_common.go:293] "Volume detached for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/c079746c-0f0e-403a-bb6e-b1866a063050-etc-nvme\") on node \"crc\" DevicePath \"\"" Dec 01 09:05:13 crc kubenswrapper[4813]: I1201 09:05:13.669742 4813 reconciler_common.go:293] "Volume detached for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/c079746c-0f0e-403a-bb6e-b1866a063050-sys\") on node \"crc\" DevicePath \"\"" Dec 01 09:05:13 crc kubenswrapper[4813]: I1201 09:05:13.670058 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/54c06e84-7478-4fbc-9e2c-bdc6e1be658b-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "54c06e84-7478-4fbc-9e2c-bdc6e1be658b" (UID: "54c06e84-7478-4fbc-9e2c-bdc6e1be658b"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 09:05:13 crc kubenswrapper[4813]: I1201 09:05:13.670725 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c079746c-0f0e-403a-bb6e-b1866a063050-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "c079746c-0f0e-403a-bb6e-b1866a063050" (UID: "c079746c-0f0e-403a-bb6e-b1866a063050"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 09:05:13 crc kubenswrapper[4813]: I1201 09:05:13.673142 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage05-crc" (OuterVolumeSpecName: "glance") pod "c079746c-0f0e-403a-bb6e-b1866a063050" (UID: "c079746c-0f0e-403a-bb6e-b1866a063050"). InnerVolumeSpecName "local-storage05-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 01 09:05:13 crc kubenswrapper[4813]: I1201 09:05:13.673188 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/54c06e84-7478-4fbc-9e2c-bdc6e1be658b-kube-api-access-4r5z5" (OuterVolumeSpecName: "kube-api-access-4r5z5") pod "54c06e84-7478-4fbc-9e2c-bdc6e1be658b" (UID: "54c06e84-7478-4fbc-9e2c-bdc6e1be658b"). InnerVolumeSpecName "kube-api-access-4r5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:05:13 crc kubenswrapper[4813]: I1201 09:05:13.673736 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c079746c-0f0e-403a-bb6e-b1866a063050-kube-api-access-s6459" (OuterVolumeSpecName: "kube-api-access-s6459") pod "c079746c-0f0e-403a-bb6e-b1866a063050" (UID: "c079746c-0f0e-403a-bb6e-b1866a063050"). InnerVolumeSpecName "kube-api-access-s6459". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:05:13 crc kubenswrapper[4813]: I1201 09:05:13.673945 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage01-crc" (OuterVolumeSpecName: "glance-cache") pod "c079746c-0f0e-403a-bb6e-b1866a063050" (UID: "c079746c-0f0e-403a-bb6e-b1866a063050"). InnerVolumeSpecName "local-storage01-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 01 09:05:13 crc kubenswrapper[4813]: I1201 09:05:13.674638 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c079746c-0f0e-403a-bb6e-b1866a063050-scripts" (OuterVolumeSpecName: "scripts") pod "c079746c-0f0e-403a-bb6e-b1866a063050" (UID: "c079746c-0f0e-403a-bb6e-b1866a063050"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:05:13 crc kubenswrapper[4813]: I1201 09:05:13.747800 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c079746c-0f0e-403a-bb6e-b1866a063050-config-data" (OuterVolumeSpecName: "config-data") pod "c079746c-0f0e-403a-bb6e-b1866a063050" (UID: "c079746c-0f0e-403a-bb6e-b1866a063050"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:05:13 crc kubenswrapper[4813]: I1201 09:05:13.771190 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4r5z5\" (UniqueName: \"kubernetes.io/projected/54c06e84-7478-4fbc-9e2c-bdc6e1be658b-kube-api-access-4r5z5\") on node \"crc\" DevicePath \"\"" Dec 01 09:05:13 crc kubenswrapper[4813]: I1201 09:05:13.771227 4813 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c079746c-0f0e-403a-bb6e-b1866a063050-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 09:05:13 crc kubenswrapper[4813]: I1201 09:05:13.771242 4813 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c079746c-0f0e-403a-bb6e-b1866a063050-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 09:05:13 crc kubenswrapper[4813]: I1201 09:05:13.771253 4813 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/54c06e84-7478-4fbc-9e2c-bdc6e1be658b-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 09:05:13 crc kubenswrapper[4813]: I1201 09:05:13.772070 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s6459\" (UniqueName: \"kubernetes.io/projected/c079746c-0f0e-403a-bb6e-b1866a063050-kube-api-access-s6459\") on node \"crc\" DevicePath \"\"" Dec 01 09:05:13 crc kubenswrapper[4813]: I1201 09:05:13.772142 4813 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/c079746c-0f0e-403a-bb6e-b1866a063050-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 01 09:05:13 crc kubenswrapper[4813]: I1201 09:05:13.772205 4813 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" " Dec 01 09:05:13 crc kubenswrapper[4813]: I1201 09:05:13.772235 4813 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" " Dec 01 09:05:13 crc kubenswrapper[4813]: I1201 09:05:13.789854 4813 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage05-crc" (UniqueName: "kubernetes.io/local-volume/local-storage05-crc") on node "crc" Dec 01 09:05:13 crc kubenswrapper[4813]: I1201 09:05:13.796558 4813 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage01-crc" (UniqueName: "kubernetes.io/local-volume/local-storage01-crc") on node "crc" Dec 01 09:05:13 crc kubenswrapper[4813]: I1201 09:05:13.874036 4813 reconciler_common.go:293] "Volume detached for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" DevicePath \"\"" Dec 01 09:05:13 crc kubenswrapper[4813]: I1201 09:05:13.874100 4813 reconciler_common.go:293] "Volume detached for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" DevicePath \"\"" Dec 01 09:05:14 crc kubenswrapper[4813]: I1201 09:05:14.141020 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glancea9c4-account-delete-wl47f" event={"ID":"54c06e84-7478-4fbc-9e2c-bdc6e1be658b","Type":"ContainerDied","Data":"17d938babbce43861ca8450a7ec26271d9ef23e64fde291843ab3616d7793eac"} Dec 01 09:05:14 crc kubenswrapper[4813]: I1201 09:05:14.141115 4813 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="17d938babbce43861ca8450a7ec26271d9ef23e64fde291843ab3616d7793eac" Dec 01 09:05:14 crc kubenswrapper[4813]: I1201 09:05:14.141123 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glancea9c4-account-delete-wl47f" Dec 01 09:05:14 crc kubenswrapper[4813]: I1201 09:05:14.146722 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-1" event={"ID":"c079746c-0f0e-403a-bb6e-b1866a063050","Type":"ContainerDied","Data":"597e9606975353626d3b79d9dd36fc9fcfb029ca4c8d3da2851ae76fe368b589"} Dec 01 09:05:14 crc kubenswrapper[4813]: I1201 09:05:14.146737 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-1" Dec 01 09:05:14 crc kubenswrapper[4813]: I1201 09:05:14.146809 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-1" Dec 01 09:05:14 crc kubenswrapper[4813]: I1201 09:05:14.147276 4813 scope.go:117] "RemoveContainer" containerID="d191ad43cbe03d01275077b6d3666ca27b942e2e476948af348780daa2824a34" Dec 01 09:05:14 crc kubenswrapper[4813]: I1201 09:05:14.197481 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-1"] Dec 01 09:05:14 crc kubenswrapper[4813]: I1201 09:05:14.204689 4813 scope.go:117] "RemoveContainer" containerID="d0e434dc9af434eac9e19163ad506f93f21a1b45f5335294ea9b9a8d72872b0f" Dec 01 09:05:14 crc kubenswrapper[4813]: I1201 09:05:14.209773 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-1"] Dec 01 09:05:14 crc kubenswrapper[4813]: I1201 09:05:14.224871 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-1"] Dec 01 09:05:14 crc kubenswrapper[4813]: I1201 09:05:14.230754 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-1"] Dec 01 09:05:14 crc kubenswrapper[4813]: I1201 09:05:14.416868 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b4c39a03-2938-4b10-b430-369066c79ce4" path="/var/lib/kubelet/pods/b4c39a03-2938-4b10-b430-369066c79ce4/volumes" Dec 01 09:05:14 crc kubenswrapper[4813]: I1201 09:05:14.417835 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c079746c-0f0e-403a-bb6e-b1866a063050" path="/var/lib/kubelet/pods/c079746c-0f0e-403a-bb6e-b1866a063050/volumes" Dec 01 09:05:14 crc kubenswrapper[4813]: I1201 09:05:14.793635 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-0" Dec 01 09:05:14 crc kubenswrapper[4813]: I1201 09:05:14.794737 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/d6f4323d-561b-4725-a1fe-140db5f6a33d-dev\") pod \"d6f4323d-561b-4725-a1fe-140db5f6a33d\" (UID: \"d6f4323d-561b-4725-a1fe-140db5f6a33d\") " Dec 01 09:05:14 crc kubenswrapper[4813]: I1201 09:05:14.794818 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance-cache\" (UniqueName: \"kubernetes.io/local-volume/local-storage16-crc\") pod \"d6f4323d-561b-4725-a1fe-140db5f6a33d\" (UID: \"d6f4323d-561b-4725-a1fe-140db5f6a33d\") " Dec 01 09:05:14 crc kubenswrapper[4813]: I1201 09:05:14.794853 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/d6f4323d-561b-4725-a1fe-140db5f6a33d-run\") pod \"d6f4323d-561b-4725-a1fe-140db5f6a33d\" (UID: \"d6f4323d-561b-4725-a1fe-140db5f6a33d\") " Dec 01 09:05:14 crc kubenswrapper[4813]: I1201 09:05:14.794880 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d6f4323d-561b-4725-a1fe-140db5f6a33d-dev" (OuterVolumeSpecName: "dev") pod "d6f4323d-561b-4725-a1fe-140db5f6a33d" (UID: "d6f4323d-561b-4725-a1fe-140db5f6a33d"). InnerVolumeSpecName "dev". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 09:05:14 crc kubenswrapper[4813]: I1201 09:05:14.794903 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/d6f4323d-561b-4725-a1fe-140db5f6a33d-sys\") pod \"d6f4323d-561b-4725-a1fe-140db5f6a33d\" (UID: \"d6f4323d-561b-4725-a1fe-140db5f6a33d\") " Dec 01 09:05:14 crc kubenswrapper[4813]: I1201 09:05:14.794955 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d6f4323d-561b-4725-a1fe-140db5f6a33d-sys" (OuterVolumeSpecName: "sys") pod "d6f4323d-561b-4725-a1fe-140db5f6a33d" (UID: "d6f4323d-561b-4725-a1fe-140db5f6a33d"). InnerVolumeSpecName "sys". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 09:05:14 crc kubenswrapper[4813]: I1201 09:05:14.795045 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d6f4323d-561b-4725-a1fe-140db5f6a33d-run" (OuterVolumeSpecName: "run") pod "d6f4323d-561b-4725-a1fe-140db5f6a33d" (UID: "d6f4323d-561b-4725-a1fe-140db5f6a33d"). InnerVolumeSpecName "run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 09:05:14 crc kubenswrapper[4813]: I1201 09:05:14.795066 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d6f4323d-561b-4725-a1fe-140db5f6a33d-httpd-run\") pod \"d6f4323d-561b-4725-a1fe-140db5f6a33d\" (UID: \"d6f4323d-561b-4725-a1fe-140db5f6a33d\") " Dec 01 09:05:14 crc kubenswrapper[4813]: I1201 09:05:14.795171 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/d6f4323d-561b-4725-a1fe-140db5f6a33d-etc-nvme\") pod \"d6f4323d-561b-4725-a1fe-140db5f6a33d\" (UID: \"d6f4323d-561b-4725-a1fe-140db5f6a33d\") " Dec 01 09:05:14 crc kubenswrapper[4813]: I1201 09:05:14.795213 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d6f4323d-561b-4725-a1fe-140db5f6a33d-logs\") pod \"d6f4323d-561b-4725-a1fe-140db5f6a33d\" (UID: \"d6f4323d-561b-4725-a1fe-140db5f6a33d\") " Dec 01 09:05:14 crc kubenswrapper[4813]: I1201 09:05:14.795299 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d6f4323d-561b-4725-a1fe-140db5f6a33d-etc-nvme" (OuterVolumeSpecName: "etc-nvme") pod "d6f4323d-561b-4725-a1fe-140db5f6a33d" (UID: "d6f4323d-561b-4725-a1fe-140db5f6a33d"). InnerVolumeSpecName "etc-nvme". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 09:05:14 crc kubenswrapper[4813]: I1201 09:05:14.795321 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d6f4323d-561b-4725-a1fe-140db5f6a33d-config-data\") pod \"d6f4323d-561b-4725-a1fe-140db5f6a33d\" (UID: \"d6f4323d-561b-4725-a1fe-140db5f6a33d\") " Dec 01 09:05:14 crc kubenswrapper[4813]: I1201 09:05:14.795401 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d6f4323d-561b-4725-a1fe-140db5f6a33d-scripts\") pod \"d6f4323d-561b-4725-a1fe-140db5f6a33d\" (UID: \"d6f4323d-561b-4725-a1fe-140db5f6a33d\") " Dec 01 09:05:14 crc kubenswrapper[4813]: I1201 09:05:14.795432 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/d6f4323d-561b-4725-a1fe-140db5f6a33d-var-locks-brick\") pod \"d6f4323d-561b-4725-a1fe-140db5f6a33d\" (UID: \"d6f4323d-561b-4725-a1fe-140db5f6a33d\") " Dec 01 09:05:14 crc kubenswrapper[4813]: I1201 09:05:14.795488 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage20-crc\") pod \"d6f4323d-561b-4725-a1fe-140db5f6a33d\" (UID: \"d6f4323d-561b-4725-a1fe-140db5f6a33d\") " Dec 01 09:05:14 crc kubenswrapper[4813]: I1201 09:05:14.795558 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/d6f4323d-561b-4725-a1fe-140db5f6a33d-lib-modules\") pod \"d6f4323d-561b-4725-a1fe-140db5f6a33d\" (UID: \"d6f4323d-561b-4725-a1fe-140db5f6a33d\") " Dec 01 09:05:14 crc kubenswrapper[4813]: I1201 09:05:14.795590 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jtfbh\" (UniqueName: \"kubernetes.io/projected/d6f4323d-561b-4725-a1fe-140db5f6a33d-kube-api-access-jtfbh\") pod \"d6f4323d-561b-4725-a1fe-140db5f6a33d\" (UID: \"d6f4323d-561b-4725-a1fe-140db5f6a33d\") " Dec 01 09:05:14 crc kubenswrapper[4813]: I1201 09:05:14.795669 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/d6f4323d-561b-4725-a1fe-140db5f6a33d-etc-iscsi\") pod \"d6f4323d-561b-4725-a1fe-140db5f6a33d\" (UID: \"d6f4323d-561b-4725-a1fe-140db5f6a33d\") " Dec 01 09:05:14 crc kubenswrapper[4813]: I1201 09:05:14.795834 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d6f4323d-561b-4725-a1fe-140db5f6a33d-logs" (OuterVolumeSpecName: "logs") pod "d6f4323d-561b-4725-a1fe-140db5f6a33d" (UID: "d6f4323d-561b-4725-a1fe-140db5f6a33d"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 09:05:14 crc kubenswrapper[4813]: I1201 09:05:14.795908 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d6f4323d-561b-4725-a1fe-140db5f6a33d-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "d6f4323d-561b-4725-a1fe-140db5f6a33d" (UID: "d6f4323d-561b-4725-a1fe-140db5f6a33d"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 09:05:14 crc kubenswrapper[4813]: I1201 09:05:14.795983 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d6f4323d-561b-4725-a1fe-140db5f6a33d-var-locks-brick" (OuterVolumeSpecName: "var-locks-brick") pod "d6f4323d-561b-4725-a1fe-140db5f6a33d" (UID: "d6f4323d-561b-4725-a1fe-140db5f6a33d"). InnerVolumeSpecName "var-locks-brick". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 09:05:14 crc kubenswrapper[4813]: I1201 09:05:14.796120 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d6f4323d-561b-4725-a1fe-140db5f6a33d-etc-iscsi" (OuterVolumeSpecName: "etc-iscsi") pod "d6f4323d-561b-4725-a1fe-140db5f6a33d" (UID: "d6f4323d-561b-4725-a1fe-140db5f6a33d"). InnerVolumeSpecName "etc-iscsi". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 09:05:14 crc kubenswrapper[4813]: I1201 09:05:14.796249 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d6f4323d-561b-4725-a1fe-140db5f6a33d-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "d6f4323d-561b-4725-a1fe-140db5f6a33d" (UID: "d6f4323d-561b-4725-a1fe-140db5f6a33d"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 09:05:14 crc kubenswrapper[4813]: I1201 09:05:14.797119 4813 reconciler_common.go:293] "Volume detached for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/d6f4323d-561b-4725-a1fe-140db5f6a33d-var-locks-brick\") on node \"crc\" DevicePath \"\"" Dec 01 09:05:14 crc kubenswrapper[4813]: I1201 09:05:14.797142 4813 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/d6f4323d-561b-4725-a1fe-140db5f6a33d-lib-modules\") on node \"crc\" DevicePath \"\"" Dec 01 09:05:14 crc kubenswrapper[4813]: I1201 09:05:14.797152 4813 reconciler_common.go:293] "Volume detached for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/d6f4323d-561b-4725-a1fe-140db5f6a33d-etc-iscsi\") on node \"crc\" DevicePath \"\"" Dec 01 09:05:14 crc kubenswrapper[4813]: I1201 09:05:14.797163 4813 reconciler_common.go:293] "Volume detached for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/d6f4323d-561b-4725-a1fe-140db5f6a33d-dev\") on node \"crc\" DevicePath \"\"" Dec 01 09:05:14 crc kubenswrapper[4813]: I1201 09:05:14.797175 4813 reconciler_common.go:293] "Volume detached for volume \"run\" (UniqueName: \"kubernetes.io/host-path/d6f4323d-561b-4725-a1fe-140db5f6a33d-run\") on node \"crc\" DevicePath \"\"" Dec 01 09:05:14 crc kubenswrapper[4813]: I1201 09:05:14.797191 4813 reconciler_common.go:293] "Volume detached for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/d6f4323d-561b-4725-a1fe-140db5f6a33d-sys\") on node \"crc\" DevicePath \"\"" Dec 01 09:05:14 crc kubenswrapper[4813]: I1201 09:05:14.797204 4813 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d6f4323d-561b-4725-a1fe-140db5f6a33d-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 01 09:05:14 crc kubenswrapper[4813]: I1201 09:05:14.797213 4813 reconciler_common.go:293] "Volume detached for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/d6f4323d-561b-4725-a1fe-140db5f6a33d-etc-nvme\") on node \"crc\" DevicePath \"\"" Dec 01 09:05:14 crc kubenswrapper[4813]: I1201 09:05:14.797222 4813 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d6f4323d-561b-4725-a1fe-140db5f6a33d-logs\") on node \"crc\" DevicePath \"\"" Dec 01 09:05:14 crc kubenswrapper[4813]: I1201 09:05:14.800163 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d6f4323d-561b-4725-a1fe-140db5f6a33d-kube-api-access-jtfbh" (OuterVolumeSpecName: "kube-api-access-jtfbh") pod "d6f4323d-561b-4725-a1fe-140db5f6a33d" (UID: "d6f4323d-561b-4725-a1fe-140db5f6a33d"). InnerVolumeSpecName "kube-api-access-jtfbh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:05:14 crc kubenswrapper[4813]: I1201 09:05:14.801122 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage20-crc" (OuterVolumeSpecName: "glance") pod "d6f4323d-561b-4725-a1fe-140db5f6a33d" (UID: "d6f4323d-561b-4725-a1fe-140db5f6a33d"). InnerVolumeSpecName "local-storage20-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 01 09:05:14 crc kubenswrapper[4813]: I1201 09:05:14.802495 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d6f4323d-561b-4725-a1fe-140db5f6a33d-scripts" (OuterVolumeSpecName: "scripts") pod "d6f4323d-561b-4725-a1fe-140db5f6a33d" (UID: "d6f4323d-561b-4725-a1fe-140db5f6a33d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:05:14 crc kubenswrapper[4813]: I1201 09:05:14.803810 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage16-crc" (OuterVolumeSpecName: "glance-cache") pod "d6f4323d-561b-4725-a1fe-140db5f6a33d" (UID: "d6f4323d-561b-4725-a1fe-140db5f6a33d"). InnerVolumeSpecName "local-storage16-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 01 09:05:14 crc kubenswrapper[4813]: I1201 09:05:14.841864 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d6f4323d-561b-4725-a1fe-140db5f6a33d-config-data" (OuterVolumeSpecName: "config-data") pod "d6f4323d-561b-4725-a1fe-140db5f6a33d" (UID: "d6f4323d-561b-4725-a1fe-140db5f6a33d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:05:14 crc kubenswrapper[4813]: I1201 09:05:14.862912 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:05:14 crc kubenswrapper[4813]: I1201 09:05:14.898700 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/522f36f6-9da6-46b0-8ca6-9753311b3dfc-scripts\") pod \"522f36f6-9da6-46b0-8ca6-9753311b3dfc\" (UID: \"522f36f6-9da6-46b0-8ca6-9753311b3dfc\") " Dec 01 09:05:14 crc kubenswrapper[4813]: I1201 09:05:14.898803 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/522f36f6-9da6-46b0-8ca6-9753311b3dfc-dev\") pod \"522f36f6-9da6-46b0-8ca6-9753311b3dfc\" (UID: \"522f36f6-9da6-46b0-8ca6-9753311b3dfc\") " Dec 01 09:05:14 crc kubenswrapper[4813]: I1201 09:05:14.898844 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/522f36f6-9da6-46b0-8ca6-9753311b3dfc-lib-modules\") pod \"522f36f6-9da6-46b0-8ca6-9753311b3dfc\" (UID: \"522f36f6-9da6-46b0-8ca6-9753311b3dfc\") " Dec 01 09:05:14 crc kubenswrapper[4813]: I1201 09:05:14.898877 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/522f36f6-9da6-46b0-8ca6-9753311b3dfc-config-data\") pod \"522f36f6-9da6-46b0-8ca6-9753311b3dfc\" (UID: \"522f36f6-9da6-46b0-8ca6-9753311b3dfc\") " Dec 01 09:05:14 crc kubenswrapper[4813]: I1201 09:05:14.898918 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/522f36f6-9da6-46b0-8ca6-9753311b3dfc-logs\") pod \"522f36f6-9da6-46b0-8ca6-9753311b3dfc\" (UID: \"522f36f6-9da6-46b0-8ca6-9753311b3dfc\") " Dec 01 09:05:14 crc kubenswrapper[4813]: I1201 09:05:14.898950 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/522f36f6-9da6-46b0-8ca6-9753311b3dfc-etc-nvme\") pod \"522f36f6-9da6-46b0-8ca6-9753311b3dfc\" (UID: \"522f36f6-9da6-46b0-8ca6-9753311b3dfc\") " Dec 01 09:05:14 crc kubenswrapper[4813]: I1201 09:05:14.898990 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"522f36f6-9da6-46b0-8ca6-9753311b3dfc\" (UID: \"522f36f6-9da6-46b0-8ca6-9753311b3dfc\") " Dec 01 09:05:14 crc kubenswrapper[4813]: I1201 09:05:14.899031 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/522f36f6-9da6-46b0-8ca6-9753311b3dfc-etc-iscsi\") pod \"522f36f6-9da6-46b0-8ca6-9753311b3dfc\" (UID: \"522f36f6-9da6-46b0-8ca6-9753311b3dfc\") " Dec 01 09:05:14 crc kubenswrapper[4813]: I1201 09:05:14.899062 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/522f36f6-9da6-46b0-8ca6-9753311b3dfc-sys\") pod \"522f36f6-9da6-46b0-8ca6-9753311b3dfc\" (UID: \"522f36f6-9da6-46b0-8ca6-9753311b3dfc\") " Dec 01 09:05:14 crc kubenswrapper[4813]: I1201 09:05:14.899088 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/522f36f6-9da6-46b0-8ca6-9753311b3dfc-etc-nvme" (OuterVolumeSpecName: "etc-nvme") pod "522f36f6-9da6-46b0-8ca6-9753311b3dfc" (UID: "522f36f6-9da6-46b0-8ca6-9753311b3dfc"). InnerVolumeSpecName "etc-nvme". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 09:05:14 crc kubenswrapper[4813]: I1201 09:05:14.899104 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/522f36f6-9da6-46b0-8ca6-9753311b3dfc-httpd-run\") pod \"522f36f6-9da6-46b0-8ca6-9753311b3dfc\" (UID: \"522f36f6-9da6-46b0-8ca6-9753311b3dfc\") " Dec 01 09:05:14 crc kubenswrapper[4813]: I1201 09:05:14.899144 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/522f36f6-9da6-46b0-8ca6-9753311b3dfc-etc-iscsi" (OuterVolumeSpecName: "etc-iscsi") pod "522f36f6-9da6-46b0-8ca6-9753311b3dfc" (UID: "522f36f6-9da6-46b0-8ca6-9753311b3dfc"). InnerVolumeSpecName "etc-iscsi". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 09:05:14 crc kubenswrapper[4813]: I1201 09:05:14.899146 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/522f36f6-9da6-46b0-8ca6-9753311b3dfc-run\") pod \"522f36f6-9da6-46b0-8ca6-9753311b3dfc\" (UID: \"522f36f6-9da6-46b0-8ca6-9753311b3dfc\") " Dec 01 09:05:14 crc kubenswrapper[4813]: I1201 09:05:14.899176 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/522f36f6-9da6-46b0-8ca6-9753311b3dfc-run" (OuterVolumeSpecName: "run") pod "522f36f6-9da6-46b0-8ca6-9753311b3dfc" (UID: "522f36f6-9da6-46b0-8ca6-9753311b3dfc"). InnerVolumeSpecName "run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 09:05:14 crc kubenswrapper[4813]: I1201 09:05:14.899189 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v7h7b\" (UniqueName: \"kubernetes.io/projected/522f36f6-9da6-46b0-8ca6-9753311b3dfc-kube-api-access-v7h7b\") pod \"522f36f6-9da6-46b0-8ca6-9753311b3dfc\" (UID: \"522f36f6-9da6-46b0-8ca6-9753311b3dfc\") " Dec 01 09:05:14 crc kubenswrapper[4813]: I1201 09:05:14.899206 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/522f36f6-9da6-46b0-8ca6-9753311b3dfc-sys" (OuterVolumeSpecName: "sys") pod "522f36f6-9da6-46b0-8ca6-9753311b3dfc" (UID: "522f36f6-9da6-46b0-8ca6-9753311b3dfc"). InnerVolumeSpecName "sys". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 09:05:14 crc kubenswrapper[4813]: I1201 09:05:14.899256 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance-cache\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"522f36f6-9da6-46b0-8ca6-9753311b3dfc\" (UID: \"522f36f6-9da6-46b0-8ca6-9753311b3dfc\") " Dec 01 09:05:14 crc kubenswrapper[4813]: I1201 09:05:14.899310 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/522f36f6-9da6-46b0-8ca6-9753311b3dfc-var-locks-brick\") pod \"522f36f6-9da6-46b0-8ca6-9753311b3dfc\" (UID: \"522f36f6-9da6-46b0-8ca6-9753311b3dfc\") " Dec 01 09:05:14 crc kubenswrapper[4813]: I1201 09:05:14.899460 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/522f36f6-9da6-46b0-8ca6-9753311b3dfc-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "522f36f6-9da6-46b0-8ca6-9753311b3dfc" (UID: "522f36f6-9da6-46b0-8ca6-9753311b3dfc"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 09:05:14 crc kubenswrapper[4813]: I1201 09:05:14.899519 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/522f36f6-9da6-46b0-8ca6-9753311b3dfc-logs" (OuterVolumeSpecName: "logs") pod "522f36f6-9da6-46b0-8ca6-9753311b3dfc" (UID: "522f36f6-9da6-46b0-8ca6-9753311b3dfc"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 09:05:14 crc kubenswrapper[4813]: I1201 09:05:14.899578 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/522f36f6-9da6-46b0-8ca6-9753311b3dfc-var-locks-brick" (OuterVolumeSpecName: "var-locks-brick") pod "522f36f6-9da6-46b0-8ca6-9753311b3dfc" (UID: "522f36f6-9da6-46b0-8ca6-9753311b3dfc"). InnerVolumeSpecName "var-locks-brick". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 09:05:14 crc kubenswrapper[4813]: I1201 09:05:14.899755 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/522f36f6-9da6-46b0-8ca6-9753311b3dfc-dev" (OuterVolumeSpecName: "dev") pod "522f36f6-9da6-46b0-8ca6-9753311b3dfc" (UID: "522f36f6-9da6-46b0-8ca6-9753311b3dfc"). InnerVolumeSpecName "dev". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 09:05:14 crc kubenswrapper[4813]: I1201 09:05:14.900019 4813 reconciler_common.go:293] "Volume detached for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/522f36f6-9da6-46b0-8ca6-9753311b3dfc-var-locks-brick\") on node \"crc\" DevicePath \"\"" Dec 01 09:05:14 crc kubenswrapper[4813]: I1201 09:05:14.900095 4813 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage16-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage16-crc\") on node \"crc\" " Dec 01 09:05:14 crc kubenswrapper[4813]: I1201 09:05:14.900114 4813 reconciler_common.go:293] "Volume detached for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/522f36f6-9da6-46b0-8ca6-9753311b3dfc-dev\") on node \"crc\" DevicePath \"\"" Dec 01 09:05:14 crc kubenswrapper[4813]: I1201 09:05:14.900127 4813 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d6f4323d-561b-4725-a1fe-140db5f6a33d-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 09:05:14 crc kubenswrapper[4813]: I1201 09:05:14.900159 4813 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/522f36f6-9da6-46b0-8ca6-9753311b3dfc-logs\") on node \"crc\" DevicePath \"\"" Dec 01 09:05:14 crc kubenswrapper[4813]: I1201 09:05:14.900169 4813 reconciler_common.go:293] "Volume detached for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/522f36f6-9da6-46b0-8ca6-9753311b3dfc-etc-nvme\") on node \"crc\" DevicePath \"\"" Dec 01 09:05:14 crc kubenswrapper[4813]: I1201 09:05:14.900177 4813 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d6f4323d-561b-4725-a1fe-140db5f6a33d-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 09:05:14 crc kubenswrapper[4813]: I1201 09:05:14.900185 4813 reconciler_common.go:293] "Volume detached for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/522f36f6-9da6-46b0-8ca6-9753311b3dfc-etc-iscsi\") on node \"crc\" DevicePath \"\"" Dec 01 09:05:14 crc kubenswrapper[4813]: I1201 09:05:14.900193 4813 reconciler_common.go:293] "Volume detached for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/522f36f6-9da6-46b0-8ca6-9753311b3dfc-sys\") on node \"crc\" DevicePath \"\"" Dec 01 09:05:14 crc kubenswrapper[4813]: I1201 09:05:14.900213 4813 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage20-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage20-crc\") on node \"crc\" " Dec 01 09:05:14 crc kubenswrapper[4813]: I1201 09:05:14.900244 4813 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/522f36f6-9da6-46b0-8ca6-9753311b3dfc-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 01 09:05:14 crc kubenswrapper[4813]: I1201 09:05:14.900256 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jtfbh\" (UniqueName: \"kubernetes.io/projected/d6f4323d-561b-4725-a1fe-140db5f6a33d-kube-api-access-jtfbh\") on node \"crc\" DevicePath \"\"" Dec 01 09:05:14 crc kubenswrapper[4813]: I1201 09:05:14.900278 4813 reconciler_common.go:293] "Volume detached for volume \"run\" (UniqueName: \"kubernetes.io/host-path/522f36f6-9da6-46b0-8ca6-9753311b3dfc-run\") on node \"crc\" DevicePath \"\"" Dec 01 09:05:14 crc kubenswrapper[4813]: I1201 09:05:14.900371 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/522f36f6-9da6-46b0-8ca6-9753311b3dfc-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "522f36f6-9da6-46b0-8ca6-9753311b3dfc" (UID: "522f36f6-9da6-46b0-8ca6-9753311b3dfc"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 09:05:14 crc kubenswrapper[4813]: I1201 09:05:14.901717 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage07-crc" (OuterVolumeSpecName: "glance") pod "522f36f6-9da6-46b0-8ca6-9753311b3dfc" (UID: "522f36f6-9da6-46b0-8ca6-9753311b3dfc"). InnerVolumeSpecName "local-storage07-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 01 09:05:14 crc kubenswrapper[4813]: I1201 09:05:14.903327 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage03-crc" (OuterVolumeSpecName: "glance-cache") pod "522f36f6-9da6-46b0-8ca6-9753311b3dfc" (UID: "522f36f6-9da6-46b0-8ca6-9753311b3dfc"). InnerVolumeSpecName "local-storage03-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 01 09:05:14 crc kubenswrapper[4813]: I1201 09:05:14.904023 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/522f36f6-9da6-46b0-8ca6-9753311b3dfc-kube-api-access-v7h7b" (OuterVolumeSpecName: "kube-api-access-v7h7b") pod "522f36f6-9da6-46b0-8ca6-9753311b3dfc" (UID: "522f36f6-9da6-46b0-8ca6-9753311b3dfc"). InnerVolumeSpecName "kube-api-access-v7h7b". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:05:14 crc kubenswrapper[4813]: I1201 09:05:14.905910 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/522f36f6-9da6-46b0-8ca6-9753311b3dfc-scripts" (OuterVolumeSpecName: "scripts") pod "522f36f6-9da6-46b0-8ca6-9753311b3dfc" (UID: "522f36f6-9da6-46b0-8ca6-9753311b3dfc"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:05:14 crc kubenswrapper[4813]: I1201 09:05:14.917543 4813 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage16-crc" (UniqueName: "kubernetes.io/local-volume/local-storage16-crc") on node "crc" Dec 01 09:05:14 crc kubenswrapper[4813]: I1201 09:05:14.917783 4813 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage20-crc" (UniqueName: "kubernetes.io/local-volume/local-storage20-crc") on node "crc" Dec 01 09:05:14 crc kubenswrapper[4813]: I1201 09:05:14.947242 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/522f36f6-9da6-46b0-8ca6-9753311b3dfc-config-data" (OuterVolumeSpecName: "config-data") pod "522f36f6-9da6-46b0-8ca6-9753311b3dfc" (UID: "522f36f6-9da6-46b0-8ca6-9753311b3dfc"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:05:15 crc kubenswrapper[4813]: I1201 09:05:15.001698 4813 reconciler_common.go:293] "Volume detached for volume \"local-storage16-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage16-crc\") on node \"crc\" DevicePath \"\"" Dec 01 09:05:15 crc kubenswrapper[4813]: I1201 09:05:15.001733 4813 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/522f36f6-9da6-46b0-8ca6-9753311b3dfc-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 09:05:15 crc kubenswrapper[4813]: I1201 09:05:15.001744 4813 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/522f36f6-9da6-46b0-8ca6-9753311b3dfc-lib-modules\") on node \"crc\" DevicePath \"\"" Dec 01 09:05:15 crc kubenswrapper[4813]: I1201 09:05:15.001753 4813 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/522f36f6-9da6-46b0-8ca6-9753311b3dfc-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 09:05:15 crc kubenswrapper[4813]: I1201 09:05:15.001787 4813 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" " Dec 01 09:05:15 crc kubenswrapper[4813]: I1201 09:05:15.001798 4813 reconciler_common.go:293] "Volume detached for volume \"local-storage20-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage20-crc\") on node \"crc\" DevicePath \"\"" Dec 01 09:05:15 crc kubenswrapper[4813]: I1201 09:05:15.001806 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v7h7b\" (UniqueName: \"kubernetes.io/projected/522f36f6-9da6-46b0-8ca6-9753311b3dfc-kube-api-access-v7h7b\") on node \"crc\" DevicePath \"\"" Dec 01 09:05:15 crc kubenswrapper[4813]: I1201 09:05:15.001820 4813 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" " Dec 01 09:05:15 crc kubenswrapper[4813]: I1201 09:05:15.014524 4813 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage03-crc" (UniqueName: "kubernetes.io/local-volume/local-storage03-crc") on node "crc" Dec 01 09:05:15 crc kubenswrapper[4813]: I1201 09:05:15.031191 4813 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage07-crc" (UniqueName: "kubernetes.io/local-volume/local-storage07-crc") on node "crc" Dec 01 09:05:15 crc kubenswrapper[4813]: I1201 09:05:15.103698 4813 reconciler_common.go:293] "Volume detached for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" DevicePath \"\"" Dec 01 09:05:15 crc kubenswrapper[4813]: I1201 09:05:15.103759 4813 reconciler_common.go:293] "Volume detached for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" DevicePath \"\"" Dec 01 09:05:15 crc kubenswrapper[4813]: I1201 09:05:15.160136 4813 generic.go:334] "Generic (PLEG): container finished" podID="d6f4323d-561b-4725-a1fe-140db5f6a33d" containerID="9ed932164815080b18b0a321c5504fcf333b242a507ed815a0fcb1e071c97b5c" exitCode=0 Dec 01 09:05:15 crc kubenswrapper[4813]: I1201 09:05:15.160220 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-0" Dec 01 09:05:15 crc kubenswrapper[4813]: I1201 09:05:15.160297 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-0" event={"ID":"d6f4323d-561b-4725-a1fe-140db5f6a33d","Type":"ContainerDied","Data":"9ed932164815080b18b0a321c5504fcf333b242a507ed815a0fcb1e071c97b5c"} Dec 01 09:05:15 crc kubenswrapper[4813]: I1201 09:05:15.160398 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-0" event={"ID":"d6f4323d-561b-4725-a1fe-140db5f6a33d","Type":"ContainerDied","Data":"10b31d489e5914c49ad6b4a668c2fa1494515f79057c148aeb333ecc1b27a574"} Dec 01 09:05:15 crc kubenswrapper[4813]: I1201 09:05:15.160442 4813 scope.go:117] "RemoveContainer" containerID="9ed932164815080b18b0a321c5504fcf333b242a507ed815a0fcb1e071c97b5c" Dec 01 09:05:15 crc kubenswrapper[4813]: I1201 09:05:15.167854 4813 generic.go:334] "Generic (PLEG): container finished" podID="522f36f6-9da6-46b0-8ca6-9753311b3dfc" containerID="269455c89215c29899616183ae9d07b026d8fc34efad53048b988fc15a0f5889" exitCode=0 Dec 01 09:05:15 crc kubenswrapper[4813]: I1201 09:05:15.167987 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-0" Dec 01 09:05:15 crc kubenswrapper[4813]: I1201 09:05:15.168023 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"522f36f6-9da6-46b0-8ca6-9753311b3dfc","Type":"ContainerDied","Data":"269455c89215c29899616183ae9d07b026d8fc34efad53048b988fc15a0f5889"} Dec 01 09:05:15 crc kubenswrapper[4813]: I1201 09:05:15.169009 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"522f36f6-9da6-46b0-8ca6-9753311b3dfc","Type":"ContainerDied","Data":"011eaf2acf8bb7bc009d418aad121279c5ff5d494e6e7b6e142025263a66edeb"} Dec 01 09:05:15 crc kubenswrapper[4813]: I1201 09:05:15.204372 4813 scope.go:117] "RemoveContainer" containerID="83cba63dc3ec859be23efccb9eb8d4d3a8dbe205a4e16dbcad3a7a0ffd16a54e" Dec 01 09:05:15 crc kubenswrapper[4813]: I1201 09:05:15.260468 4813 scope.go:117] "RemoveContainer" containerID="9ed932164815080b18b0a321c5504fcf333b242a507ed815a0fcb1e071c97b5c" Dec 01 09:05:15 crc kubenswrapper[4813]: E1201 09:05:15.264180 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9ed932164815080b18b0a321c5504fcf333b242a507ed815a0fcb1e071c97b5c\": container with ID starting with 9ed932164815080b18b0a321c5504fcf333b242a507ed815a0fcb1e071c97b5c not found: ID does not exist" containerID="9ed932164815080b18b0a321c5504fcf333b242a507ed815a0fcb1e071c97b5c" Dec 01 09:05:15 crc kubenswrapper[4813]: I1201 09:05:15.264393 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9ed932164815080b18b0a321c5504fcf333b242a507ed815a0fcb1e071c97b5c"} err="failed to get container status \"9ed932164815080b18b0a321c5504fcf333b242a507ed815a0fcb1e071c97b5c\": rpc error: code = NotFound desc = could not find container \"9ed932164815080b18b0a321c5504fcf333b242a507ed815a0fcb1e071c97b5c\": container with ID starting with 9ed932164815080b18b0a321c5504fcf333b242a507ed815a0fcb1e071c97b5c not found: ID does not exist" Dec 01 09:05:15 crc kubenswrapper[4813]: I1201 09:05:15.264559 4813 scope.go:117] "RemoveContainer" containerID="83cba63dc3ec859be23efccb9eb8d4d3a8dbe205a4e16dbcad3a7a0ffd16a54e" Dec 01 09:05:15 crc kubenswrapper[4813]: E1201 09:05:15.265536 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"83cba63dc3ec859be23efccb9eb8d4d3a8dbe205a4e16dbcad3a7a0ffd16a54e\": container with ID starting with 83cba63dc3ec859be23efccb9eb8d4d3a8dbe205a4e16dbcad3a7a0ffd16a54e not found: ID does not exist" containerID="83cba63dc3ec859be23efccb9eb8d4d3a8dbe205a4e16dbcad3a7a0ffd16a54e" Dec 01 09:05:15 crc kubenswrapper[4813]: I1201 09:05:15.265695 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"83cba63dc3ec859be23efccb9eb8d4d3a8dbe205a4e16dbcad3a7a0ffd16a54e"} err="failed to get container status \"83cba63dc3ec859be23efccb9eb8d4d3a8dbe205a4e16dbcad3a7a0ffd16a54e\": rpc error: code = NotFound desc = could not find container \"83cba63dc3ec859be23efccb9eb8d4d3a8dbe205a4e16dbcad3a7a0ffd16a54e\": container with ID starting with 83cba63dc3ec859be23efccb9eb8d4d3a8dbe205a4e16dbcad3a7a0ffd16a54e not found: ID does not exist" Dec 01 09:05:15 crc kubenswrapper[4813]: I1201 09:05:15.265731 4813 scope.go:117] "RemoveContainer" containerID="269455c89215c29899616183ae9d07b026d8fc34efad53048b988fc15a0f5889" Dec 01 09:05:15 crc kubenswrapper[4813]: I1201 09:05:15.270378 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-0"] Dec 01 09:05:15 crc kubenswrapper[4813]: I1201 09:05:15.281034 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-0"] Dec 01 09:05:15 crc kubenswrapper[4813]: I1201 09:05:15.290544 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Dec 01 09:05:15 crc kubenswrapper[4813]: I1201 09:05:15.298745 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Dec 01 09:05:15 crc kubenswrapper[4813]: I1201 09:05:15.341528 4813 scope.go:117] "RemoveContainer" containerID="954887d4b9b9439ab9fdfc4078054b7d9d86884474571c85ee8708f64ca6198a" Dec 01 09:05:15 crc kubenswrapper[4813]: I1201 09:05:15.362327 4813 scope.go:117] "RemoveContainer" containerID="269455c89215c29899616183ae9d07b026d8fc34efad53048b988fc15a0f5889" Dec 01 09:05:15 crc kubenswrapper[4813]: E1201 09:05:15.362925 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"269455c89215c29899616183ae9d07b026d8fc34efad53048b988fc15a0f5889\": container with ID starting with 269455c89215c29899616183ae9d07b026d8fc34efad53048b988fc15a0f5889 not found: ID does not exist" containerID="269455c89215c29899616183ae9d07b026d8fc34efad53048b988fc15a0f5889" Dec 01 09:05:15 crc kubenswrapper[4813]: I1201 09:05:15.362997 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"269455c89215c29899616183ae9d07b026d8fc34efad53048b988fc15a0f5889"} err="failed to get container status \"269455c89215c29899616183ae9d07b026d8fc34efad53048b988fc15a0f5889\": rpc error: code = NotFound desc = could not find container \"269455c89215c29899616183ae9d07b026d8fc34efad53048b988fc15a0f5889\": container with ID starting with 269455c89215c29899616183ae9d07b026d8fc34efad53048b988fc15a0f5889 not found: ID does not exist" Dec 01 09:05:15 crc kubenswrapper[4813]: I1201 09:05:15.363031 4813 scope.go:117] "RemoveContainer" containerID="954887d4b9b9439ab9fdfc4078054b7d9d86884474571c85ee8708f64ca6198a" Dec 01 09:05:15 crc kubenswrapper[4813]: E1201 09:05:15.363537 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"954887d4b9b9439ab9fdfc4078054b7d9d86884474571c85ee8708f64ca6198a\": container with ID starting with 954887d4b9b9439ab9fdfc4078054b7d9d86884474571c85ee8708f64ca6198a not found: ID does not exist" containerID="954887d4b9b9439ab9fdfc4078054b7d9d86884474571c85ee8708f64ca6198a" Dec 01 09:05:15 crc kubenswrapper[4813]: I1201 09:05:15.363580 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"954887d4b9b9439ab9fdfc4078054b7d9d86884474571c85ee8708f64ca6198a"} err="failed to get container status \"954887d4b9b9439ab9fdfc4078054b7d9d86884474571c85ee8708f64ca6198a\": rpc error: code = NotFound desc = could not find container \"954887d4b9b9439ab9fdfc4078054b7d9d86884474571c85ee8708f64ca6198a\": container with ID starting with 954887d4b9b9439ab9fdfc4078054b7d9d86884474571c85ee8708f64ca6198a not found: ID does not exist" Dec 01 09:05:16 crc kubenswrapper[4813]: I1201 09:05:16.064840 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-db-create-fcjt9"] Dec 01 09:05:16 crc kubenswrapper[4813]: I1201 09:05:16.078209 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-db-create-fcjt9"] Dec 01 09:05:16 crc kubenswrapper[4813]: I1201 09:05:16.087895 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glancea9c4-account-delete-wl47f"] Dec 01 09:05:16 crc kubenswrapper[4813]: I1201 09:05:16.094714 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-a9c4-account-create-update-4g8nj"] Dec 01 09:05:16 crc kubenswrapper[4813]: I1201 09:05:16.104921 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-a9c4-account-create-update-4g8nj"] Dec 01 09:05:16 crc kubenswrapper[4813]: I1201 09:05:16.112679 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glancea9c4-account-delete-wl47f"] Dec 01 09:05:16 crc kubenswrapper[4813]: I1201 09:05:16.405619 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3de75509-b8b6-4b97-b08e-77057f86ea77" path="/var/lib/kubelet/pods/3de75509-b8b6-4b97-b08e-77057f86ea77/volumes" Dec 01 09:05:16 crc kubenswrapper[4813]: I1201 09:05:16.406785 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="522f36f6-9da6-46b0-8ca6-9753311b3dfc" path="/var/lib/kubelet/pods/522f36f6-9da6-46b0-8ca6-9753311b3dfc/volumes" Dec 01 09:05:16 crc kubenswrapper[4813]: I1201 09:05:16.407349 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="54c06e84-7478-4fbc-9e2c-bdc6e1be658b" path="/var/lib/kubelet/pods/54c06e84-7478-4fbc-9e2c-bdc6e1be658b/volumes" Dec 01 09:05:16 crc kubenswrapper[4813]: I1201 09:05:16.408257 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="77e7d212-07ac-4c29-acb2-4e83bcf6d7e0" path="/var/lib/kubelet/pods/77e7d212-07ac-4c29-acb2-4e83bcf6d7e0/volumes" Dec 01 09:05:16 crc kubenswrapper[4813]: I1201 09:05:16.408855 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d6f4323d-561b-4725-a1fe-140db5f6a33d" path="/var/lib/kubelet/pods/d6f4323d-561b-4725-a1fe-140db5f6a33d/volumes" Dec 01 09:05:17 crc kubenswrapper[4813]: I1201 09:05:17.221654 4813 patch_prober.go:28] interesting pod/machine-config-daemon-9wn4q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 09:05:17 crc kubenswrapper[4813]: I1201 09:05:17.222034 4813 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9wn4q" podUID="3d423402-e08f-42e9-b60c-6eb9bf2f6d48" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 09:05:17 crc kubenswrapper[4813]: I1201 09:05:17.222104 4813 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-9wn4q" Dec 01 09:05:17 crc kubenswrapper[4813]: I1201 09:05:17.222787 4813 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"09b49517239f156b4ac50c72e4d184fde70e2daaf7b602adcfc3b42ec90a71e0"} pod="openshift-machine-config-operator/machine-config-daemon-9wn4q" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 01 09:05:17 crc kubenswrapper[4813]: I1201 09:05:17.222866 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-9wn4q" podUID="3d423402-e08f-42e9-b60c-6eb9bf2f6d48" containerName="machine-config-daemon" containerID="cri-o://09b49517239f156b4ac50c72e4d184fde70e2daaf7b602adcfc3b42ec90a71e0" gracePeriod=600 Dec 01 09:05:17 crc kubenswrapper[4813]: E1201 09:05:17.343992 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9wn4q_openshift-machine-config-operator(3d423402-e08f-42e9-b60c-6eb9bf2f6d48)\"" pod="openshift-machine-config-operator/machine-config-daemon-9wn4q" podUID="3d423402-e08f-42e9-b60c-6eb9bf2f6d48" Dec 01 09:05:17 crc kubenswrapper[4813]: I1201 09:05:17.539154 4813 scope.go:117] "RemoveContainer" containerID="20f85026200b9a4f47c934240e34d5be7ca58f51c42699d023d34b52854312d9" Dec 01 09:05:17 crc kubenswrapper[4813]: I1201 09:05:17.563120 4813 scope.go:117] "RemoveContainer" containerID="47f3a60a2c5f83d711d86643c6edac1f3b9962268dc6d7aed6aa40747477311b" Dec 01 09:05:17 crc kubenswrapper[4813]: I1201 09:05:17.633281 4813 scope.go:117] "RemoveContainer" containerID="68ba99704890378fa2659521be07b13eef554ebf618621fd31d7efe2d3fe1049" Dec 01 09:05:18 crc kubenswrapper[4813]: I1201 09:05:18.196509 4813 generic.go:334] "Generic (PLEG): container finished" podID="3d423402-e08f-42e9-b60c-6eb9bf2f6d48" containerID="09b49517239f156b4ac50c72e4d184fde70e2daaf7b602adcfc3b42ec90a71e0" exitCode=0 Dec 01 09:05:18 crc kubenswrapper[4813]: I1201 09:05:18.196567 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-9wn4q" event={"ID":"3d423402-e08f-42e9-b60c-6eb9bf2f6d48","Type":"ContainerDied","Data":"09b49517239f156b4ac50c72e4d184fde70e2daaf7b602adcfc3b42ec90a71e0"} Dec 01 09:05:18 crc kubenswrapper[4813]: I1201 09:05:18.196607 4813 scope.go:117] "RemoveContainer" containerID="e5115101af56b8d932bd5a35078afaf20c64fdf9b230d78f43cbc4c834a68bf5" Dec 01 09:05:18 crc kubenswrapper[4813]: I1201 09:05:18.197543 4813 scope.go:117] "RemoveContainer" containerID="09b49517239f156b4ac50c72e4d184fde70e2daaf7b602adcfc3b42ec90a71e0" Dec 01 09:05:18 crc kubenswrapper[4813]: E1201 09:05:18.197859 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9wn4q_openshift-machine-config-operator(3d423402-e08f-42e9-b60c-6eb9bf2f6d48)\"" pod="openshift-machine-config-operator/machine-config-daemon-9wn4q" podUID="3d423402-e08f-42e9-b60c-6eb9bf2f6d48" Dec 01 09:05:24 crc kubenswrapper[4813]: I1201 09:05:24.855465 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/swift-ring-rebalance-6bsvb"] Dec 01 09:05:24 crc kubenswrapper[4813]: I1201 09:05:24.862191 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/swift-storage-0"] Dec 01 09:05:24 crc kubenswrapper[4813]: I1201 09:05:24.862874 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/swift-storage-0" podUID="995100a7-4c5a-409a-81f2-9dbaf2fa1d47" containerName="account-server" containerID="cri-o://c9f5cb19a620cfc7b9df0bad296f8f90075cf217fda34e7c9818acd5277ecc4b" gracePeriod=30 Dec 01 09:05:24 crc kubenswrapper[4813]: I1201 09:05:24.862933 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/swift-storage-0" podUID="995100a7-4c5a-409a-81f2-9dbaf2fa1d47" containerName="object-server" containerID="cri-o://90152b553c2c0696cb360f3d2b45d4e4f1d069d88a813084717ae44e3d7f007e" gracePeriod=30 Dec 01 09:05:24 crc kubenswrapper[4813]: I1201 09:05:24.863065 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/swift-storage-0" podUID="995100a7-4c5a-409a-81f2-9dbaf2fa1d47" containerName="container-updater" containerID="cri-o://71ea72b9275533c0074b122f73f7beebac12af351e19c3c9c0e6c0b137e24f6f" gracePeriod=30 Dec 01 09:05:24 crc kubenswrapper[4813]: I1201 09:05:24.863127 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/swift-storage-0" podUID="995100a7-4c5a-409a-81f2-9dbaf2fa1d47" containerName="container-auditor" containerID="cri-o://6e3df495999a12b94a6ca78f0149ccd17c13cf04f3fc31edaef7a77490b259d1" gracePeriod=30 Dec 01 09:05:24 crc kubenswrapper[4813]: I1201 09:05:24.863218 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/swift-storage-0" podUID="995100a7-4c5a-409a-81f2-9dbaf2fa1d47" containerName="container-replicator" containerID="cri-o://d2d428af1b1d67ed1e1389a7fb3e3e1ea61d4ffd00a4526c2b45c9655f43d7ca" gracePeriod=30 Dec 01 09:05:24 crc kubenswrapper[4813]: I1201 09:05:24.863260 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/swift-storage-0" podUID="995100a7-4c5a-409a-81f2-9dbaf2fa1d47" containerName="container-server" containerID="cri-o://4a635b3fb091ce09c36d9181c2aa094b2a84e366b1bb19e73b4059e4af3205fa" gracePeriod=30 Dec 01 09:05:24 crc kubenswrapper[4813]: I1201 09:05:24.863304 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/swift-storage-0" podUID="995100a7-4c5a-409a-81f2-9dbaf2fa1d47" containerName="account-reaper" containerID="cri-o://2b9455a05903cfc3c8f0fe13912f2aeb0af2368d2f612165ece50ff474a9e935" gracePeriod=30 Dec 01 09:05:24 crc kubenswrapper[4813]: I1201 09:05:24.863348 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/swift-storage-0" podUID="995100a7-4c5a-409a-81f2-9dbaf2fa1d47" containerName="account-auditor" containerID="cri-o://0b01d236c0fd4b7e6153fcbb9c688913b9adc82fea28ce6cfa31368fb7072a52" gracePeriod=30 Dec 01 09:05:24 crc kubenswrapper[4813]: I1201 09:05:24.863389 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/swift-storage-0" podUID="995100a7-4c5a-409a-81f2-9dbaf2fa1d47" containerName="account-replicator" containerID="cri-o://532f0c05bdb2cee021285ec588aac0effe66a65c6f37749e25369d978d0515e4" gracePeriod=30 Dec 01 09:05:24 crc kubenswrapper[4813]: I1201 09:05:24.863404 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/swift-storage-0" podUID="995100a7-4c5a-409a-81f2-9dbaf2fa1d47" containerName="swift-recon-cron" containerID="cri-o://08b235b36c9b32c3a32345cc647c1bd488a16d47403f387ba92c8337ae7120e1" gracePeriod=30 Dec 01 09:05:24 crc kubenswrapper[4813]: I1201 09:05:24.863472 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/swift-storage-0" podUID="995100a7-4c5a-409a-81f2-9dbaf2fa1d47" containerName="rsync" containerID="cri-o://8a6edcb51ab7c3bb702c42f67b2432c8ace1cb2fa02e6a3db3fe9519ecf7f7f3" gracePeriod=30 Dec 01 09:05:24 crc kubenswrapper[4813]: I1201 09:05:24.863528 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/swift-storage-0" podUID="995100a7-4c5a-409a-81f2-9dbaf2fa1d47" containerName="object-expirer" containerID="cri-o://b5f856516db2a807cb93878232098a03c9c22a32b0221eba5c835097c972aca5" gracePeriod=30 Dec 01 09:05:24 crc kubenswrapper[4813]: I1201 09:05:24.863585 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/swift-storage-0" podUID="995100a7-4c5a-409a-81f2-9dbaf2fa1d47" containerName="object-updater" containerID="cri-o://cfcbba573ff80c783a56d8b194b0bfee5c5bf61df702623a4752329024c48781" gracePeriod=30 Dec 01 09:05:24 crc kubenswrapper[4813]: I1201 09:05:24.863641 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/swift-storage-0" podUID="995100a7-4c5a-409a-81f2-9dbaf2fa1d47" containerName="object-auditor" containerID="cri-o://72d0f4ae012d4d6cd410f0e6e9f0041810afd974e12e68981d6f390c1ce1cb8a" gracePeriod=30 Dec 01 09:05:24 crc kubenswrapper[4813]: I1201 09:05:24.863706 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/swift-storage-0" podUID="995100a7-4c5a-409a-81f2-9dbaf2fa1d47" containerName="object-replicator" containerID="cri-o://4cd563e572aa4851bd7a2a6fdce31ea0962d58c89ab9c06f628d341281b7a914" gracePeriod=30 Dec 01 09:05:24 crc kubenswrapper[4813]: I1201 09:05:24.872265 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/swift-ring-rebalance-6bsvb"] Dec 01 09:05:24 crc kubenswrapper[4813]: I1201 09:05:24.922659 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/swift-proxy-6bd58cfcf7-vhtgp"] Dec 01 09:05:24 crc kubenswrapper[4813]: I1201 09:05:24.923686 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-vhtgp" podUID="5504ea2a-5701-4d37-8f4f-613caab65413" containerName="proxy-httpd" containerID="cri-o://d57b6a9a7684f7aceb73a0c233d97f348bb0a09a608815152313412c1a8f0cdc" gracePeriod=30 Dec 01 09:05:24 crc kubenswrapper[4813]: I1201 09:05:24.923754 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-vhtgp" podUID="5504ea2a-5701-4d37-8f4f-613caab65413" containerName="proxy-server" containerID="cri-o://af790186878c3a0dfbe4b88e5964cbbe3ef87f73830bc59c1d00f90328cc6e30" gracePeriod=30 Dec 01 09:05:25 crc kubenswrapper[4813]: I1201 09:05:25.263524 4813 generic.go:334] "Generic (PLEG): container finished" podID="5504ea2a-5701-4d37-8f4f-613caab65413" containerID="d57b6a9a7684f7aceb73a0c233d97f348bb0a09a608815152313412c1a8f0cdc" exitCode=0 Dec 01 09:05:25 crc kubenswrapper[4813]: I1201 09:05:25.263627 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-vhtgp" event={"ID":"5504ea2a-5701-4d37-8f4f-613caab65413","Type":"ContainerDied","Data":"d57b6a9a7684f7aceb73a0c233d97f348bb0a09a608815152313412c1a8f0cdc"} Dec 01 09:05:25 crc kubenswrapper[4813]: I1201 09:05:25.270442 4813 generic.go:334] "Generic (PLEG): container finished" podID="995100a7-4c5a-409a-81f2-9dbaf2fa1d47" containerID="b5f856516db2a807cb93878232098a03c9c22a32b0221eba5c835097c972aca5" exitCode=0 Dec 01 09:05:25 crc kubenswrapper[4813]: I1201 09:05:25.270493 4813 generic.go:334] "Generic (PLEG): container finished" podID="995100a7-4c5a-409a-81f2-9dbaf2fa1d47" containerID="cfcbba573ff80c783a56d8b194b0bfee5c5bf61df702623a4752329024c48781" exitCode=0 Dec 01 09:05:25 crc kubenswrapper[4813]: I1201 09:05:25.270507 4813 generic.go:334] "Generic (PLEG): container finished" podID="995100a7-4c5a-409a-81f2-9dbaf2fa1d47" containerID="72d0f4ae012d4d6cd410f0e6e9f0041810afd974e12e68981d6f390c1ce1cb8a" exitCode=0 Dec 01 09:05:25 crc kubenswrapper[4813]: I1201 09:05:25.270525 4813 generic.go:334] "Generic (PLEG): container finished" podID="995100a7-4c5a-409a-81f2-9dbaf2fa1d47" containerID="4cd563e572aa4851bd7a2a6fdce31ea0962d58c89ab9c06f628d341281b7a914" exitCode=0 Dec 01 09:05:25 crc kubenswrapper[4813]: I1201 09:05:25.270537 4813 generic.go:334] "Generic (PLEG): container finished" podID="995100a7-4c5a-409a-81f2-9dbaf2fa1d47" containerID="71ea72b9275533c0074b122f73f7beebac12af351e19c3c9c0e6c0b137e24f6f" exitCode=0 Dec 01 09:05:25 crc kubenswrapper[4813]: I1201 09:05:25.270548 4813 generic.go:334] "Generic (PLEG): container finished" podID="995100a7-4c5a-409a-81f2-9dbaf2fa1d47" containerID="6e3df495999a12b94a6ca78f0149ccd17c13cf04f3fc31edaef7a77490b259d1" exitCode=0 Dec 01 09:05:25 crc kubenswrapper[4813]: I1201 09:05:25.270559 4813 generic.go:334] "Generic (PLEG): container finished" podID="995100a7-4c5a-409a-81f2-9dbaf2fa1d47" containerID="d2d428af1b1d67ed1e1389a7fb3e3e1ea61d4ffd00a4526c2b45c9655f43d7ca" exitCode=0 Dec 01 09:05:25 crc kubenswrapper[4813]: I1201 09:05:25.270569 4813 generic.go:334] "Generic (PLEG): container finished" podID="995100a7-4c5a-409a-81f2-9dbaf2fa1d47" containerID="2b9455a05903cfc3c8f0fe13912f2aeb0af2368d2f612165ece50ff474a9e935" exitCode=0 Dec 01 09:05:25 crc kubenswrapper[4813]: I1201 09:05:25.270581 4813 generic.go:334] "Generic (PLEG): container finished" podID="995100a7-4c5a-409a-81f2-9dbaf2fa1d47" containerID="0b01d236c0fd4b7e6153fcbb9c688913b9adc82fea28ce6cfa31368fb7072a52" exitCode=0 Dec 01 09:05:25 crc kubenswrapper[4813]: I1201 09:05:25.270589 4813 generic.go:334] "Generic (PLEG): container finished" podID="995100a7-4c5a-409a-81f2-9dbaf2fa1d47" containerID="532f0c05bdb2cee021285ec588aac0effe66a65c6f37749e25369d978d0515e4" exitCode=0 Dec 01 09:05:25 crc kubenswrapper[4813]: I1201 09:05:25.270598 4813 generic.go:334] "Generic (PLEG): container finished" podID="995100a7-4c5a-409a-81f2-9dbaf2fa1d47" containerID="c9f5cb19a620cfc7b9df0bad296f8f90075cf217fda34e7c9818acd5277ecc4b" exitCode=0 Dec 01 09:05:25 crc kubenswrapper[4813]: I1201 09:05:25.270626 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"995100a7-4c5a-409a-81f2-9dbaf2fa1d47","Type":"ContainerDied","Data":"b5f856516db2a807cb93878232098a03c9c22a32b0221eba5c835097c972aca5"} Dec 01 09:05:25 crc kubenswrapper[4813]: I1201 09:05:25.270661 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"995100a7-4c5a-409a-81f2-9dbaf2fa1d47","Type":"ContainerDied","Data":"cfcbba573ff80c783a56d8b194b0bfee5c5bf61df702623a4752329024c48781"} Dec 01 09:05:25 crc kubenswrapper[4813]: I1201 09:05:25.270676 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"995100a7-4c5a-409a-81f2-9dbaf2fa1d47","Type":"ContainerDied","Data":"72d0f4ae012d4d6cd410f0e6e9f0041810afd974e12e68981d6f390c1ce1cb8a"} Dec 01 09:05:25 crc kubenswrapper[4813]: I1201 09:05:25.270688 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"995100a7-4c5a-409a-81f2-9dbaf2fa1d47","Type":"ContainerDied","Data":"4cd563e572aa4851bd7a2a6fdce31ea0962d58c89ab9c06f628d341281b7a914"} Dec 01 09:05:25 crc kubenswrapper[4813]: I1201 09:05:25.270700 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"995100a7-4c5a-409a-81f2-9dbaf2fa1d47","Type":"ContainerDied","Data":"71ea72b9275533c0074b122f73f7beebac12af351e19c3c9c0e6c0b137e24f6f"} Dec 01 09:05:25 crc kubenswrapper[4813]: I1201 09:05:25.270714 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"995100a7-4c5a-409a-81f2-9dbaf2fa1d47","Type":"ContainerDied","Data":"6e3df495999a12b94a6ca78f0149ccd17c13cf04f3fc31edaef7a77490b259d1"} Dec 01 09:05:25 crc kubenswrapper[4813]: I1201 09:05:25.270726 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"995100a7-4c5a-409a-81f2-9dbaf2fa1d47","Type":"ContainerDied","Data":"d2d428af1b1d67ed1e1389a7fb3e3e1ea61d4ffd00a4526c2b45c9655f43d7ca"} Dec 01 09:05:25 crc kubenswrapper[4813]: I1201 09:05:25.270738 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"995100a7-4c5a-409a-81f2-9dbaf2fa1d47","Type":"ContainerDied","Data":"2b9455a05903cfc3c8f0fe13912f2aeb0af2368d2f612165ece50ff474a9e935"} Dec 01 09:05:25 crc kubenswrapper[4813]: I1201 09:05:25.270750 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"995100a7-4c5a-409a-81f2-9dbaf2fa1d47","Type":"ContainerDied","Data":"0b01d236c0fd4b7e6153fcbb9c688913b9adc82fea28ce6cfa31368fb7072a52"} Dec 01 09:05:25 crc kubenswrapper[4813]: I1201 09:05:25.270762 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"995100a7-4c5a-409a-81f2-9dbaf2fa1d47","Type":"ContainerDied","Data":"532f0c05bdb2cee021285ec588aac0effe66a65c6f37749e25369d978d0515e4"} Dec 01 09:05:25 crc kubenswrapper[4813]: I1201 09:05:25.270774 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"995100a7-4c5a-409a-81f2-9dbaf2fa1d47","Type":"ContainerDied","Data":"c9f5cb19a620cfc7b9df0bad296f8f90075cf217fda34e7c9818acd5277ecc4b"} Dec 01 09:05:25 crc kubenswrapper[4813]: I1201 09:05:25.805485 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-vhtgp" Dec 01 09:05:25 crc kubenswrapper[4813]: I1201 09:05:25.970612 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/5504ea2a-5701-4d37-8f4f-613caab65413-etc-swift\") pod \"5504ea2a-5701-4d37-8f4f-613caab65413\" (UID: \"5504ea2a-5701-4d37-8f4f-613caab65413\") " Dec 01 09:05:25 crc kubenswrapper[4813]: I1201 09:05:25.970731 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5504ea2a-5701-4d37-8f4f-613caab65413-log-httpd\") pod \"5504ea2a-5701-4d37-8f4f-613caab65413\" (UID: \"5504ea2a-5701-4d37-8f4f-613caab65413\") " Dec 01 09:05:25 crc kubenswrapper[4813]: I1201 09:05:25.970783 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5504ea2a-5701-4d37-8f4f-613caab65413-config-data\") pod \"5504ea2a-5701-4d37-8f4f-613caab65413\" (UID: \"5504ea2a-5701-4d37-8f4f-613caab65413\") " Dec 01 09:05:25 crc kubenswrapper[4813]: I1201 09:05:25.970817 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gbncc\" (UniqueName: \"kubernetes.io/projected/5504ea2a-5701-4d37-8f4f-613caab65413-kube-api-access-gbncc\") pod \"5504ea2a-5701-4d37-8f4f-613caab65413\" (UID: \"5504ea2a-5701-4d37-8f4f-613caab65413\") " Dec 01 09:05:25 crc kubenswrapper[4813]: I1201 09:05:25.970880 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5504ea2a-5701-4d37-8f4f-613caab65413-run-httpd\") pod \"5504ea2a-5701-4d37-8f4f-613caab65413\" (UID: \"5504ea2a-5701-4d37-8f4f-613caab65413\") " Dec 01 09:05:25 crc kubenswrapper[4813]: I1201 09:05:25.971281 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5504ea2a-5701-4d37-8f4f-613caab65413-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "5504ea2a-5701-4d37-8f4f-613caab65413" (UID: "5504ea2a-5701-4d37-8f4f-613caab65413"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 09:05:25 crc kubenswrapper[4813]: I1201 09:05:25.971655 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5504ea2a-5701-4d37-8f4f-613caab65413-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "5504ea2a-5701-4d37-8f4f-613caab65413" (UID: "5504ea2a-5701-4d37-8f4f-613caab65413"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 09:05:25 crc kubenswrapper[4813]: I1201 09:05:25.976428 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5504ea2a-5701-4d37-8f4f-613caab65413-kube-api-access-gbncc" (OuterVolumeSpecName: "kube-api-access-gbncc") pod "5504ea2a-5701-4d37-8f4f-613caab65413" (UID: "5504ea2a-5701-4d37-8f4f-613caab65413"). InnerVolumeSpecName "kube-api-access-gbncc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:05:25 crc kubenswrapper[4813]: I1201 09:05:25.978532 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5504ea2a-5701-4d37-8f4f-613caab65413-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "5504ea2a-5701-4d37-8f4f-613caab65413" (UID: "5504ea2a-5701-4d37-8f4f-613caab65413"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:05:26 crc kubenswrapper[4813]: I1201 09:05:26.008015 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5504ea2a-5701-4d37-8f4f-613caab65413-config-data" (OuterVolumeSpecName: "config-data") pod "5504ea2a-5701-4d37-8f4f-613caab65413" (UID: "5504ea2a-5701-4d37-8f4f-613caab65413"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:05:26 crc kubenswrapper[4813]: I1201 09:05:26.073135 4813 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/5504ea2a-5701-4d37-8f4f-613caab65413-etc-swift\") on node \"crc\" DevicePath \"\"" Dec 01 09:05:26 crc kubenswrapper[4813]: I1201 09:05:26.073180 4813 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5504ea2a-5701-4d37-8f4f-613caab65413-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 01 09:05:26 crc kubenswrapper[4813]: I1201 09:05:26.073192 4813 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5504ea2a-5701-4d37-8f4f-613caab65413-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 09:05:26 crc kubenswrapper[4813]: I1201 09:05:26.073206 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gbncc\" (UniqueName: \"kubernetes.io/projected/5504ea2a-5701-4d37-8f4f-613caab65413-kube-api-access-gbncc\") on node \"crc\" DevicePath \"\"" Dec 01 09:05:26 crc kubenswrapper[4813]: I1201 09:05:26.073219 4813 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5504ea2a-5701-4d37-8f4f-613caab65413-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 01 09:05:26 crc kubenswrapper[4813]: I1201 09:05:26.096664 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/keystone-db-sync-gjdp8"] Dec 01 09:05:26 crc kubenswrapper[4813]: I1201 09:05:26.111695 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/keystone-bootstrap-sv5w6"] Dec 01 09:05:26 crc kubenswrapper[4813]: I1201 09:05:26.118285 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/keystone-bootstrap-sv5w6"] Dec 01 09:05:26 crc kubenswrapper[4813]: I1201 09:05:26.125571 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/keystone-db-sync-gjdp8"] Dec 01 09:05:26 crc kubenswrapper[4813]: I1201 09:05:26.138830 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/keystone-cron-29409661-442fd"] Dec 01 09:05:26 crc kubenswrapper[4813]: I1201 09:05:26.145573 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/keystone-cron-29409661-442fd"] Dec 01 09:05:26 crc kubenswrapper[4813]: I1201 09:05:26.151386 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/keystone-86b5f6997f-8d5h9"] Dec 01 09:05:26 crc kubenswrapper[4813]: I1201 09:05:26.151638 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/keystone-86b5f6997f-8d5h9" podUID="a80abda8-962d-4b0a-bc3a-2e6d693d0393" containerName="keystone-api" containerID="cri-o://54095cf0e86899dcdcc5b94f8681013eb42cc52529b80cec4da277170498a551" gracePeriod=30 Dec 01 09:05:26 crc kubenswrapper[4813]: I1201 09:05:26.163956 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/keystone05f1-account-delete-j85h9"] Dec 01 09:05:26 crc kubenswrapper[4813]: E1201 09:05:26.164343 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5504ea2a-5701-4d37-8f4f-613caab65413" containerName="proxy-server" Dec 01 09:05:26 crc kubenswrapper[4813]: I1201 09:05:26.164373 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="5504ea2a-5701-4d37-8f4f-613caab65413" containerName="proxy-server" Dec 01 09:05:26 crc kubenswrapper[4813]: E1201 09:05:26.164391 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="522f36f6-9da6-46b0-8ca6-9753311b3dfc" containerName="glance-log" Dec 01 09:05:26 crc kubenswrapper[4813]: I1201 09:05:26.164397 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="522f36f6-9da6-46b0-8ca6-9753311b3dfc" containerName="glance-log" Dec 01 09:05:26 crc kubenswrapper[4813]: E1201 09:05:26.164411 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c079746c-0f0e-403a-bb6e-b1866a063050" containerName="glance-httpd" Dec 01 09:05:26 crc kubenswrapper[4813]: I1201 09:05:26.164421 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="c079746c-0f0e-403a-bb6e-b1866a063050" containerName="glance-httpd" Dec 01 09:05:26 crc kubenswrapper[4813]: E1201 09:05:26.164432 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b4c39a03-2938-4b10-b430-369066c79ce4" containerName="glance-log" Dec 01 09:05:26 crc kubenswrapper[4813]: I1201 09:05:26.164438 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="b4c39a03-2938-4b10-b430-369066c79ce4" containerName="glance-log" Dec 01 09:05:26 crc kubenswrapper[4813]: E1201 09:05:26.164450 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c079746c-0f0e-403a-bb6e-b1866a063050" containerName="glance-log" Dec 01 09:05:26 crc kubenswrapper[4813]: I1201 09:05:26.164456 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="c079746c-0f0e-403a-bb6e-b1866a063050" containerName="glance-log" Dec 01 09:05:26 crc kubenswrapper[4813]: E1201 09:05:26.164467 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5504ea2a-5701-4d37-8f4f-613caab65413" containerName="proxy-httpd" Dec 01 09:05:26 crc kubenswrapper[4813]: I1201 09:05:26.164472 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="5504ea2a-5701-4d37-8f4f-613caab65413" containerName="proxy-httpd" Dec 01 09:05:26 crc kubenswrapper[4813]: E1201 09:05:26.164486 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d6f4323d-561b-4725-a1fe-140db5f6a33d" containerName="glance-httpd" Dec 01 09:05:26 crc kubenswrapper[4813]: I1201 09:05:26.164491 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="d6f4323d-561b-4725-a1fe-140db5f6a33d" containerName="glance-httpd" Dec 01 09:05:26 crc kubenswrapper[4813]: E1201 09:05:26.164500 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d6f4323d-561b-4725-a1fe-140db5f6a33d" containerName="glance-log" Dec 01 09:05:26 crc kubenswrapper[4813]: I1201 09:05:26.164505 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="d6f4323d-561b-4725-a1fe-140db5f6a33d" containerName="glance-log" Dec 01 09:05:26 crc kubenswrapper[4813]: E1201 09:05:26.164513 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="54c06e84-7478-4fbc-9e2c-bdc6e1be658b" containerName="mariadb-account-delete" Dec 01 09:05:26 crc kubenswrapper[4813]: I1201 09:05:26.164519 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="54c06e84-7478-4fbc-9e2c-bdc6e1be658b" containerName="mariadb-account-delete" Dec 01 09:05:26 crc kubenswrapper[4813]: E1201 09:05:26.164530 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b4c39a03-2938-4b10-b430-369066c79ce4" containerName="glance-httpd" Dec 01 09:05:26 crc kubenswrapper[4813]: I1201 09:05:26.164535 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="b4c39a03-2938-4b10-b430-369066c79ce4" containerName="glance-httpd" Dec 01 09:05:26 crc kubenswrapper[4813]: E1201 09:05:26.164546 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="522f36f6-9da6-46b0-8ca6-9753311b3dfc" containerName="glance-httpd" Dec 01 09:05:26 crc kubenswrapper[4813]: I1201 09:05:26.164552 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="522f36f6-9da6-46b0-8ca6-9753311b3dfc" containerName="glance-httpd" Dec 01 09:05:26 crc kubenswrapper[4813]: I1201 09:05:26.164689 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="d6f4323d-561b-4725-a1fe-140db5f6a33d" containerName="glance-log" Dec 01 09:05:26 crc kubenswrapper[4813]: I1201 09:05:26.164702 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="522f36f6-9da6-46b0-8ca6-9753311b3dfc" containerName="glance-httpd" Dec 01 09:05:26 crc kubenswrapper[4813]: I1201 09:05:26.164711 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="c079746c-0f0e-403a-bb6e-b1866a063050" containerName="glance-log" Dec 01 09:05:26 crc kubenswrapper[4813]: I1201 09:05:26.164720 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="522f36f6-9da6-46b0-8ca6-9753311b3dfc" containerName="glance-log" Dec 01 09:05:26 crc kubenswrapper[4813]: I1201 09:05:26.164729 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="5504ea2a-5701-4d37-8f4f-613caab65413" containerName="proxy-httpd" Dec 01 09:05:26 crc kubenswrapper[4813]: I1201 09:05:26.164738 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="5504ea2a-5701-4d37-8f4f-613caab65413" containerName="proxy-server" Dec 01 09:05:26 crc kubenswrapper[4813]: I1201 09:05:26.164746 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="c079746c-0f0e-403a-bb6e-b1866a063050" containerName="glance-httpd" Dec 01 09:05:26 crc kubenswrapper[4813]: I1201 09:05:26.164756 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="b4c39a03-2938-4b10-b430-369066c79ce4" containerName="glance-log" Dec 01 09:05:26 crc kubenswrapper[4813]: I1201 09:05:26.164763 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="b4c39a03-2938-4b10-b430-369066c79ce4" containerName="glance-httpd" Dec 01 09:05:26 crc kubenswrapper[4813]: I1201 09:05:26.164771 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="d6f4323d-561b-4725-a1fe-140db5f6a33d" containerName="glance-httpd" Dec 01 09:05:26 crc kubenswrapper[4813]: I1201 09:05:26.164779 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="54c06e84-7478-4fbc-9e2c-bdc6e1be658b" containerName="mariadb-account-delete" Dec 01 09:05:26 crc kubenswrapper[4813]: I1201 09:05:26.165283 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone05f1-account-delete-j85h9" Dec 01 09:05:26 crc kubenswrapper[4813]: I1201 09:05:26.170600 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/keystone05f1-account-delete-j85h9"] Dec 01 09:05:26 crc kubenswrapper[4813]: I1201 09:05:26.275932 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/16ecbca6-953b-4f9f-981e-d9f011703629-operator-scripts\") pod \"keystone05f1-account-delete-j85h9\" (UID: \"16ecbca6-953b-4f9f-981e-d9f011703629\") " pod="glance-kuttl-tests/keystone05f1-account-delete-j85h9" Dec 01 09:05:26 crc kubenswrapper[4813]: I1201 09:05:26.276028 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lwj2f\" (UniqueName: \"kubernetes.io/projected/16ecbca6-953b-4f9f-981e-d9f011703629-kube-api-access-lwj2f\") pod \"keystone05f1-account-delete-j85h9\" (UID: \"16ecbca6-953b-4f9f-981e-d9f011703629\") " pod="glance-kuttl-tests/keystone05f1-account-delete-j85h9" Dec 01 09:05:26 crc kubenswrapper[4813]: I1201 09:05:26.283750 4813 generic.go:334] "Generic (PLEG): container finished" podID="995100a7-4c5a-409a-81f2-9dbaf2fa1d47" containerID="8a6edcb51ab7c3bb702c42f67b2432c8ace1cb2fa02e6a3db3fe9519ecf7f7f3" exitCode=0 Dec 01 09:05:26 crc kubenswrapper[4813]: I1201 09:05:26.283787 4813 generic.go:334] "Generic (PLEG): container finished" podID="995100a7-4c5a-409a-81f2-9dbaf2fa1d47" containerID="90152b553c2c0696cb360f3d2b45d4e4f1d069d88a813084717ae44e3d7f007e" exitCode=0 Dec 01 09:05:26 crc kubenswrapper[4813]: I1201 09:05:26.283796 4813 generic.go:334] "Generic (PLEG): container finished" podID="995100a7-4c5a-409a-81f2-9dbaf2fa1d47" containerID="4a635b3fb091ce09c36d9181c2aa094b2a84e366b1bb19e73b4059e4af3205fa" exitCode=0 Dec 01 09:05:26 crc kubenswrapper[4813]: I1201 09:05:26.283835 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"995100a7-4c5a-409a-81f2-9dbaf2fa1d47","Type":"ContainerDied","Data":"8a6edcb51ab7c3bb702c42f67b2432c8ace1cb2fa02e6a3db3fe9519ecf7f7f3"} Dec 01 09:05:26 crc kubenswrapper[4813]: I1201 09:05:26.283861 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"995100a7-4c5a-409a-81f2-9dbaf2fa1d47","Type":"ContainerDied","Data":"90152b553c2c0696cb360f3d2b45d4e4f1d069d88a813084717ae44e3d7f007e"} Dec 01 09:05:26 crc kubenswrapper[4813]: I1201 09:05:26.283869 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"995100a7-4c5a-409a-81f2-9dbaf2fa1d47","Type":"ContainerDied","Data":"4a635b3fb091ce09c36d9181c2aa094b2a84e366b1bb19e73b4059e4af3205fa"} Dec 01 09:05:26 crc kubenswrapper[4813]: I1201 09:05:26.286556 4813 generic.go:334] "Generic (PLEG): container finished" podID="5504ea2a-5701-4d37-8f4f-613caab65413" containerID="af790186878c3a0dfbe4b88e5964cbbe3ef87f73830bc59c1d00f90328cc6e30" exitCode=0 Dec 01 09:05:26 crc kubenswrapper[4813]: I1201 09:05:26.286580 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-vhtgp" event={"ID":"5504ea2a-5701-4d37-8f4f-613caab65413","Type":"ContainerDied","Data":"af790186878c3a0dfbe4b88e5964cbbe3ef87f73830bc59c1d00f90328cc6e30"} Dec 01 09:05:26 crc kubenswrapper[4813]: I1201 09:05:26.286598 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-vhtgp" event={"ID":"5504ea2a-5701-4d37-8f4f-613caab65413","Type":"ContainerDied","Data":"c12ece2b54c95f80ea4622f335f60a10f8a8d8daed227503d298eff7feff8e5a"} Dec 01 09:05:26 crc kubenswrapper[4813]: I1201 09:05:26.286615 4813 scope.go:117] "RemoveContainer" containerID="af790186878c3a0dfbe4b88e5964cbbe3ef87f73830bc59c1d00f90328cc6e30" Dec 01 09:05:26 crc kubenswrapper[4813]: I1201 09:05:26.286730 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-vhtgp" Dec 01 09:05:26 crc kubenswrapper[4813]: I1201 09:05:26.306593 4813 scope.go:117] "RemoveContainer" containerID="d57b6a9a7684f7aceb73a0c233d97f348bb0a09a608815152313412c1a8f0cdc" Dec 01 09:05:26 crc kubenswrapper[4813]: I1201 09:05:26.318526 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/swift-proxy-6bd58cfcf7-vhtgp"] Dec 01 09:05:26 crc kubenswrapper[4813]: I1201 09:05:26.323789 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/swift-proxy-6bd58cfcf7-vhtgp"] Dec 01 09:05:26 crc kubenswrapper[4813]: I1201 09:05:26.328569 4813 scope.go:117] "RemoveContainer" containerID="af790186878c3a0dfbe4b88e5964cbbe3ef87f73830bc59c1d00f90328cc6e30" Dec 01 09:05:26 crc kubenswrapper[4813]: E1201 09:05:26.329097 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"af790186878c3a0dfbe4b88e5964cbbe3ef87f73830bc59c1d00f90328cc6e30\": container with ID starting with af790186878c3a0dfbe4b88e5964cbbe3ef87f73830bc59c1d00f90328cc6e30 not found: ID does not exist" containerID="af790186878c3a0dfbe4b88e5964cbbe3ef87f73830bc59c1d00f90328cc6e30" Dec 01 09:05:26 crc kubenswrapper[4813]: I1201 09:05:26.329140 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"af790186878c3a0dfbe4b88e5964cbbe3ef87f73830bc59c1d00f90328cc6e30"} err="failed to get container status \"af790186878c3a0dfbe4b88e5964cbbe3ef87f73830bc59c1d00f90328cc6e30\": rpc error: code = NotFound desc = could not find container \"af790186878c3a0dfbe4b88e5964cbbe3ef87f73830bc59c1d00f90328cc6e30\": container with ID starting with af790186878c3a0dfbe4b88e5964cbbe3ef87f73830bc59c1d00f90328cc6e30 not found: ID does not exist" Dec 01 09:05:26 crc kubenswrapper[4813]: I1201 09:05:26.329169 4813 scope.go:117] "RemoveContainer" containerID="d57b6a9a7684f7aceb73a0c233d97f348bb0a09a608815152313412c1a8f0cdc" Dec 01 09:05:26 crc kubenswrapper[4813]: E1201 09:05:26.329448 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d57b6a9a7684f7aceb73a0c233d97f348bb0a09a608815152313412c1a8f0cdc\": container with ID starting with d57b6a9a7684f7aceb73a0c233d97f348bb0a09a608815152313412c1a8f0cdc not found: ID does not exist" containerID="d57b6a9a7684f7aceb73a0c233d97f348bb0a09a608815152313412c1a8f0cdc" Dec 01 09:05:26 crc kubenswrapper[4813]: I1201 09:05:26.329471 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d57b6a9a7684f7aceb73a0c233d97f348bb0a09a608815152313412c1a8f0cdc"} err="failed to get container status \"d57b6a9a7684f7aceb73a0c233d97f348bb0a09a608815152313412c1a8f0cdc\": rpc error: code = NotFound desc = could not find container \"d57b6a9a7684f7aceb73a0c233d97f348bb0a09a608815152313412c1a8f0cdc\": container with ID starting with d57b6a9a7684f7aceb73a0c233d97f348bb0a09a608815152313412c1a8f0cdc not found: ID does not exist" Dec 01 09:05:26 crc kubenswrapper[4813]: I1201 09:05:26.377540 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/16ecbca6-953b-4f9f-981e-d9f011703629-operator-scripts\") pod \"keystone05f1-account-delete-j85h9\" (UID: \"16ecbca6-953b-4f9f-981e-d9f011703629\") " pod="glance-kuttl-tests/keystone05f1-account-delete-j85h9" Dec 01 09:05:26 crc kubenswrapper[4813]: I1201 09:05:26.377660 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lwj2f\" (UniqueName: \"kubernetes.io/projected/16ecbca6-953b-4f9f-981e-d9f011703629-kube-api-access-lwj2f\") pod \"keystone05f1-account-delete-j85h9\" (UID: \"16ecbca6-953b-4f9f-981e-d9f011703629\") " pod="glance-kuttl-tests/keystone05f1-account-delete-j85h9" Dec 01 09:05:26 crc kubenswrapper[4813]: I1201 09:05:26.378371 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/16ecbca6-953b-4f9f-981e-d9f011703629-operator-scripts\") pod \"keystone05f1-account-delete-j85h9\" (UID: \"16ecbca6-953b-4f9f-981e-d9f011703629\") " pod="glance-kuttl-tests/keystone05f1-account-delete-j85h9" Dec 01 09:05:26 crc kubenswrapper[4813]: I1201 09:05:26.409108 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5504ea2a-5701-4d37-8f4f-613caab65413" path="/var/lib/kubelet/pods/5504ea2a-5701-4d37-8f4f-613caab65413/volumes" Dec 01 09:05:26 crc kubenswrapper[4813]: I1201 09:05:26.410011 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="56b5c9f0-c484-4d62-88d9-67c039dd7bca" path="/var/lib/kubelet/pods/56b5c9f0-c484-4d62-88d9-67c039dd7bca/volumes" Dec 01 09:05:26 crc kubenswrapper[4813]: I1201 09:05:26.410698 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6edb6678-d60d-4e1e-92aa-5018a9c1c58d" path="/var/lib/kubelet/pods/6edb6678-d60d-4e1e-92aa-5018a9c1c58d/volumes" Dec 01 09:05:26 crc kubenswrapper[4813]: I1201 09:05:26.412624 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="881622cc-f398-4ea0-8fef-a443848b664f" path="/var/lib/kubelet/pods/881622cc-f398-4ea0-8fef-a443848b664f/volumes" Dec 01 09:05:26 crc kubenswrapper[4813]: I1201 09:05:26.413319 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b247a555-0805-4884-a0d5-d2799423bf94" path="/var/lib/kubelet/pods/b247a555-0805-4884-a0d5-d2799423bf94/volumes" Dec 01 09:05:26 crc kubenswrapper[4813]: I1201 09:05:26.417616 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lwj2f\" (UniqueName: \"kubernetes.io/projected/16ecbca6-953b-4f9f-981e-d9f011703629-kube-api-access-lwj2f\") pod \"keystone05f1-account-delete-j85h9\" (UID: \"16ecbca6-953b-4f9f-981e-d9f011703629\") " pod="glance-kuttl-tests/keystone05f1-account-delete-j85h9" Dec 01 09:05:26 crc kubenswrapper[4813]: I1201 09:05:26.446526 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-g4dzh"] Dec 01 09:05:26 crc kubenswrapper[4813]: I1201 09:05:26.451082 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-g4dzh" Dec 01 09:05:26 crc kubenswrapper[4813]: I1201 09:05:26.464740 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-g4dzh"] Dec 01 09:05:26 crc kubenswrapper[4813]: I1201 09:05:26.485073 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone05f1-account-delete-j85h9" Dec 01 09:05:26 crc kubenswrapper[4813]: I1201 09:05:26.580738 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/07bbfadb-d0a3-478c-a2ed-36ddc6081a49-catalog-content\") pod \"community-operators-g4dzh\" (UID: \"07bbfadb-d0a3-478c-a2ed-36ddc6081a49\") " pod="openshift-marketplace/community-operators-g4dzh" Dec 01 09:05:26 crc kubenswrapper[4813]: I1201 09:05:26.581092 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/07bbfadb-d0a3-478c-a2ed-36ddc6081a49-utilities\") pod \"community-operators-g4dzh\" (UID: \"07bbfadb-d0a3-478c-a2ed-36ddc6081a49\") " pod="openshift-marketplace/community-operators-g4dzh" Dec 01 09:05:26 crc kubenswrapper[4813]: I1201 09:05:26.581213 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mq4l5\" (UniqueName: \"kubernetes.io/projected/07bbfadb-d0a3-478c-a2ed-36ddc6081a49-kube-api-access-mq4l5\") pod \"community-operators-g4dzh\" (UID: \"07bbfadb-d0a3-478c-a2ed-36ddc6081a49\") " pod="openshift-marketplace/community-operators-g4dzh" Dec 01 09:05:26 crc kubenswrapper[4813]: I1201 09:05:26.682989 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mq4l5\" (UniqueName: \"kubernetes.io/projected/07bbfadb-d0a3-478c-a2ed-36ddc6081a49-kube-api-access-mq4l5\") pod \"community-operators-g4dzh\" (UID: \"07bbfadb-d0a3-478c-a2ed-36ddc6081a49\") " pod="openshift-marketplace/community-operators-g4dzh" Dec 01 09:05:26 crc kubenswrapper[4813]: I1201 09:05:26.683058 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/07bbfadb-d0a3-478c-a2ed-36ddc6081a49-catalog-content\") pod \"community-operators-g4dzh\" (UID: \"07bbfadb-d0a3-478c-a2ed-36ddc6081a49\") " pod="openshift-marketplace/community-operators-g4dzh" Dec 01 09:05:26 crc kubenswrapper[4813]: I1201 09:05:26.683108 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/07bbfadb-d0a3-478c-a2ed-36ddc6081a49-utilities\") pod \"community-operators-g4dzh\" (UID: \"07bbfadb-d0a3-478c-a2ed-36ddc6081a49\") " pod="openshift-marketplace/community-operators-g4dzh" Dec 01 09:05:26 crc kubenswrapper[4813]: I1201 09:05:26.683805 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/07bbfadb-d0a3-478c-a2ed-36ddc6081a49-utilities\") pod \"community-operators-g4dzh\" (UID: \"07bbfadb-d0a3-478c-a2ed-36ddc6081a49\") " pod="openshift-marketplace/community-operators-g4dzh" Dec 01 09:05:26 crc kubenswrapper[4813]: I1201 09:05:26.684445 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/07bbfadb-d0a3-478c-a2ed-36ddc6081a49-catalog-content\") pod \"community-operators-g4dzh\" (UID: \"07bbfadb-d0a3-478c-a2ed-36ddc6081a49\") " pod="openshift-marketplace/community-operators-g4dzh" Dec 01 09:05:26 crc kubenswrapper[4813]: I1201 09:05:26.700482 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mq4l5\" (UniqueName: \"kubernetes.io/projected/07bbfadb-d0a3-478c-a2ed-36ddc6081a49-kube-api-access-mq4l5\") pod \"community-operators-g4dzh\" (UID: \"07bbfadb-d0a3-478c-a2ed-36ddc6081a49\") " pod="openshift-marketplace/community-operators-g4dzh" Dec 01 09:05:26 crc kubenswrapper[4813]: I1201 09:05:26.772009 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-g4dzh" Dec 01 09:05:27 crc kubenswrapper[4813]: I1201 09:05:27.401872 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/keystone05f1-account-delete-j85h9"] Dec 01 09:05:27 crc kubenswrapper[4813]: I1201 09:05:27.513036 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-g4dzh"] Dec 01 09:05:28 crc kubenswrapper[4813]: I1201 09:05:28.304130 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone05f1-account-delete-j85h9" event={"ID":"16ecbca6-953b-4f9f-981e-d9f011703629","Type":"ContainerStarted","Data":"65dc583e88fe1abbbaf461f2f189e99d2cefde16fd6c99f281b62ca2548dc02a"} Dec 01 09:05:28 crc kubenswrapper[4813]: I1201 09:05:28.304484 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone05f1-account-delete-j85h9" event={"ID":"16ecbca6-953b-4f9f-981e-d9f011703629","Type":"ContainerStarted","Data":"6710c7830741be27bd771bfe54743d0cf61017ebc676538798dc5d73146dfc00"} Dec 01 09:05:28 crc kubenswrapper[4813]: I1201 09:05:28.305857 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-g4dzh" event={"ID":"07bbfadb-d0a3-478c-a2ed-36ddc6081a49","Type":"ContainerStarted","Data":"5da0e0319711e8d1c40aa687a5875d18ae14d499d0d65c03d94a8e6ab8fc38e3"} Dec 01 09:05:29 crc kubenswrapper[4813]: I1201 09:05:29.317032 4813 generic.go:334] "Generic (PLEG): container finished" podID="07bbfadb-d0a3-478c-a2ed-36ddc6081a49" containerID="8933d3fb6c6148d55759aacf3a471c3b08aa2b10859ecd7a6d3421892e4018df" exitCode=0 Dec 01 09:05:29 crc kubenswrapper[4813]: I1201 09:05:29.318438 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-g4dzh" event={"ID":"07bbfadb-d0a3-478c-a2ed-36ddc6081a49","Type":"ContainerDied","Data":"8933d3fb6c6148d55759aacf3a471c3b08aa2b10859ecd7a6d3421892e4018df"} Dec 01 09:05:29 crc kubenswrapper[4813]: I1201 09:05:29.319394 4813 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 01 09:05:29 crc kubenswrapper[4813]: I1201 09:05:29.363680 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/keystone05f1-account-delete-j85h9" podStartSLOduration=3.363634023 podStartE2EDuration="3.363634023s" podCreationTimestamp="2025-12-01 09:05:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 09:05:29.355160672 +0000 UTC m=+1513.978082258" watchObservedRunningTime="2025-12-01 09:05:29.363634023 +0000 UTC m=+1513.986555619" Dec 01 09:05:29 crc kubenswrapper[4813]: I1201 09:05:29.504376 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/openstack-galera-1"] Dec 01 09:05:29 crc kubenswrapper[4813]: I1201 09:05:29.514093 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/openstack-galera-2"] Dec 01 09:05:29 crc kubenswrapper[4813]: I1201 09:05:29.524833 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/openstack-galera-0"] Dec 01 09:05:29 crc kubenswrapper[4813]: I1201 09:05:29.661048 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/openstack-galera-2" podUID="73988763-722d-4652-8fac-053ba8217547" containerName="galera" containerID="cri-o://f7024039c493e9165151be3914c80e6c5d5142a4eaf9d9c184364cc64a3f8f0e" gracePeriod=30 Dec 01 09:05:29 crc kubenswrapper[4813]: I1201 09:05:29.694257 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-86b5f6997f-8d5h9" Dec 01 09:05:29 crc kubenswrapper[4813]: I1201 09:05:29.848885 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a80abda8-962d-4b0a-bc3a-2e6d693d0393-config-data\") pod \"a80abda8-962d-4b0a-bc3a-2e6d693d0393\" (UID: \"a80abda8-962d-4b0a-bc3a-2e6d693d0393\") " Dec 01 09:05:29 crc kubenswrapper[4813]: I1201 09:05:29.848937 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a80abda8-962d-4b0a-bc3a-2e6d693d0393-scripts\") pod \"a80abda8-962d-4b0a-bc3a-2e6d693d0393\" (UID: \"a80abda8-962d-4b0a-bc3a-2e6d693d0393\") " Dec 01 09:05:29 crc kubenswrapper[4813]: I1201 09:05:29.849033 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fjj2b\" (UniqueName: \"kubernetes.io/projected/a80abda8-962d-4b0a-bc3a-2e6d693d0393-kube-api-access-fjj2b\") pod \"a80abda8-962d-4b0a-bc3a-2e6d693d0393\" (UID: \"a80abda8-962d-4b0a-bc3a-2e6d693d0393\") " Dec 01 09:05:29 crc kubenswrapper[4813]: I1201 09:05:29.849105 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/a80abda8-962d-4b0a-bc3a-2e6d693d0393-fernet-keys\") pod \"a80abda8-962d-4b0a-bc3a-2e6d693d0393\" (UID: \"a80abda8-962d-4b0a-bc3a-2e6d693d0393\") " Dec 01 09:05:29 crc kubenswrapper[4813]: I1201 09:05:29.849149 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/a80abda8-962d-4b0a-bc3a-2e6d693d0393-credential-keys\") pod \"a80abda8-962d-4b0a-bc3a-2e6d693d0393\" (UID: \"a80abda8-962d-4b0a-bc3a-2e6d693d0393\") " Dec 01 09:05:29 crc kubenswrapper[4813]: I1201 09:05:29.859641 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a80abda8-962d-4b0a-bc3a-2e6d693d0393-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "a80abda8-962d-4b0a-bc3a-2e6d693d0393" (UID: "a80abda8-962d-4b0a-bc3a-2e6d693d0393"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:05:29 crc kubenswrapper[4813]: I1201 09:05:29.860190 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a80abda8-962d-4b0a-bc3a-2e6d693d0393-kube-api-access-fjj2b" (OuterVolumeSpecName: "kube-api-access-fjj2b") pod "a80abda8-962d-4b0a-bc3a-2e6d693d0393" (UID: "a80abda8-962d-4b0a-bc3a-2e6d693d0393"). InnerVolumeSpecName "kube-api-access-fjj2b". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:05:29 crc kubenswrapper[4813]: I1201 09:05:29.861556 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a80abda8-962d-4b0a-bc3a-2e6d693d0393-scripts" (OuterVolumeSpecName: "scripts") pod "a80abda8-962d-4b0a-bc3a-2e6d693d0393" (UID: "a80abda8-962d-4b0a-bc3a-2e6d693d0393"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:05:29 crc kubenswrapper[4813]: I1201 09:05:29.861660 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a80abda8-962d-4b0a-bc3a-2e6d693d0393-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "a80abda8-962d-4b0a-bc3a-2e6d693d0393" (UID: "a80abda8-962d-4b0a-bc3a-2e6d693d0393"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:05:29 crc kubenswrapper[4813]: I1201 09:05:29.877471 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a80abda8-962d-4b0a-bc3a-2e6d693d0393-config-data" (OuterVolumeSpecName: "config-data") pod "a80abda8-962d-4b0a-bc3a-2e6d693d0393" (UID: "a80abda8-962d-4b0a-bc3a-2e6d693d0393"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:05:29 crc kubenswrapper[4813]: I1201 09:05:29.951078 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fjj2b\" (UniqueName: \"kubernetes.io/projected/a80abda8-962d-4b0a-bc3a-2e6d693d0393-kube-api-access-fjj2b\") on node \"crc\" DevicePath \"\"" Dec 01 09:05:29 crc kubenswrapper[4813]: I1201 09:05:29.951116 4813 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/a80abda8-962d-4b0a-bc3a-2e6d693d0393-fernet-keys\") on node \"crc\" DevicePath \"\"" Dec 01 09:05:29 crc kubenswrapper[4813]: I1201 09:05:29.951127 4813 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/a80abda8-962d-4b0a-bc3a-2e6d693d0393-credential-keys\") on node \"crc\" DevicePath \"\"" Dec 01 09:05:29 crc kubenswrapper[4813]: I1201 09:05:29.951138 4813 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a80abda8-962d-4b0a-bc3a-2e6d693d0393-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 09:05:29 crc kubenswrapper[4813]: I1201 09:05:29.951148 4813 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a80abda8-962d-4b0a-bc3a-2e6d693d0393-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 09:05:30 crc kubenswrapper[4813]: I1201 09:05:30.202731 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/memcached-0"] Dec 01 09:05:30 crc kubenswrapper[4813]: I1201 09:05:30.203317 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/memcached-0" podUID="88d8ade8-c257-464c-9dab-0052bed99037" containerName="memcached" containerID="cri-o://13647640697f73f4b7f03b0bcbd720969780ef3b11f627085ddd2eb78f70d473" gracePeriod=30 Dec 01 09:05:30 crc kubenswrapper[4813]: I1201 09:05:30.326944 4813 generic.go:334] "Generic (PLEG): container finished" podID="16ecbca6-953b-4f9f-981e-d9f011703629" containerID="65dc583e88fe1abbbaf461f2f189e99d2cefde16fd6c99f281b62ca2548dc02a" exitCode=0 Dec 01 09:05:30 crc kubenswrapper[4813]: I1201 09:05:30.327070 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone05f1-account-delete-j85h9" event={"ID":"16ecbca6-953b-4f9f-981e-d9f011703629","Type":"ContainerDied","Data":"65dc583e88fe1abbbaf461f2f189e99d2cefde16fd6c99f281b62ca2548dc02a"} Dec 01 09:05:30 crc kubenswrapper[4813]: I1201 09:05:30.328814 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-g4dzh" event={"ID":"07bbfadb-d0a3-478c-a2ed-36ddc6081a49","Type":"ContainerStarted","Data":"b0a66c5eb5f9de05c01d2a0ac255e1682f4a982e3742f4ccd7de893a7165976b"} Dec 01 09:05:30 crc kubenswrapper[4813]: I1201 09:05:30.330125 4813 generic.go:334] "Generic (PLEG): container finished" podID="a80abda8-962d-4b0a-bc3a-2e6d693d0393" containerID="54095cf0e86899dcdcc5b94f8681013eb42cc52529b80cec4da277170498a551" exitCode=0 Dec 01 09:05:30 crc kubenswrapper[4813]: I1201 09:05:30.330165 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-86b5f6997f-8d5h9" event={"ID":"a80abda8-962d-4b0a-bc3a-2e6d693d0393","Type":"ContainerDied","Data":"54095cf0e86899dcdcc5b94f8681013eb42cc52529b80cec4da277170498a551"} Dec 01 09:05:30 crc kubenswrapper[4813]: I1201 09:05:30.330192 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-86b5f6997f-8d5h9" event={"ID":"a80abda8-962d-4b0a-bc3a-2e6d693d0393","Type":"ContainerDied","Data":"5700b50bae4953f211699772cee4993299ebab0fbba90fc0e8fd84d8f4f66a25"} Dec 01 09:05:30 crc kubenswrapper[4813]: I1201 09:05:30.330214 4813 scope.go:117] "RemoveContainer" containerID="54095cf0e86899dcdcc5b94f8681013eb42cc52529b80cec4da277170498a551" Dec 01 09:05:30 crc kubenswrapper[4813]: I1201 09:05:30.330216 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-86b5f6997f-8d5h9" Dec 01 09:05:30 crc kubenswrapper[4813]: I1201 09:05:30.350567 4813 scope.go:117] "RemoveContainer" containerID="54095cf0e86899dcdcc5b94f8681013eb42cc52529b80cec4da277170498a551" Dec 01 09:05:30 crc kubenswrapper[4813]: E1201 09:05:30.350947 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"54095cf0e86899dcdcc5b94f8681013eb42cc52529b80cec4da277170498a551\": container with ID starting with 54095cf0e86899dcdcc5b94f8681013eb42cc52529b80cec4da277170498a551 not found: ID does not exist" containerID="54095cf0e86899dcdcc5b94f8681013eb42cc52529b80cec4da277170498a551" Dec 01 09:05:30 crc kubenswrapper[4813]: I1201 09:05:30.351007 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"54095cf0e86899dcdcc5b94f8681013eb42cc52529b80cec4da277170498a551"} err="failed to get container status \"54095cf0e86899dcdcc5b94f8681013eb42cc52529b80cec4da277170498a551\": rpc error: code = NotFound desc = could not find container \"54095cf0e86899dcdcc5b94f8681013eb42cc52529b80cec4da277170498a551\": container with ID starting with 54095cf0e86899dcdcc5b94f8681013eb42cc52529b80cec4da277170498a551 not found: ID does not exist" Dec 01 09:05:30 crc kubenswrapper[4813]: I1201 09:05:30.388230 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/keystone-86b5f6997f-8d5h9"] Dec 01 09:05:30 crc kubenswrapper[4813]: I1201 09:05:30.401249 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/keystone-86b5f6997f-8d5h9"] Dec 01 09:05:30 crc kubenswrapper[4813]: I1201 09:05:30.614598 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/rabbitmq-server-0"] Dec 01 09:05:30 crc kubenswrapper[4813]: I1201 09:05:30.748132 4813 prober.go:107] "Probe failed" probeType="Readiness" pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-vhtgp" podUID="5504ea2a-5701-4d37-8f4f-613caab65413" containerName="proxy-httpd" probeResult="failure" output="Get \"http://10.217.0.90:8080/healthcheck\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 01 09:05:30 crc kubenswrapper[4813]: I1201 09:05:30.748183 4813 prober.go:107] "Probe failed" probeType="Readiness" pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-vhtgp" podUID="5504ea2a-5701-4d37-8f4f-613caab65413" containerName="proxy-server" probeResult="failure" output="Get \"http://10.217.0.90:8080/healthcheck\": dial tcp 10.217.0.90:8080: i/o timeout (Client.Timeout exceeded while awaiting headers)" Dec 01 09:05:31 crc kubenswrapper[4813]: I1201 09:05:31.011432 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/rabbitmq-server-0"] Dec 01 09:05:31 crc kubenswrapper[4813]: I1201 09:05:31.085529 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/openstack-galera-2" Dec 01 09:05:31 crc kubenswrapper[4813]: I1201 09:05:31.151628 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/keystone-db-create-lv649"] Dec 01 09:05:31 crc kubenswrapper[4813]: I1201 09:05:31.160641 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/keystone-db-create-lv649"] Dec 01 09:05:31 crc kubenswrapper[4813]: I1201 09:05:31.171243 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/keystone-05f1-account-create-update-r6np6"] Dec 01 09:05:31 crc kubenswrapper[4813]: I1201 09:05:31.174866 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/73988763-722d-4652-8fac-053ba8217547-config-data-default\") pod \"73988763-722d-4652-8fac-053ba8217547\" (UID: \"73988763-722d-4652-8fac-053ba8217547\") " Dec 01 09:05:31 crc kubenswrapper[4813]: I1201 09:05:31.174912 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mysql-db\" (UniqueName: \"kubernetes.io/local-volume/local-storage15-crc\") pod \"73988763-722d-4652-8fac-053ba8217547\" (UID: \"73988763-722d-4652-8fac-053ba8217547\") " Dec 01 09:05:31 crc kubenswrapper[4813]: I1201 09:05:31.175013 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/73988763-722d-4652-8fac-053ba8217547-operator-scripts\") pod \"73988763-722d-4652-8fac-053ba8217547\" (UID: \"73988763-722d-4652-8fac-053ba8217547\") " Dec 01 09:05:31 crc kubenswrapper[4813]: I1201 09:05:31.175048 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/73988763-722d-4652-8fac-053ba8217547-kolla-config\") pod \"73988763-722d-4652-8fac-053ba8217547\" (UID: \"73988763-722d-4652-8fac-053ba8217547\") " Dec 01 09:05:31 crc kubenswrapper[4813]: I1201 09:05:31.175093 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-str7c\" (UniqueName: \"kubernetes.io/projected/73988763-722d-4652-8fac-053ba8217547-kube-api-access-str7c\") pod \"73988763-722d-4652-8fac-053ba8217547\" (UID: \"73988763-722d-4652-8fac-053ba8217547\") " Dec 01 09:05:31 crc kubenswrapper[4813]: I1201 09:05:31.175124 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/73988763-722d-4652-8fac-053ba8217547-config-data-generated\") pod \"73988763-722d-4652-8fac-053ba8217547\" (UID: \"73988763-722d-4652-8fac-053ba8217547\") " Dec 01 09:05:31 crc kubenswrapper[4813]: I1201 09:05:31.175592 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/73988763-722d-4652-8fac-053ba8217547-kolla-config" (OuterVolumeSpecName: "kolla-config") pod "73988763-722d-4652-8fac-053ba8217547" (UID: "73988763-722d-4652-8fac-053ba8217547"). InnerVolumeSpecName "kolla-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 09:05:31 crc kubenswrapper[4813]: I1201 09:05:31.176211 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/73988763-722d-4652-8fac-053ba8217547-config-data-default" (OuterVolumeSpecName: "config-data-default") pod "73988763-722d-4652-8fac-053ba8217547" (UID: "73988763-722d-4652-8fac-053ba8217547"). InnerVolumeSpecName "config-data-default". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 09:05:31 crc kubenswrapper[4813]: I1201 09:05:31.176267 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/73988763-722d-4652-8fac-053ba8217547-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "73988763-722d-4652-8fac-053ba8217547" (UID: "73988763-722d-4652-8fac-053ba8217547"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 09:05:31 crc kubenswrapper[4813]: I1201 09:05:31.176504 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/73988763-722d-4652-8fac-053ba8217547-config-data-generated" (OuterVolumeSpecName: "config-data-generated") pod "73988763-722d-4652-8fac-053ba8217547" (UID: "73988763-722d-4652-8fac-053ba8217547"). InnerVolumeSpecName "config-data-generated". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 09:05:31 crc kubenswrapper[4813]: I1201 09:05:31.181693 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/keystone-05f1-account-create-update-r6np6"] Dec 01 09:05:31 crc kubenswrapper[4813]: I1201 09:05:31.181846 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/73988763-722d-4652-8fac-053ba8217547-kube-api-access-str7c" (OuterVolumeSpecName: "kube-api-access-str7c") pod "73988763-722d-4652-8fac-053ba8217547" (UID: "73988763-722d-4652-8fac-053ba8217547"). InnerVolumeSpecName "kube-api-access-str7c". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:05:31 crc kubenswrapper[4813]: I1201 09:05:31.186237 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/keystone05f1-account-delete-j85h9"] Dec 01 09:05:31 crc kubenswrapper[4813]: I1201 09:05:31.187888 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage15-crc" (OuterVolumeSpecName: "mysql-db") pod "73988763-722d-4652-8fac-053ba8217547" (UID: "73988763-722d-4652-8fac-053ba8217547"). InnerVolumeSpecName "local-storage15-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 01 09:05:31 crc kubenswrapper[4813]: I1201 09:05:31.277119 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-str7c\" (UniqueName: \"kubernetes.io/projected/73988763-722d-4652-8fac-053ba8217547-kube-api-access-str7c\") on node \"crc\" DevicePath \"\"" Dec 01 09:05:31 crc kubenswrapper[4813]: I1201 09:05:31.277184 4813 reconciler_common.go:293] "Volume detached for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/73988763-722d-4652-8fac-053ba8217547-config-data-generated\") on node \"crc\" DevicePath \"\"" Dec 01 09:05:31 crc kubenswrapper[4813]: I1201 09:05:31.277207 4813 reconciler_common.go:293] "Volume detached for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/73988763-722d-4652-8fac-053ba8217547-config-data-default\") on node \"crc\" DevicePath \"\"" Dec 01 09:05:31 crc kubenswrapper[4813]: I1201 09:05:31.277266 4813 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage15-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage15-crc\") on node \"crc\" " Dec 01 09:05:31 crc kubenswrapper[4813]: I1201 09:05:31.277287 4813 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/73988763-722d-4652-8fac-053ba8217547-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 09:05:31 crc kubenswrapper[4813]: I1201 09:05:31.277305 4813 reconciler_common.go:293] "Volume detached for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/73988763-722d-4652-8fac-053ba8217547-kolla-config\") on node \"crc\" DevicePath \"\"" Dec 01 09:05:31 crc kubenswrapper[4813]: I1201 09:05:31.294282 4813 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage15-crc" (UniqueName: "kubernetes.io/local-volume/local-storage15-crc") on node "crc" Dec 01 09:05:31 crc kubenswrapper[4813]: I1201 09:05:31.344676 4813 generic.go:334] "Generic (PLEG): container finished" podID="88d8ade8-c257-464c-9dab-0052bed99037" containerID="13647640697f73f4b7f03b0bcbd720969780ef3b11f627085ddd2eb78f70d473" exitCode=0 Dec 01 09:05:31 crc kubenswrapper[4813]: I1201 09:05:31.344785 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/memcached-0" event={"ID":"88d8ade8-c257-464c-9dab-0052bed99037","Type":"ContainerDied","Data":"13647640697f73f4b7f03b0bcbd720969780ef3b11f627085ddd2eb78f70d473"} Dec 01 09:05:31 crc kubenswrapper[4813]: I1201 09:05:31.348478 4813 generic.go:334] "Generic (PLEG): container finished" podID="07bbfadb-d0a3-478c-a2ed-36ddc6081a49" containerID="b0a66c5eb5f9de05c01d2a0ac255e1682f4a982e3742f4ccd7de893a7165976b" exitCode=0 Dec 01 09:05:31 crc kubenswrapper[4813]: I1201 09:05:31.348655 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-g4dzh" event={"ID":"07bbfadb-d0a3-478c-a2ed-36ddc6081a49","Type":"ContainerDied","Data":"b0a66c5eb5f9de05c01d2a0ac255e1682f4a982e3742f4ccd7de893a7165976b"} Dec 01 09:05:31 crc kubenswrapper[4813]: I1201 09:05:31.356109 4813 generic.go:334] "Generic (PLEG): container finished" podID="73988763-722d-4652-8fac-053ba8217547" containerID="f7024039c493e9165151be3914c80e6c5d5142a4eaf9d9c184364cc64a3f8f0e" exitCode=0 Dec 01 09:05:31 crc kubenswrapper[4813]: I1201 09:05:31.356155 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstack-galera-2" event={"ID":"73988763-722d-4652-8fac-053ba8217547","Type":"ContainerDied","Data":"f7024039c493e9165151be3914c80e6c5d5142a4eaf9d9c184364cc64a3f8f0e"} Dec 01 09:05:31 crc kubenswrapper[4813]: I1201 09:05:31.356201 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstack-galera-2" event={"ID":"73988763-722d-4652-8fac-053ba8217547","Type":"ContainerDied","Data":"ab843b8177030ce9338918bf46de7538032e92a628fb669df6123fd1ee964d6c"} Dec 01 09:05:31 crc kubenswrapper[4813]: I1201 09:05:31.356223 4813 scope.go:117] "RemoveContainer" containerID="f7024039c493e9165151be3914c80e6c5d5142a4eaf9d9c184364cc64a3f8f0e" Dec 01 09:05:31 crc kubenswrapper[4813]: I1201 09:05:31.356276 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/openstack-galera-2" Dec 01 09:05:31 crc kubenswrapper[4813]: I1201 09:05:31.378561 4813 reconciler_common.go:293] "Volume detached for volume \"local-storage15-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage15-crc\") on node \"crc\" DevicePath \"\"" Dec 01 09:05:31 crc kubenswrapper[4813]: I1201 09:05:31.393220 4813 scope.go:117] "RemoveContainer" containerID="3642faeb2d7a5ddd953c03fd847808337be3eca66a804d81a464654fcd35ae82" Dec 01 09:05:31 crc kubenswrapper[4813]: I1201 09:05:31.418457 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/rabbitmq-server-0" podUID="a274574f-1dfe-4afd-8dfa-661d14669e73" containerName="rabbitmq" containerID="cri-o://e9eef734cf66c90ff9f0f04280a9c11fc9ecb34a8b6da3055227e88543656ffe" gracePeriod=604800 Dec 01 09:05:31 crc kubenswrapper[4813]: I1201 09:05:31.462258 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/openstack-galera-2"] Dec 01 09:05:31 crc kubenswrapper[4813]: I1201 09:05:31.471526 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/openstack-galera-2"] Dec 01 09:05:31 crc kubenswrapper[4813]: I1201 09:05:31.474466 4813 scope.go:117] "RemoveContainer" containerID="f7024039c493e9165151be3914c80e6c5d5142a4eaf9d9c184364cc64a3f8f0e" Dec 01 09:05:31 crc kubenswrapper[4813]: E1201 09:05:31.475274 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f7024039c493e9165151be3914c80e6c5d5142a4eaf9d9c184364cc64a3f8f0e\": container with ID starting with f7024039c493e9165151be3914c80e6c5d5142a4eaf9d9c184364cc64a3f8f0e not found: ID does not exist" containerID="f7024039c493e9165151be3914c80e6c5d5142a4eaf9d9c184364cc64a3f8f0e" Dec 01 09:05:31 crc kubenswrapper[4813]: I1201 09:05:31.475335 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f7024039c493e9165151be3914c80e6c5d5142a4eaf9d9c184364cc64a3f8f0e"} err="failed to get container status \"f7024039c493e9165151be3914c80e6c5d5142a4eaf9d9c184364cc64a3f8f0e\": rpc error: code = NotFound desc = could not find container \"f7024039c493e9165151be3914c80e6c5d5142a4eaf9d9c184364cc64a3f8f0e\": container with ID starting with f7024039c493e9165151be3914c80e6c5d5142a4eaf9d9c184364cc64a3f8f0e not found: ID does not exist" Dec 01 09:05:31 crc kubenswrapper[4813]: I1201 09:05:31.475374 4813 scope.go:117] "RemoveContainer" containerID="3642faeb2d7a5ddd953c03fd847808337be3eca66a804d81a464654fcd35ae82" Dec 01 09:05:31 crc kubenswrapper[4813]: E1201 09:05:31.478152 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3642faeb2d7a5ddd953c03fd847808337be3eca66a804d81a464654fcd35ae82\": container with ID starting with 3642faeb2d7a5ddd953c03fd847808337be3eca66a804d81a464654fcd35ae82 not found: ID does not exist" containerID="3642faeb2d7a5ddd953c03fd847808337be3eca66a804d81a464654fcd35ae82" Dec 01 09:05:31 crc kubenswrapper[4813]: I1201 09:05:31.478204 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3642faeb2d7a5ddd953c03fd847808337be3eca66a804d81a464654fcd35ae82"} err="failed to get container status \"3642faeb2d7a5ddd953c03fd847808337be3eca66a804d81a464654fcd35ae82\": rpc error: code = NotFound desc = could not find container \"3642faeb2d7a5ddd953c03fd847808337be3eca66a804d81a464654fcd35ae82\": container with ID starting with 3642faeb2d7a5ddd953c03fd847808337be3eca66a804d81a464654fcd35ae82 not found: ID does not exist" Dec 01 09:05:31 crc kubenswrapper[4813]: I1201 09:05:31.738536 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone05f1-account-delete-j85h9" Dec 01 09:05:31 crc kubenswrapper[4813]: I1201 09:05:31.845710 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/glance-operator-controller-manager-59d87c8d7c-vcbsn"] Dec 01 09:05:31 crc kubenswrapper[4813]: I1201 09:05:31.846266 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/glance-operator-controller-manager-59d87c8d7c-vcbsn" podUID="3e52d3b1-2c61-4e2b-8222-35745063e838" containerName="manager" containerID="cri-o://46fb724958d5e9eced4e641498c92ea22e8ac57e844321ca8c097a007224858b" gracePeriod=10 Dec 01 09:05:31 crc kubenswrapper[4813]: I1201 09:05:31.890749 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/openstack-galera-1" podUID="41360e5e-2620-4912-8035-0e1a9e0e715d" containerName="galera" containerID="cri-o://f186dd2ac93b4f28668082627ae26d9d2e81b0ec2304484ef52e31414559378e" gracePeriod=28 Dec 01 09:05:31 crc kubenswrapper[4813]: I1201 09:05:31.944264 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/16ecbca6-953b-4f9f-981e-d9f011703629-operator-scripts\") pod \"16ecbca6-953b-4f9f-981e-d9f011703629\" (UID: \"16ecbca6-953b-4f9f-981e-d9f011703629\") " Dec 01 09:05:31 crc kubenswrapper[4813]: I1201 09:05:31.944317 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lwj2f\" (UniqueName: \"kubernetes.io/projected/16ecbca6-953b-4f9f-981e-d9f011703629-kube-api-access-lwj2f\") pod \"16ecbca6-953b-4f9f-981e-d9f011703629\" (UID: \"16ecbca6-953b-4f9f-981e-d9f011703629\") " Dec 01 09:05:31 crc kubenswrapper[4813]: I1201 09:05:31.945502 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/16ecbca6-953b-4f9f-981e-d9f011703629-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "16ecbca6-953b-4f9f-981e-d9f011703629" (UID: "16ecbca6-953b-4f9f-981e-d9f011703629"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 09:05:31 crc kubenswrapper[4813]: I1201 09:05:31.953250 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/16ecbca6-953b-4f9f-981e-d9f011703629-kube-api-access-lwj2f" (OuterVolumeSpecName: "kube-api-access-lwj2f") pod "16ecbca6-953b-4f9f-981e-d9f011703629" (UID: "16ecbca6-953b-4f9f-981e-d9f011703629"). InnerVolumeSpecName "kube-api-access-lwj2f". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:05:32 crc kubenswrapper[4813]: I1201 09:05:32.047764 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lwj2f\" (UniqueName: \"kubernetes.io/projected/16ecbca6-953b-4f9f-981e-d9f011703629-kube-api-access-lwj2f\") on node \"crc\" DevicePath \"\"" Dec 01 09:05:32 crc kubenswrapper[4813]: I1201 09:05:32.047792 4813 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/16ecbca6-953b-4f9f-981e-d9f011703629-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 09:05:32 crc kubenswrapper[4813]: I1201 09:05:32.078299 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/glance-operator-index-vlgfp"] Dec 01 09:05:32 crc kubenswrapper[4813]: I1201 09:05:32.108142 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/254139c32b21a718739d6f558ac0bb26827dad1d88305bb3f8b52462e1xbkjm"] Dec 01 09:05:32 crc kubenswrapper[4813]: I1201 09:05:32.120212 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/254139c32b21a718739d6f558ac0bb26827dad1d88305bb3f8b52462e1xbkjm"] Dec 01 09:05:32 crc kubenswrapper[4813]: I1201 09:05:32.224488 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/memcached-0" Dec 01 09:05:32 crc kubenswrapper[4813]: I1201 09:05:32.252080 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/88d8ade8-c257-464c-9dab-0052bed99037-kolla-config\") pod \"88d8ade8-c257-464c-9dab-0052bed99037\" (UID: \"88d8ade8-c257-464c-9dab-0052bed99037\") " Dec 01 09:05:32 crc kubenswrapper[4813]: I1201 09:05:32.252130 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/88d8ade8-c257-464c-9dab-0052bed99037-config-data\") pod \"88d8ade8-c257-464c-9dab-0052bed99037\" (UID: \"88d8ade8-c257-464c-9dab-0052bed99037\") " Dec 01 09:05:32 crc kubenswrapper[4813]: I1201 09:05:32.252190 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j5c68\" (UniqueName: \"kubernetes.io/projected/88d8ade8-c257-464c-9dab-0052bed99037-kube-api-access-j5c68\") pod \"88d8ade8-c257-464c-9dab-0052bed99037\" (UID: \"88d8ade8-c257-464c-9dab-0052bed99037\") " Dec 01 09:05:32 crc kubenswrapper[4813]: I1201 09:05:32.252787 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/88d8ade8-c257-464c-9dab-0052bed99037-kolla-config" (OuterVolumeSpecName: "kolla-config") pod "88d8ade8-c257-464c-9dab-0052bed99037" (UID: "88d8ade8-c257-464c-9dab-0052bed99037"). InnerVolumeSpecName "kolla-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 09:05:32 crc kubenswrapper[4813]: I1201 09:05:32.252857 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/88d8ade8-c257-464c-9dab-0052bed99037-config-data" (OuterVolumeSpecName: "config-data") pod "88d8ade8-c257-464c-9dab-0052bed99037" (UID: "88d8ade8-c257-464c-9dab-0052bed99037"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 09:05:32 crc kubenswrapper[4813]: I1201 09:05:32.257725 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/88d8ade8-c257-464c-9dab-0052bed99037-kube-api-access-j5c68" (OuterVolumeSpecName: "kube-api-access-j5c68") pod "88d8ade8-c257-464c-9dab-0052bed99037" (UID: "88d8ade8-c257-464c-9dab-0052bed99037"). InnerVolumeSpecName "kube-api-access-j5c68". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:05:32 crc kubenswrapper[4813]: I1201 09:05:32.335289 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-59d87c8d7c-vcbsn" Dec 01 09:05:32 crc kubenswrapper[4813]: I1201 09:05:32.353565 4813 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/88d8ade8-c257-464c-9dab-0052bed99037-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 09:05:32 crc kubenswrapper[4813]: I1201 09:05:32.353596 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j5c68\" (UniqueName: \"kubernetes.io/projected/88d8ade8-c257-464c-9dab-0052bed99037-kube-api-access-j5c68\") on node \"crc\" DevicePath \"\"" Dec 01 09:05:32 crc kubenswrapper[4813]: I1201 09:05:32.353607 4813 reconciler_common.go:293] "Volume detached for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/88d8ade8-c257-464c-9dab-0052bed99037-kolla-config\") on node \"crc\" DevicePath \"\"" Dec 01 09:05:32 crc kubenswrapper[4813]: I1201 09:05:32.376919 4813 generic.go:334] "Generic (PLEG): container finished" podID="3e52d3b1-2c61-4e2b-8222-35745063e838" containerID="46fb724958d5e9eced4e641498c92ea22e8ac57e844321ca8c097a007224858b" exitCode=0 Dec 01 09:05:32 crc kubenswrapper[4813]: I1201 09:05:32.377015 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-59d87c8d7c-vcbsn" event={"ID":"3e52d3b1-2c61-4e2b-8222-35745063e838","Type":"ContainerDied","Data":"46fb724958d5e9eced4e641498c92ea22e8ac57e844321ca8c097a007224858b"} Dec 01 09:05:32 crc kubenswrapper[4813]: I1201 09:05:32.377053 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-59d87c8d7c-vcbsn" event={"ID":"3e52d3b1-2c61-4e2b-8222-35745063e838","Type":"ContainerDied","Data":"712ef1886d55f92e2e2b344468bfbf42d6fb5ade1f15d3b2e6b3cf9b6eeeb647"} Dec 01 09:05:32 crc kubenswrapper[4813]: I1201 09:05:32.377078 4813 scope.go:117] "RemoveContainer" containerID="46fb724958d5e9eced4e641498c92ea22e8ac57e844321ca8c097a007224858b" Dec 01 09:05:32 crc kubenswrapper[4813]: I1201 09:05:32.377680 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-59d87c8d7c-vcbsn" Dec 01 09:05:32 crc kubenswrapper[4813]: I1201 09:05:32.384908 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone05f1-account-delete-j85h9" Dec 01 09:05:32 crc kubenswrapper[4813]: I1201 09:05:32.384923 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone05f1-account-delete-j85h9" event={"ID":"16ecbca6-953b-4f9f-981e-d9f011703629","Type":"ContainerDied","Data":"6710c7830741be27bd771bfe54743d0cf61017ebc676538798dc5d73146dfc00"} Dec 01 09:05:32 crc kubenswrapper[4813]: I1201 09:05:32.384980 4813 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6710c7830741be27bd771bfe54743d0cf61017ebc676538798dc5d73146dfc00" Dec 01 09:05:32 crc kubenswrapper[4813]: I1201 09:05:32.388453 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/memcached-0" event={"ID":"88d8ade8-c257-464c-9dab-0052bed99037","Type":"ContainerDied","Data":"782914df518ac8b07705916b0c20a4afa91c8fa9818a88f110730f48a8000674"} Dec 01 09:05:32 crc kubenswrapper[4813]: I1201 09:05:32.388546 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/memcached-0" Dec 01 09:05:32 crc kubenswrapper[4813]: I1201 09:05:32.395349 4813 scope.go:117] "RemoveContainer" containerID="09b49517239f156b4ac50c72e4d184fde70e2daaf7b602adcfc3b42ec90a71e0" Dec 01 09:05:32 crc kubenswrapper[4813]: E1201 09:05:32.395623 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9wn4q_openshift-machine-config-operator(3d423402-e08f-42e9-b60c-6eb9bf2f6d48)\"" pod="openshift-machine-config-operator/machine-config-daemon-9wn4q" podUID="3d423402-e08f-42e9-b60c-6eb9bf2f6d48" Dec 01 09:05:32 crc kubenswrapper[4813]: I1201 09:05:32.396172 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/glance-operator-index-vlgfp" podUID="b95238b9-4f2a-4ee8-8b18-44857550a8d1" containerName="registry-server" containerID="cri-o://d5551be8c8531c0a6392a5c6b07fcad3f7680e8c8d36c88e535facac73b021db" gracePeriod=30 Dec 01 09:05:32 crc kubenswrapper[4813]: I1201 09:05:32.411437 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="387b5367-4ea4-421c-98e9-0c71fca3084c" path="/var/lib/kubelet/pods/387b5367-4ea4-421c-98e9-0c71fca3084c/volumes" Dec 01 09:05:32 crc kubenswrapper[4813]: I1201 09:05:32.420113 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="73988763-722d-4652-8fac-053ba8217547" path="/var/lib/kubelet/pods/73988763-722d-4652-8fac-053ba8217547/volumes" Dec 01 09:05:32 crc kubenswrapper[4813]: I1201 09:05:32.421239 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a80abda8-962d-4b0a-bc3a-2e6d693d0393" path="/var/lib/kubelet/pods/a80abda8-962d-4b0a-bc3a-2e6d693d0393/volumes" Dec 01 09:05:32 crc kubenswrapper[4813]: I1201 09:05:32.423107 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a9912b4c-f609-4897-bb27-459294633dc7" path="/var/lib/kubelet/pods/a9912b4c-f609-4897-bb27-459294633dc7/volumes" Dec 01 09:05:32 crc kubenswrapper[4813]: I1201 09:05:32.423646 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f9042460-3e84-44c7-a99d-2939f19d78d2" path="/var/lib/kubelet/pods/f9042460-3e84-44c7-a99d-2939f19d78d2/volumes" Dec 01 09:05:32 crc kubenswrapper[4813]: I1201 09:05:32.437417 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/memcached-0"] Dec 01 09:05:32 crc kubenswrapper[4813]: I1201 09:05:32.441994 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/memcached-0"] Dec 01 09:05:32 crc kubenswrapper[4813]: I1201 09:05:32.447119 4813 scope.go:117] "RemoveContainer" containerID="46fb724958d5e9eced4e641498c92ea22e8ac57e844321ca8c097a007224858b" Dec 01 09:05:32 crc kubenswrapper[4813]: E1201 09:05:32.450143 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"46fb724958d5e9eced4e641498c92ea22e8ac57e844321ca8c097a007224858b\": container with ID starting with 46fb724958d5e9eced4e641498c92ea22e8ac57e844321ca8c097a007224858b not found: ID does not exist" containerID="46fb724958d5e9eced4e641498c92ea22e8ac57e844321ca8c097a007224858b" Dec 01 09:05:32 crc kubenswrapper[4813]: I1201 09:05:32.450195 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"46fb724958d5e9eced4e641498c92ea22e8ac57e844321ca8c097a007224858b"} err="failed to get container status \"46fb724958d5e9eced4e641498c92ea22e8ac57e844321ca8c097a007224858b\": rpc error: code = NotFound desc = could not find container \"46fb724958d5e9eced4e641498c92ea22e8ac57e844321ca8c097a007224858b\": container with ID starting with 46fb724958d5e9eced4e641498c92ea22e8ac57e844321ca8c097a007224858b not found: ID does not exist" Dec 01 09:05:32 crc kubenswrapper[4813]: I1201 09:05:32.450232 4813 scope.go:117] "RemoveContainer" containerID="13647640697f73f4b7f03b0bcbd720969780ef3b11f627085ddd2eb78f70d473" Dec 01 09:05:32 crc kubenswrapper[4813]: I1201 09:05:32.455159 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/keystone05f1-account-delete-j85h9"] Dec 01 09:05:32 crc kubenswrapper[4813]: I1201 09:05:32.456768 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-58dbm\" (UniqueName: \"kubernetes.io/projected/3e52d3b1-2c61-4e2b-8222-35745063e838-kube-api-access-58dbm\") pod \"3e52d3b1-2c61-4e2b-8222-35745063e838\" (UID: \"3e52d3b1-2c61-4e2b-8222-35745063e838\") " Dec 01 09:05:32 crc kubenswrapper[4813]: I1201 09:05:32.456819 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/3e52d3b1-2c61-4e2b-8222-35745063e838-apiservice-cert\") pod \"3e52d3b1-2c61-4e2b-8222-35745063e838\" (UID: \"3e52d3b1-2c61-4e2b-8222-35745063e838\") " Dec 01 09:05:32 crc kubenswrapper[4813]: I1201 09:05:32.456862 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/3e52d3b1-2c61-4e2b-8222-35745063e838-webhook-cert\") pod \"3e52d3b1-2c61-4e2b-8222-35745063e838\" (UID: \"3e52d3b1-2c61-4e2b-8222-35745063e838\") " Dec 01 09:05:32 crc kubenswrapper[4813]: I1201 09:05:32.462197 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/keystone05f1-account-delete-j85h9"] Dec 01 09:05:32 crc kubenswrapper[4813]: I1201 09:05:32.482439 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3e52d3b1-2c61-4e2b-8222-35745063e838-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "3e52d3b1-2c61-4e2b-8222-35745063e838" (UID: "3e52d3b1-2c61-4e2b-8222-35745063e838"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:05:32 crc kubenswrapper[4813]: I1201 09:05:32.493854 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3e52d3b1-2c61-4e2b-8222-35745063e838-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "3e52d3b1-2c61-4e2b-8222-35745063e838" (UID: "3e52d3b1-2c61-4e2b-8222-35745063e838"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:05:32 crc kubenswrapper[4813]: I1201 09:05:32.493973 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3e52d3b1-2c61-4e2b-8222-35745063e838-kube-api-access-58dbm" (OuterVolumeSpecName: "kube-api-access-58dbm") pod "3e52d3b1-2c61-4e2b-8222-35745063e838" (UID: "3e52d3b1-2c61-4e2b-8222-35745063e838"). InnerVolumeSpecName "kube-api-access-58dbm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:05:32 crc kubenswrapper[4813]: I1201 09:05:32.558426 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-58dbm\" (UniqueName: \"kubernetes.io/projected/3e52d3b1-2c61-4e2b-8222-35745063e838-kube-api-access-58dbm\") on node \"crc\" DevicePath \"\"" Dec 01 09:05:32 crc kubenswrapper[4813]: I1201 09:05:32.558473 4813 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/3e52d3b1-2c61-4e2b-8222-35745063e838-apiservice-cert\") on node \"crc\" DevicePath \"\"" Dec 01 09:05:32 crc kubenswrapper[4813]: I1201 09:05:32.558486 4813 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/3e52d3b1-2c61-4e2b-8222-35745063e838-webhook-cert\") on node \"crc\" DevicePath \"\"" Dec 01 09:05:32 crc kubenswrapper[4813]: I1201 09:05:32.705807 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/glance-operator-controller-manager-59d87c8d7c-vcbsn"] Dec 01 09:05:32 crc kubenswrapper[4813]: I1201 09:05:32.710072 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/glance-operator-controller-manager-59d87c8d7c-vcbsn"] Dec 01 09:05:33 crc kubenswrapper[4813]: I1201 09:05:33.134934 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-index-vlgfp" Dec 01 09:05:33 crc kubenswrapper[4813]: I1201 09:05:33.173735 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ftnrl\" (UniqueName: \"kubernetes.io/projected/b95238b9-4f2a-4ee8-8b18-44857550a8d1-kube-api-access-ftnrl\") pod \"b95238b9-4f2a-4ee8-8b18-44857550a8d1\" (UID: \"b95238b9-4f2a-4ee8-8b18-44857550a8d1\") " Dec 01 09:05:33 crc kubenswrapper[4813]: I1201 09:05:33.178610 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b95238b9-4f2a-4ee8-8b18-44857550a8d1-kube-api-access-ftnrl" (OuterVolumeSpecName: "kube-api-access-ftnrl") pod "b95238b9-4f2a-4ee8-8b18-44857550a8d1" (UID: "b95238b9-4f2a-4ee8-8b18-44857550a8d1"). InnerVolumeSpecName "kube-api-access-ftnrl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:05:33 crc kubenswrapper[4813]: I1201 09:05:33.274465 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ftnrl\" (UniqueName: \"kubernetes.io/projected/b95238b9-4f2a-4ee8-8b18-44857550a8d1-kube-api-access-ftnrl\") on node \"crc\" DevicePath \"\"" Dec 01 09:05:33 crc kubenswrapper[4813]: E1201 09:05:33.281193 4813 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda274574f_1dfe_4afd_8dfa_661d14669e73.slice/crio-e9eef734cf66c90ff9f0f04280a9c11fc9ecb34a8b6da3055227e88543656ffe.scope\": RecentStats: unable to find data in memory cache]" Dec 01 09:05:33 crc kubenswrapper[4813]: I1201 09:05:33.409839 4813 generic.go:334] "Generic (PLEG): container finished" podID="a274574f-1dfe-4afd-8dfa-661d14669e73" containerID="e9eef734cf66c90ff9f0f04280a9c11fc9ecb34a8b6da3055227e88543656ffe" exitCode=0 Dec 01 09:05:33 crc kubenswrapper[4813]: I1201 09:05:33.410338 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/rabbitmq-server-0" event={"ID":"a274574f-1dfe-4afd-8dfa-661d14669e73","Type":"ContainerDied","Data":"e9eef734cf66c90ff9f0f04280a9c11fc9ecb34a8b6da3055227e88543656ffe"} Dec 01 09:05:33 crc kubenswrapper[4813]: I1201 09:05:33.413678 4813 generic.go:334] "Generic (PLEG): container finished" podID="b95238b9-4f2a-4ee8-8b18-44857550a8d1" containerID="d5551be8c8531c0a6392a5c6b07fcad3f7680e8c8d36c88e535facac73b021db" exitCode=0 Dec 01 09:05:33 crc kubenswrapper[4813]: I1201 09:05:33.413792 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-index-vlgfp" event={"ID":"b95238b9-4f2a-4ee8-8b18-44857550a8d1","Type":"ContainerDied","Data":"d5551be8c8531c0a6392a5c6b07fcad3f7680e8c8d36c88e535facac73b021db"} Dec 01 09:05:33 crc kubenswrapper[4813]: I1201 09:05:33.413794 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-index-vlgfp" Dec 01 09:05:33 crc kubenswrapper[4813]: I1201 09:05:33.413828 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-index-vlgfp" event={"ID":"b95238b9-4f2a-4ee8-8b18-44857550a8d1","Type":"ContainerDied","Data":"5af7b5b24658c333787f5403d3cccd1a018f85e48e1fb76dff3d2b0a66974019"} Dec 01 09:05:33 crc kubenswrapper[4813]: I1201 09:05:33.413849 4813 scope.go:117] "RemoveContainer" containerID="d5551be8c8531c0a6392a5c6b07fcad3f7680e8c8d36c88e535facac73b021db" Dec 01 09:05:33 crc kubenswrapper[4813]: I1201 09:05:33.417498 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-g4dzh" event={"ID":"07bbfadb-d0a3-478c-a2ed-36ddc6081a49","Type":"ContainerStarted","Data":"a3fbb9a3859ee92e312c0bd6fdf78ffd535f9a6a2df6bcbaad5f44c45e2cb309"} Dec 01 09:05:33 crc kubenswrapper[4813]: I1201 09:05:33.440994 4813 scope.go:117] "RemoveContainer" containerID="d5551be8c8531c0a6392a5c6b07fcad3f7680e8c8d36c88e535facac73b021db" Dec 01 09:05:33 crc kubenswrapper[4813]: E1201 09:05:33.448352 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d5551be8c8531c0a6392a5c6b07fcad3f7680e8c8d36c88e535facac73b021db\": container with ID starting with d5551be8c8531c0a6392a5c6b07fcad3f7680e8c8d36c88e535facac73b021db not found: ID does not exist" containerID="d5551be8c8531c0a6392a5c6b07fcad3f7680e8c8d36c88e535facac73b021db" Dec 01 09:05:33 crc kubenswrapper[4813]: I1201 09:05:33.448415 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d5551be8c8531c0a6392a5c6b07fcad3f7680e8c8d36c88e535facac73b021db"} err="failed to get container status \"d5551be8c8531c0a6392a5c6b07fcad3f7680e8c8d36c88e535facac73b021db\": rpc error: code = NotFound desc = could not find container \"d5551be8c8531c0a6392a5c6b07fcad3f7680e8c8d36c88e535facac73b021db\": container with ID starting with d5551be8c8531c0a6392a5c6b07fcad3f7680e8c8d36c88e535facac73b021db not found: ID does not exist" Dec 01 09:05:33 crc kubenswrapper[4813]: I1201 09:05:33.449422 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-g4dzh" podStartSLOduration=4.642865836 podStartE2EDuration="7.449371028s" podCreationTimestamp="2025-12-01 09:05:26 +0000 UTC" firstStartedPulling="2025-12-01 09:05:29.319094275 +0000 UTC m=+1513.942015861" lastFinishedPulling="2025-12-01 09:05:32.125599467 +0000 UTC m=+1516.748521053" observedRunningTime="2025-12-01 09:05:33.434630909 +0000 UTC m=+1518.057552525" watchObservedRunningTime="2025-12-01 09:05:33.449371028 +0000 UTC m=+1518.072292614" Dec 01 09:05:33 crc kubenswrapper[4813]: I1201 09:05:33.460715 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/glance-operator-index-vlgfp"] Dec 01 09:05:33 crc kubenswrapper[4813]: I1201 09:05:33.467699 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/glance-operator-index-vlgfp"] Dec 01 09:05:33 crc kubenswrapper[4813]: I1201 09:05:33.660356 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/rabbitmq-server-0" Dec 01 09:05:33 crc kubenswrapper[4813]: I1201 09:05:33.774470 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/openstack-galera-0" podUID="43c6449e-e9d4-4a5b-9231-d719d2a9e488" containerName="galera" containerID="cri-o://d5ede1c2b7da645352abfeb35f403960e3939f0491ad66bc5289579290d4c59c" gracePeriod=26 Dec 01 09:05:33 crc kubenswrapper[4813]: I1201 09:05:33.782390 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/a274574f-1dfe-4afd-8dfa-661d14669e73-rabbitmq-erlang-cookie\") pod \"a274574f-1dfe-4afd-8dfa-661d14669e73\" (UID: \"a274574f-1dfe-4afd-8dfa-661d14669e73\") " Dec 01 09:05:33 crc kubenswrapper[4813]: I1201 09:05:33.782472 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/a274574f-1dfe-4afd-8dfa-661d14669e73-erlang-cookie-secret\") pod \"a274574f-1dfe-4afd-8dfa-661d14669e73\" (UID: \"a274574f-1dfe-4afd-8dfa-661d14669e73\") " Dec 01 09:05:33 crc kubenswrapper[4813]: I1201 09:05:33.782501 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/a274574f-1dfe-4afd-8dfa-661d14669e73-pod-info\") pod \"a274574f-1dfe-4afd-8dfa-661d14669e73\" (UID: \"a274574f-1dfe-4afd-8dfa-661d14669e73\") " Dec 01 09:05:33 crc kubenswrapper[4813]: I1201 09:05:33.782560 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/a274574f-1dfe-4afd-8dfa-661d14669e73-rabbitmq-plugins\") pod \"a274574f-1dfe-4afd-8dfa-661d14669e73\" (UID: \"a274574f-1dfe-4afd-8dfa-661d14669e73\") " Dec 01 09:05:33 crc kubenswrapper[4813]: I1201 09:05:33.782579 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-slrjg\" (UniqueName: \"kubernetes.io/projected/a274574f-1dfe-4afd-8dfa-661d14669e73-kube-api-access-slrjg\") pod \"a274574f-1dfe-4afd-8dfa-661d14669e73\" (UID: \"a274574f-1dfe-4afd-8dfa-661d14669e73\") " Dec 01 09:05:33 crc kubenswrapper[4813]: I1201 09:05:33.782713 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-8bac5ff1-dca8-4c26-86da-88e356fbd887\") pod \"a274574f-1dfe-4afd-8dfa-661d14669e73\" (UID: \"a274574f-1dfe-4afd-8dfa-661d14669e73\") " Dec 01 09:05:33 crc kubenswrapper[4813]: I1201 09:05:33.782764 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/a274574f-1dfe-4afd-8dfa-661d14669e73-plugins-conf\") pod \"a274574f-1dfe-4afd-8dfa-661d14669e73\" (UID: \"a274574f-1dfe-4afd-8dfa-661d14669e73\") " Dec 01 09:05:33 crc kubenswrapper[4813]: I1201 09:05:33.782825 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/a274574f-1dfe-4afd-8dfa-661d14669e73-rabbitmq-confd\") pod \"a274574f-1dfe-4afd-8dfa-661d14669e73\" (UID: \"a274574f-1dfe-4afd-8dfa-661d14669e73\") " Dec 01 09:05:33 crc kubenswrapper[4813]: I1201 09:05:33.782937 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a274574f-1dfe-4afd-8dfa-661d14669e73-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "a274574f-1dfe-4afd-8dfa-661d14669e73" (UID: "a274574f-1dfe-4afd-8dfa-661d14669e73"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 09:05:33 crc kubenswrapper[4813]: I1201 09:05:33.783123 4813 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/a274574f-1dfe-4afd-8dfa-661d14669e73-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Dec 01 09:05:33 crc kubenswrapper[4813]: I1201 09:05:33.783106 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a274574f-1dfe-4afd-8dfa-661d14669e73-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "a274574f-1dfe-4afd-8dfa-661d14669e73" (UID: "a274574f-1dfe-4afd-8dfa-661d14669e73"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 09:05:33 crc kubenswrapper[4813]: I1201 09:05:33.783404 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a274574f-1dfe-4afd-8dfa-661d14669e73-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "a274574f-1dfe-4afd-8dfa-661d14669e73" (UID: "a274574f-1dfe-4afd-8dfa-661d14669e73"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 09:05:33 crc kubenswrapper[4813]: I1201 09:05:33.789127 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/a274574f-1dfe-4afd-8dfa-661d14669e73-pod-info" (OuterVolumeSpecName: "pod-info") pod "a274574f-1dfe-4afd-8dfa-661d14669e73" (UID: "a274574f-1dfe-4afd-8dfa-661d14669e73"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Dec 01 09:05:33 crc kubenswrapper[4813]: I1201 09:05:33.803201 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a274574f-1dfe-4afd-8dfa-661d14669e73-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "a274574f-1dfe-4afd-8dfa-661d14669e73" (UID: "a274574f-1dfe-4afd-8dfa-661d14669e73"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:05:33 crc kubenswrapper[4813]: I1201 09:05:33.803277 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a274574f-1dfe-4afd-8dfa-661d14669e73-kube-api-access-slrjg" (OuterVolumeSpecName: "kube-api-access-slrjg") pod "a274574f-1dfe-4afd-8dfa-661d14669e73" (UID: "a274574f-1dfe-4afd-8dfa-661d14669e73"). InnerVolumeSpecName "kube-api-access-slrjg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:05:33 crc kubenswrapper[4813]: I1201 09:05:33.804254 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-8bac5ff1-dca8-4c26-86da-88e356fbd887" (OuterVolumeSpecName: "persistence") pod "a274574f-1dfe-4afd-8dfa-661d14669e73" (UID: "a274574f-1dfe-4afd-8dfa-661d14669e73"). InnerVolumeSpecName "pvc-8bac5ff1-dca8-4c26-86da-88e356fbd887". PluginName "kubernetes.io/csi", VolumeGidValue "" Dec 01 09:05:33 crc kubenswrapper[4813]: I1201 09:05:33.854612 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a274574f-1dfe-4afd-8dfa-661d14669e73-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "a274574f-1dfe-4afd-8dfa-661d14669e73" (UID: "a274574f-1dfe-4afd-8dfa-661d14669e73"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:05:33 crc kubenswrapper[4813]: I1201 09:05:33.884701 4813 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/a274574f-1dfe-4afd-8dfa-661d14669e73-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Dec 01 09:05:33 crc kubenswrapper[4813]: I1201 09:05:33.884802 4813 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-8bac5ff1-dca8-4c26-86da-88e356fbd887\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-8bac5ff1-dca8-4c26-86da-88e356fbd887\") on node \"crc\" " Dec 01 09:05:33 crc kubenswrapper[4813]: I1201 09:05:33.884823 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-slrjg\" (UniqueName: \"kubernetes.io/projected/a274574f-1dfe-4afd-8dfa-661d14669e73-kube-api-access-slrjg\") on node \"crc\" DevicePath \"\"" Dec 01 09:05:33 crc kubenswrapper[4813]: I1201 09:05:33.884837 4813 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/a274574f-1dfe-4afd-8dfa-661d14669e73-plugins-conf\") on node \"crc\" DevicePath \"\"" Dec 01 09:05:33 crc kubenswrapper[4813]: I1201 09:05:33.884848 4813 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/a274574f-1dfe-4afd-8dfa-661d14669e73-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Dec 01 09:05:33 crc kubenswrapper[4813]: I1201 09:05:33.884862 4813 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/a274574f-1dfe-4afd-8dfa-661d14669e73-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Dec 01 09:05:33 crc kubenswrapper[4813]: I1201 09:05:33.884873 4813 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/a274574f-1dfe-4afd-8dfa-661d14669e73-pod-info\") on node \"crc\" DevicePath \"\"" Dec 01 09:05:33 crc kubenswrapper[4813]: I1201 09:05:33.902114 4813 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Dec 01 09:05:33 crc kubenswrapper[4813]: I1201 09:05:33.902495 4813 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-8bac5ff1-dca8-4c26-86da-88e356fbd887" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-8bac5ff1-dca8-4c26-86da-88e356fbd887") on node "crc" Dec 01 09:05:33 crc kubenswrapper[4813]: I1201 09:05:33.986563 4813 reconciler_common.go:293] "Volume detached for volume \"pvc-8bac5ff1-dca8-4c26-86da-88e356fbd887\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-8bac5ff1-dca8-4c26-86da-88e356fbd887\") on node \"crc\" DevicePath \"\"" Dec 01 09:05:34 crc kubenswrapper[4813]: I1201 09:05:34.403228 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="16ecbca6-953b-4f9f-981e-d9f011703629" path="/var/lib/kubelet/pods/16ecbca6-953b-4f9f-981e-d9f011703629/volumes" Dec 01 09:05:34 crc kubenswrapper[4813]: I1201 09:05:34.404055 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3e52d3b1-2c61-4e2b-8222-35745063e838" path="/var/lib/kubelet/pods/3e52d3b1-2c61-4e2b-8222-35745063e838/volumes" Dec 01 09:05:34 crc kubenswrapper[4813]: I1201 09:05:34.404753 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="88d8ade8-c257-464c-9dab-0052bed99037" path="/var/lib/kubelet/pods/88d8ade8-c257-464c-9dab-0052bed99037/volumes" Dec 01 09:05:34 crc kubenswrapper[4813]: I1201 09:05:34.405254 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b95238b9-4f2a-4ee8-8b18-44857550a8d1" path="/var/lib/kubelet/pods/b95238b9-4f2a-4ee8-8b18-44857550a8d1/volumes" Dec 01 09:05:34 crc kubenswrapper[4813]: I1201 09:05:34.427034 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/rabbitmq-server-0" event={"ID":"a274574f-1dfe-4afd-8dfa-661d14669e73","Type":"ContainerDied","Data":"79eec5e6a0c1cdd813c88f2473920813cf0c85add05c31adb8bb9f3829f2f8a4"} Dec 01 09:05:34 crc kubenswrapper[4813]: I1201 09:05:34.427058 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/rabbitmq-server-0" Dec 01 09:05:34 crc kubenswrapper[4813]: I1201 09:05:34.427093 4813 scope.go:117] "RemoveContainer" containerID="e9eef734cf66c90ff9f0f04280a9c11fc9ecb34a8b6da3055227e88543656ffe" Dec 01 09:05:34 crc kubenswrapper[4813]: I1201 09:05:34.446248 4813 scope.go:117] "RemoveContainer" containerID="a70c9b4b1ae94521445b57103c7df87633fe6b94256e1fa57ce08bd340754a02" Dec 01 09:05:34 crc kubenswrapper[4813]: I1201 09:05:34.449718 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/rabbitmq-server-0"] Dec 01 09:05:34 crc kubenswrapper[4813]: I1201 09:05:34.456205 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/rabbitmq-server-0"] Dec 01 09:05:35 crc kubenswrapper[4813]: I1201 09:05:35.429085 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/swift-operator-controller-manager-6b4988f668-x5frt"] Dec 01 09:05:35 crc kubenswrapper[4813]: I1201 09:05:35.429375 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/swift-operator-controller-manager-6b4988f668-x5frt" podUID="abb50b08-d82d-4e6a-af37-5a99d2624339" containerName="manager" containerID="cri-o://f4982383ff10a95aef6eb6240a6c60c5ba0aec1585bd468ca90f8641c3bfad63" gracePeriod=10 Dec 01 09:05:35 crc kubenswrapper[4813]: I1201 09:05:35.454033 4813 generic.go:334] "Generic (PLEG): container finished" podID="41360e5e-2620-4912-8035-0e1a9e0e715d" containerID="f186dd2ac93b4f28668082627ae26d9d2e81b0ec2304484ef52e31414559378e" exitCode=0 Dec 01 09:05:35 crc kubenswrapper[4813]: I1201 09:05:35.454134 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstack-galera-1" event={"ID":"41360e5e-2620-4912-8035-0e1a9e0e715d","Type":"ContainerDied","Data":"f186dd2ac93b4f28668082627ae26d9d2e81b0ec2304484ef52e31414559378e"} Dec 01 09:05:35 crc kubenswrapper[4813]: I1201 09:05:35.464531 4813 generic.go:334] "Generic (PLEG): container finished" podID="43c6449e-e9d4-4a5b-9231-d719d2a9e488" containerID="d5ede1c2b7da645352abfeb35f403960e3939f0491ad66bc5289579290d4c59c" exitCode=0 Dec 01 09:05:35 crc kubenswrapper[4813]: I1201 09:05:35.464634 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstack-galera-0" event={"ID":"43c6449e-e9d4-4a5b-9231-d719d2a9e488","Type":"ContainerDied","Data":"d5ede1c2b7da645352abfeb35f403960e3939f0491ad66bc5289579290d4c59c"} Dec 01 09:05:35 crc kubenswrapper[4813]: I1201 09:05:35.611216 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/openstack-galera-1" Dec 01 09:05:35 crc kubenswrapper[4813]: I1201 09:05:35.616848 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/openstack-galera-0" Dec 01 09:05:35 crc kubenswrapper[4813]: I1201 09:05:35.697107 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/swift-operator-index-vkcnv"] Dec 01 09:05:35 crc kubenswrapper[4813]: I1201 09:05:35.697358 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/swift-operator-index-vkcnv" podUID="174d5e2e-9c14-48e0-af03-91a15c96ad66" containerName="registry-server" containerID="cri-o://f9c5ee37aef3819d0bdc31bfaff6bdcdb7bcbce9deff27949a932cf2c0e39352" gracePeriod=30 Dec 01 09:05:35 crc kubenswrapper[4813]: I1201 09:05:35.711267 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mysql-db\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"41360e5e-2620-4912-8035-0e1a9e0e715d\" (UID: \"41360e5e-2620-4912-8035-0e1a9e0e715d\") " Dec 01 09:05:35 crc kubenswrapper[4813]: I1201 09:05:35.711358 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/41360e5e-2620-4912-8035-0e1a9e0e715d-operator-scripts\") pod \"41360e5e-2620-4912-8035-0e1a9e0e715d\" (UID: \"41360e5e-2620-4912-8035-0e1a9e0e715d\") " Dec 01 09:05:35 crc kubenswrapper[4813]: I1201 09:05:35.711419 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/43c6449e-e9d4-4a5b-9231-d719d2a9e488-kolla-config\") pod \"43c6449e-e9d4-4a5b-9231-d719d2a9e488\" (UID: \"43c6449e-e9d4-4a5b-9231-d719d2a9e488\") " Dec 01 09:05:35 crc kubenswrapper[4813]: I1201 09:05:35.711489 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/41360e5e-2620-4912-8035-0e1a9e0e715d-config-data-default\") pod \"41360e5e-2620-4912-8035-0e1a9e0e715d\" (UID: \"41360e5e-2620-4912-8035-0e1a9e0e715d\") " Dec 01 09:05:35 crc kubenswrapper[4813]: I1201 09:05:35.711545 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/43c6449e-e9d4-4a5b-9231-d719d2a9e488-config-data-default\") pod \"43c6449e-e9d4-4a5b-9231-d719d2a9e488\" (UID: \"43c6449e-e9d4-4a5b-9231-d719d2a9e488\") " Dec 01 09:05:35 crc kubenswrapper[4813]: I1201 09:05:35.711600 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/41360e5e-2620-4912-8035-0e1a9e0e715d-kolla-config\") pod \"41360e5e-2620-4912-8035-0e1a9e0e715d\" (UID: \"41360e5e-2620-4912-8035-0e1a9e0e715d\") " Dec 01 09:05:35 crc kubenswrapper[4813]: I1201 09:05:35.711624 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/43c6449e-e9d4-4a5b-9231-d719d2a9e488-config-data-generated\") pod \"43c6449e-e9d4-4a5b-9231-d719d2a9e488\" (UID: \"43c6449e-e9d4-4a5b-9231-d719d2a9e488\") " Dec 01 09:05:35 crc kubenswrapper[4813]: I1201 09:05:35.711665 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/43c6449e-e9d4-4a5b-9231-d719d2a9e488-operator-scripts\") pod \"43c6449e-e9d4-4a5b-9231-d719d2a9e488\" (UID: \"43c6449e-e9d4-4a5b-9231-d719d2a9e488\") " Dec 01 09:05:35 crc kubenswrapper[4813]: I1201 09:05:35.711693 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ssdv7\" (UniqueName: \"kubernetes.io/projected/41360e5e-2620-4912-8035-0e1a9e0e715d-kube-api-access-ssdv7\") pod \"41360e5e-2620-4912-8035-0e1a9e0e715d\" (UID: \"41360e5e-2620-4912-8035-0e1a9e0e715d\") " Dec 01 09:05:35 crc kubenswrapper[4813]: I1201 09:05:35.711743 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mysql-db\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"43c6449e-e9d4-4a5b-9231-d719d2a9e488\" (UID: \"43c6449e-e9d4-4a5b-9231-d719d2a9e488\") " Dec 01 09:05:35 crc kubenswrapper[4813]: I1201 09:05:35.711787 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d2wgq\" (UniqueName: \"kubernetes.io/projected/43c6449e-e9d4-4a5b-9231-d719d2a9e488-kube-api-access-d2wgq\") pod \"43c6449e-e9d4-4a5b-9231-d719d2a9e488\" (UID: \"43c6449e-e9d4-4a5b-9231-d719d2a9e488\") " Dec 01 09:05:35 crc kubenswrapper[4813]: I1201 09:05:35.711831 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/41360e5e-2620-4912-8035-0e1a9e0e715d-config-data-generated\") pod \"41360e5e-2620-4912-8035-0e1a9e0e715d\" (UID: \"41360e5e-2620-4912-8035-0e1a9e0e715d\") " Dec 01 09:05:35 crc kubenswrapper[4813]: I1201 09:05:35.712864 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/41360e5e-2620-4912-8035-0e1a9e0e715d-config-data-generated" (OuterVolumeSpecName: "config-data-generated") pod "41360e5e-2620-4912-8035-0e1a9e0e715d" (UID: "41360e5e-2620-4912-8035-0e1a9e0e715d"). InnerVolumeSpecName "config-data-generated". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 09:05:35 crc kubenswrapper[4813]: I1201 09:05:35.713207 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/41360e5e-2620-4912-8035-0e1a9e0e715d-kolla-config" (OuterVolumeSpecName: "kolla-config") pod "41360e5e-2620-4912-8035-0e1a9e0e715d" (UID: "41360e5e-2620-4912-8035-0e1a9e0e715d"). InnerVolumeSpecName "kolla-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 09:05:35 crc kubenswrapper[4813]: I1201 09:05:35.713274 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43c6449e-e9d4-4a5b-9231-d719d2a9e488-kolla-config" (OuterVolumeSpecName: "kolla-config") pod "43c6449e-e9d4-4a5b-9231-d719d2a9e488" (UID: "43c6449e-e9d4-4a5b-9231-d719d2a9e488"). InnerVolumeSpecName "kolla-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 09:05:35 crc kubenswrapper[4813]: I1201 09:05:35.713393 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/41360e5e-2620-4912-8035-0e1a9e0e715d-config-data-default" (OuterVolumeSpecName: "config-data-default") pod "41360e5e-2620-4912-8035-0e1a9e0e715d" (UID: "41360e5e-2620-4912-8035-0e1a9e0e715d"). InnerVolumeSpecName "config-data-default". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 09:05:35 crc kubenswrapper[4813]: I1201 09:05:35.713859 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43c6449e-e9d4-4a5b-9231-d719d2a9e488-config-data-default" (OuterVolumeSpecName: "config-data-default") pod "43c6449e-e9d4-4a5b-9231-d719d2a9e488" (UID: "43c6449e-e9d4-4a5b-9231-d719d2a9e488"). InnerVolumeSpecName "config-data-default". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 09:05:35 crc kubenswrapper[4813]: I1201 09:05:35.714008 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43c6449e-e9d4-4a5b-9231-d719d2a9e488-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "43c6449e-e9d4-4a5b-9231-d719d2a9e488" (UID: "43c6449e-e9d4-4a5b-9231-d719d2a9e488"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 09:05:35 crc kubenswrapper[4813]: I1201 09:05:35.714542 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/41360e5e-2620-4912-8035-0e1a9e0e715d-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "41360e5e-2620-4912-8035-0e1a9e0e715d" (UID: "41360e5e-2620-4912-8035-0e1a9e0e715d"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 09:05:35 crc kubenswrapper[4813]: I1201 09:05:35.716678 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/43c6449e-e9d4-4a5b-9231-d719d2a9e488-config-data-generated" (OuterVolumeSpecName: "config-data-generated") pod "43c6449e-e9d4-4a5b-9231-d719d2a9e488" (UID: "43c6449e-e9d4-4a5b-9231-d719d2a9e488"). InnerVolumeSpecName "config-data-generated". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 09:05:35 crc kubenswrapper[4813]: I1201 09:05:35.719243 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43c6449e-e9d4-4a5b-9231-d719d2a9e488-kube-api-access-d2wgq" (OuterVolumeSpecName: "kube-api-access-d2wgq") pod "43c6449e-e9d4-4a5b-9231-d719d2a9e488" (UID: "43c6449e-e9d4-4a5b-9231-d719d2a9e488"). InnerVolumeSpecName "kube-api-access-d2wgq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:05:35 crc kubenswrapper[4813]: I1201 09:05:35.727128 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/41360e5e-2620-4912-8035-0e1a9e0e715d-kube-api-access-ssdv7" (OuterVolumeSpecName: "kube-api-access-ssdv7") pod "41360e5e-2620-4912-8035-0e1a9e0e715d" (UID: "41360e5e-2620-4912-8035-0e1a9e0e715d"). InnerVolumeSpecName "kube-api-access-ssdv7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:05:35 crc kubenswrapper[4813]: I1201 09:05:35.750224 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage06-crc" (OuterVolumeSpecName: "mysql-db") pod "43c6449e-e9d4-4a5b-9231-d719d2a9e488" (UID: "43c6449e-e9d4-4a5b-9231-d719d2a9e488"). InnerVolumeSpecName "local-storage06-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 01 09:05:35 crc kubenswrapper[4813]: I1201 09:05:35.756200 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage09-crc" (OuterVolumeSpecName: "mysql-db") pod "41360e5e-2620-4912-8035-0e1a9e0e715d" (UID: "41360e5e-2620-4912-8035-0e1a9e0e715d"). InnerVolumeSpecName "local-storage09-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 01 09:05:35 crc kubenswrapper[4813]: I1201 09:05:35.756465 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240btqv7m"] Dec 01 09:05:35 crc kubenswrapper[4813]: I1201 09:05:35.767805 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240btqv7m"] Dec 01 09:05:35 crc kubenswrapper[4813]: I1201 09:05:35.813048 4813 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/43c6449e-e9d4-4a5b-9231-d719d2a9e488-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 09:05:35 crc kubenswrapper[4813]: I1201 09:05:35.813081 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ssdv7\" (UniqueName: \"kubernetes.io/projected/41360e5e-2620-4912-8035-0e1a9e0e715d-kube-api-access-ssdv7\") on node \"crc\" DevicePath \"\"" Dec 01 09:05:35 crc kubenswrapper[4813]: I1201 09:05:35.813116 4813 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" " Dec 01 09:05:35 crc kubenswrapper[4813]: I1201 09:05:35.813132 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d2wgq\" (UniqueName: \"kubernetes.io/projected/43c6449e-e9d4-4a5b-9231-d719d2a9e488-kube-api-access-d2wgq\") on node \"crc\" DevicePath \"\"" Dec 01 09:05:35 crc kubenswrapper[4813]: I1201 09:05:35.813144 4813 reconciler_common.go:293] "Volume detached for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/41360e5e-2620-4912-8035-0e1a9e0e715d-config-data-generated\") on node \"crc\" DevicePath \"\"" Dec 01 09:05:35 crc kubenswrapper[4813]: I1201 09:05:35.813161 4813 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" " Dec 01 09:05:35 crc kubenswrapper[4813]: I1201 09:05:35.813187 4813 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/41360e5e-2620-4912-8035-0e1a9e0e715d-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 09:05:35 crc kubenswrapper[4813]: I1201 09:05:35.813202 4813 reconciler_common.go:293] "Volume detached for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/43c6449e-e9d4-4a5b-9231-d719d2a9e488-kolla-config\") on node \"crc\" DevicePath \"\"" Dec 01 09:05:35 crc kubenswrapper[4813]: I1201 09:05:35.813214 4813 reconciler_common.go:293] "Volume detached for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/41360e5e-2620-4912-8035-0e1a9e0e715d-config-data-default\") on node \"crc\" DevicePath \"\"" Dec 01 09:05:35 crc kubenswrapper[4813]: I1201 09:05:35.813225 4813 reconciler_common.go:293] "Volume detached for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/43c6449e-e9d4-4a5b-9231-d719d2a9e488-config-data-default\") on node \"crc\" DevicePath \"\"" Dec 01 09:05:35 crc kubenswrapper[4813]: I1201 09:05:35.813235 4813 reconciler_common.go:293] "Volume detached for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/43c6449e-e9d4-4a5b-9231-d719d2a9e488-config-data-generated\") on node \"crc\" DevicePath \"\"" Dec 01 09:05:35 crc kubenswrapper[4813]: I1201 09:05:35.813245 4813 reconciler_common.go:293] "Volume detached for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/41360e5e-2620-4912-8035-0e1a9e0e715d-kolla-config\") on node \"crc\" DevicePath \"\"" Dec 01 09:05:35 crc kubenswrapper[4813]: I1201 09:05:35.849604 4813 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage06-crc" (UniqueName: "kubernetes.io/local-volume/local-storage06-crc") on node "crc" Dec 01 09:05:35 crc kubenswrapper[4813]: I1201 09:05:35.852274 4813 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage09-crc" (UniqueName: "kubernetes.io/local-volume/local-storage09-crc") on node "crc" Dec 01 09:05:35 crc kubenswrapper[4813]: I1201 09:05:35.914220 4813 reconciler_common.go:293] "Volume detached for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" DevicePath \"\"" Dec 01 09:05:35 crc kubenswrapper[4813]: I1201 09:05:35.914255 4813 reconciler_common.go:293] "Volume detached for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" DevicePath \"\"" Dec 01 09:05:36 crc kubenswrapper[4813]: I1201 09:05:36.409907 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="323401ed-8df8-4341-8d07-ab27d5162fbd" path="/var/lib/kubelet/pods/323401ed-8df8-4341-8d07-ab27d5162fbd/volumes" Dec 01 09:05:36 crc kubenswrapper[4813]: I1201 09:05:36.411418 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a274574f-1dfe-4afd-8dfa-661d14669e73" path="/var/lib/kubelet/pods/a274574f-1dfe-4afd-8dfa-661d14669e73/volumes" Dec 01 09:05:36 crc kubenswrapper[4813]: I1201 09:05:36.457041 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-6b4988f668-x5frt" Dec 01 09:05:36 crc kubenswrapper[4813]: I1201 09:05:36.545716 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstack-galera-0" event={"ID":"43c6449e-e9d4-4a5b-9231-d719d2a9e488","Type":"ContainerDied","Data":"e8a88efbe8dab4460c62a1c604a1e043a164a3ce2d479526c91ba789344981e3"} Dec 01 09:05:36 crc kubenswrapper[4813]: I1201 09:05:36.545771 4813 scope.go:117] "RemoveContainer" containerID="d5ede1c2b7da645352abfeb35f403960e3939f0491ad66bc5289579290d4c59c" Dec 01 09:05:36 crc kubenswrapper[4813]: I1201 09:05:36.546014 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/openstack-galera-0" Dec 01 09:05:36 crc kubenswrapper[4813]: I1201 09:05:36.553029 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstack-galera-1" event={"ID":"41360e5e-2620-4912-8035-0e1a9e0e715d","Type":"ContainerDied","Data":"44694b5fe25334469ed41d3b7c562e53f7524f7f68c06165c5d770d0aa2bd941"} Dec 01 09:05:36 crc kubenswrapper[4813]: I1201 09:05:36.553150 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/openstack-galera-1" Dec 01 09:05:36 crc kubenswrapper[4813]: I1201 09:05:36.570758 4813 generic.go:334] "Generic (PLEG): container finished" podID="abb50b08-d82d-4e6a-af37-5a99d2624339" containerID="f4982383ff10a95aef6eb6240a6c60c5ba0aec1585bd468ca90f8641c3bfad63" exitCode=0 Dec 01 09:05:36 crc kubenswrapper[4813]: I1201 09:05:36.570907 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-6b4988f668-x5frt" event={"ID":"abb50b08-d82d-4e6a-af37-5a99d2624339","Type":"ContainerDied","Data":"f4982383ff10a95aef6eb6240a6c60c5ba0aec1585bd468ca90f8641c3bfad63"} Dec 01 09:05:36 crc kubenswrapper[4813]: I1201 09:05:36.571013 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-6b4988f668-x5frt" Dec 01 09:05:36 crc kubenswrapper[4813]: I1201 09:05:36.571052 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-6b4988f668-x5frt" event={"ID":"abb50b08-d82d-4e6a-af37-5a99d2624339","Type":"ContainerDied","Data":"6247e4da180a31a9b17d5988bcae4b950154328f189ef91fb77c80b4f1023053"} Dec 01 09:05:36 crc kubenswrapper[4813]: I1201 09:05:36.575681 4813 generic.go:334] "Generic (PLEG): container finished" podID="174d5e2e-9c14-48e0-af03-91a15c96ad66" containerID="f9c5ee37aef3819d0bdc31bfaff6bdcdb7bcbce9deff27949a932cf2c0e39352" exitCode=0 Dec 01 09:05:36 crc kubenswrapper[4813]: I1201 09:05:36.575737 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-index-vkcnv" event={"ID":"174d5e2e-9c14-48e0-af03-91a15c96ad66","Type":"ContainerDied","Data":"f9c5ee37aef3819d0bdc31bfaff6bdcdb7bcbce9deff27949a932cf2c0e39352"} Dec 01 09:05:36 crc kubenswrapper[4813]: I1201 09:05:36.600531 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/openstack-galera-0"] Dec 01 09:05:36 crc kubenswrapper[4813]: I1201 09:05:36.605835 4813 scope.go:117] "RemoveContainer" containerID="6731c56ca57e019f8d83a6b260a34487a0f5035a833742a7aaafaaa1b9463302" Dec 01 09:05:36 crc kubenswrapper[4813]: I1201 09:05:36.611881 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/openstack-galera-0"] Dec 01 09:05:36 crc kubenswrapper[4813]: I1201 09:05:36.617735 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/openstack-galera-1"] Dec 01 09:05:36 crc kubenswrapper[4813]: I1201 09:05:36.629709 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/openstack-galera-1"] Dec 01 09:05:36 crc kubenswrapper[4813]: I1201 09:05:36.639095 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6qdh5\" (UniqueName: \"kubernetes.io/projected/abb50b08-d82d-4e6a-af37-5a99d2624339-kube-api-access-6qdh5\") pod \"abb50b08-d82d-4e6a-af37-5a99d2624339\" (UID: \"abb50b08-d82d-4e6a-af37-5a99d2624339\") " Dec 01 09:05:36 crc kubenswrapper[4813]: I1201 09:05:36.639227 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/abb50b08-d82d-4e6a-af37-5a99d2624339-webhook-cert\") pod \"abb50b08-d82d-4e6a-af37-5a99d2624339\" (UID: \"abb50b08-d82d-4e6a-af37-5a99d2624339\") " Dec 01 09:05:36 crc kubenswrapper[4813]: I1201 09:05:36.639255 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/abb50b08-d82d-4e6a-af37-5a99d2624339-apiservice-cert\") pod \"abb50b08-d82d-4e6a-af37-5a99d2624339\" (UID: \"abb50b08-d82d-4e6a-af37-5a99d2624339\") " Dec 01 09:05:36 crc kubenswrapper[4813]: I1201 09:05:36.649927 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/abb50b08-d82d-4e6a-af37-5a99d2624339-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "abb50b08-d82d-4e6a-af37-5a99d2624339" (UID: "abb50b08-d82d-4e6a-af37-5a99d2624339"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:05:36 crc kubenswrapper[4813]: I1201 09:05:36.650542 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/abb50b08-d82d-4e6a-af37-5a99d2624339-kube-api-access-6qdh5" (OuterVolumeSpecName: "kube-api-access-6qdh5") pod "abb50b08-d82d-4e6a-af37-5a99d2624339" (UID: "abb50b08-d82d-4e6a-af37-5a99d2624339"). InnerVolumeSpecName "kube-api-access-6qdh5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:05:36 crc kubenswrapper[4813]: I1201 09:05:36.655896 4813 scope.go:117] "RemoveContainer" containerID="f186dd2ac93b4f28668082627ae26d9d2e81b0ec2304484ef52e31414559378e" Dec 01 09:05:36 crc kubenswrapper[4813]: I1201 09:05:36.660870 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/abb50b08-d82d-4e6a-af37-5a99d2624339-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "abb50b08-d82d-4e6a-af37-5a99d2624339" (UID: "abb50b08-d82d-4e6a-af37-5a99d2624339"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:05:36 crc kubenswrapper[4813]: I1201 09:05:36.711437 4813 scope.go:117] "RemoveContainer" containerID="3a5a49e275f1242a56cc581a872d2ad308bdbc43109cfce09b8356e5e684bc3a" Dec 01 09:05:36 crc kubenswrapper[4813]: I1201 09:05:36.729296 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-index-vkcnv" Dec 01 09:05:36 crc kubenswrapper[4813]: I1201 09:05:36.753725 4813 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/abb50b08-d82d-4e6a-af37-5a99d2624339-webhook-cert\") on node \"crc\" DevicePath \"\"" Dec 01 09:05:36 crc kubenswrapper[4813]: I1201 09:05:36.753763 4813 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/abb50b08-d82d-4e6a-af37-5a99d2624339-apiservice-cert\") on node \"crc\" DevicePath \"\"" Dec 01 09:05:36 crc kubenswrapper[4813]: I1201 09:05:36.753775 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6qdh5\" (UniqueName: \"kubernetes.io/projected/abb50b08-d82d-4e6a-af37-5a99d2624339-kube-api-access-6qdh5\") on node \"crc\" DevicePath \"\"" Dec 01 09:05:36 crc kubenswrapper[4813]: I1201 09:05:36.772769 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-g4dzh" Dec 01 09:05:36 crc kubenswrapper[4813]: I1201 09:05:36.772820 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-g4dzh" Dec 01 09:05:36 crc kubenswrapper[4813]: I1201 09:05:36.778134 4813 scope.go:117] "RemoveContainer" containerID="f4982383ff10a95aef6eb6240a6c60c5ba0aec1585bd468ca90f8641c3bfad63" Dec 01 09:05:36 crc kubenswrapper[4813]: I1201 09:05:36.797985 4813 scope.go:117] "RemoveContainer" containerID="f4982383ff10a95aef6eb6240a6c60c5ba0aec1585bd468ca90f8641c3bfad63" Dec 01 09:05:36 crc kubenswrapper[4813]: E1201 09:05:36.798701 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f4982383ff10a95aef6eb6240a6c60c5ba0aec1585bd468ca90f8641c3bfad63\": container with ID starting with f4982383ff10a95aef6eb6240a6c60c5ba0aec1585bd468ca90f8641c3bfad63 not found: ID does not exist" containerID="f4982383ff10a95aef6eb6240a6c60c5ba0aec1585bd468ca90f8641c3bfad63" Dec 01 09:05:36 crc kubenswrapper[4813]: I1201 09:05:36.798764 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f4982383ff10a95aef6eb6240a6c60c5ba0aec1585bd468ca90f8641c3bfad63"} err="failed to get container status \"f4982383ff10a95aef6eb6240a6c60c5ba0aec1585bd468ca90f8641c3bfad63\": rpc error: code = NotFound desc = could not find container \"f4982383ff10a95aef6eb6240a6c60c5ba0aec1585bd468ca90f8641c3bfad63\": container with ID starting with f4982383ff10a95aef6eb6240a6c60c5ba0aec1585bd468ca90f8641c3bfad63 not found: ID does not exist" Dec 01 09:05:36 crc kubenswrapper[4813]: I1201 09:05:36.819110 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-g4dzh" Dec 01 09:05:36 crc kubenswrapper[4813]: I1201 09:05:36.855100 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zpvwb\" (UniqueName: \"kubernetes.io/projected/174d5e2e-9c14-48e0-af03-91a15c96ad66-kube-api-access-zpvwb\") pod \"174d5e2e-9c14-48e0-af03-91a15c96ad66\" (UID: \"174d5e2e-9c14-48e0-af03-91a15c96ad66\") " Dec 01 09:05:36 crc kubenswrapper[4813]: I1201 09:05:36.859042 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/174d5e2e-9c14-48e0-af03-91a15c96ad66-kube-api-access-zpvwb" (OuterVolumeSpecName: "kube-api-access-zpvwb") pod "174d5e2e-9c14-48e0-af03-91a15c96ad66" (UID: "174d5e2e-9c14-48e0-af03-91a15c96ad66"). InnerVolumeSpecName "kube-api-access-zpvwb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:05:36 crc kubenswrapper[4813]: I1201 09:05:36.904527 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/swift-operator-controller-manager-6b4988f668-x5frt"] Dec 01 09:05:36 crc kubenswrapper[4813]: I1201 09:05:36.910234 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/swift-operator-controller-manager-6b4988f668-x5frt"] Dec 01 09:05:36 crc kubenswrapper[4813]: I1201 09:05:36.956443 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zpvwb\" (UniqueName: \"kubernetes.io/projected/174d5e2e-9c14-48e0-af03-91a15c96ad66-kube-api-access-zpvwb\") on node \"crc\" DevicePath \"\"" Dec 01 09:05:37 crc kubenswrapper[4813]: I1201 09:05:37.702550 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-index-vkcnv" event={"ID":"174d5e2e-9c14-48e0-af03-91a15c96ad66","Type":"ContainerDied","Data":"911e826c96baa80392dfebaeb15f4b76cc73cf02b7afdaca79a2125cc94a17c1"} Dec 01 09:05:37 crc kubenswrapper[4813]: I1201 09:05:37.702880 4813 scope.go:117] "RemoveContainer" containerID="f9c5ee37aef3819d0bdc31bfaff6bdcdb7bcbce9deff27949a932cf2c0e39352" Dec 01 09:05:37 crc kubenswrapper[4813]: I1201 09:05:37.703070 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-index-vkcnv" Dec 01 09:05:37 crc kubenswrapper[4813]: I1201 09:05:37.733571 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/swift-operator-index-vkcnv"] Dec 01 09:05:37 crc kubenswrapper[4813]: I1201 09:05:37.741299 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/swift-operator-index-vkcnv"] Dec 01 09:05:37 crc kubenswrapper[4813]: I1201 09:05:37.766236 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-g4dzh" Dec 01 09:05:38 crc kubenswrapper[4813]: I1201 09:05:38.403152 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="174d5e2e-9c14-48e0-af03-91a15c96ad66" path="/var/lib/kubelet/pods/174d5e2e-9c14-48e0-af03-91a15c96ad66/volumes" Dec 01 09:05:38 crc kubenswrapper[4813]: I1201 09:05:38.404037 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="41360e5e-2620-4912-8035-0e1a9e0e715d" path="/var/lib/kubelet/pods/41360e5e-2620-4912-8035-0e1a9e0e715d/volumes" Dec 01 09:05:38 crc kubenswrapper[4813]: I1201 09:05:38.404768 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43c6449e-e9d4-4a5b-9231-d719d2a9e488" path="/var/lib/kubelet/pods/43c6449e-e9d4-4a5b-9231-d719d2a9e488/volumes" Dec 01 09:05:38 crc kubenswrapper[4813]: I1201 09:05:38.405991 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="abb50b08-d82d-4e6a-af37-5a99d2624339" path="/var/lib/kubelet/pods/abb50b08-d82d-4e6a-af37-5a99d2624339/volumes" Dec 01 09:05:39 crc kubenswrapper[4813]: I1201 09:05:39.294399 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7789cf9588-hv6bg"] Dec 01 09:05:39 crc kubenswrapper[4813]: I1201 09:05:39.294909 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/keystone-operator-controller-manager-7789cf9588-hv6bg" podUID="2a681edc-eda3-4d33-a642-24898d63a97f" containerName="manager" containerID="cri-o://48dbd3ab6f02167f321a00621dc1663966530d13f29261a0304ea32c0e6001c9" gracePeriod=10 Dec 01 09:05:39 crc kubenswrapper[4813]: I1201 09:05:39.549350 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/keystone-operator-index-jvhf5"] Dec 01 09:05:39 crc kubenswrapper[4813]: I1201 09:05:39.549791 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/keystone-operator-index-jvhf5" podUID="6fee8f87-bd86-4153-81b5-fbfc619f609c" containerName="registry-server" containerID="cri-o://7419e96d359c1e84aaf7eb33c841e99f68729bb2e4bbfb0c84284f2139f8559d" gracePeriod=30 Dec 01 09:05:39 crc kubenswrapper[4813]: I1201 09:05:39.590236 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/49c083020ae5dfe237b73a6c6b807501660a323f061d7879268c43a121stbkr"] Dec 01 09:05:39 crc kubenswrapper[4813]: I1201 09:05:39.596249 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/49c083020ae5dfe237b73a6c6b807501660a323f061d7879268c43a121stbkr"] Dec 01 09:05:39 crc kubenswrapper[4813]: I1201 09:05:39.796527 4813 generic.go:334] "Generic (PLEG): container finished" podID="6fee8f87-bd86-4153-81b5-fbfc619f609c" containerID="7419e96d359c1e84aaf7eb33c841e99f68729bb2e4bbfb0c84284f2139f8559d" exitCode=0 Dec 01 09:05:39 crc kubenswrapper[4813]: I1201 09:05:39.796614 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-index-jvhf5" event={"ID":"6fee8f87-bd86-4153-81b5-fbfc619f609c","Type":"ContainerDied","Data":"7419e96d359c1e84aaf7eb33c841e99f68729bb2e4bbfb0c84284f2139f8559d"} Dec 01 09:05:39 crc kubenswrapper[4813]: I1201 09:05:39.798095 4813 generic.go:334] "Generic (PLEG): container finished" podID="2a681edc-eda3-4d33-a642-24898d63a97f" containerID="48dbd3ab6f02167f321a00621dc1663966530d13f29261a0304ea32c0e6001c9" exitCode=0 Dec 01 09:05:39 crc kubenswrapper[4813]: I1201 09:05:39.798144 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7789cf9588-hv6bg" event={"ID":"2a681edc-eda3-4d33-a642-24898d63a97f","Type":"ContainerDied","Data":"48dbd3ab6f02167f321a00621dc1663966530d13f29261a0304ea32c0e6001c9"} Dec 01 09:05:39 crc kubenswrapper[4813]: I1201 09:05:39.798181 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7789cf9588-hv6bg" event={"ID":"2a681edc-eda3-4d33-a642-24898d63a97f","Type":"ContainerDied","Data":"667fe2ff11c27639ba0b7a62d37da55fc2d9e3891451cf302e75f1de107f3a6c"} Dec 01 09:05:39 crc kubenswrapper[4813]: I1201 09:05:39.798209 4813 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="667fe2ff11c27639ba0b7a62d37da55fc2d9e3891451cf302e75f1de107f3a6c" Dec 01 09:05:39 crc kubenswrapper[4813]: I1201 09:05:39.838441 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-7789cf9588-hv6bg" Dec 01 09:05:39 crc kubenswrapper[4813]: I1201 09:05:39.894567 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/2a681edc-eda3-4d33-a642-24898d63a97f-webhook-cert\") pod \"2a681edc-eda3-4d33-a642-24898d63a97f\" (UID: \"2a681edc-eda3-4d33-a642-24898d63a97f\") " Dec 01 09:05:39 crc kubenswrapper[4813]: I1201 09:05:39.894628 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9fwkb\" (UniqueName: \"kubernetes.io/projected/2a681edc-eda3-4d33-a642-24898d63a97f-kube-api-access-9fwkb\") pod \"2a681edc-eda3-4d33-a642-24898d63a97f\" (UID: \"2a681edc-eda3-4d33-a642-24898d63a97f\") " Dec 01 09:05:39 crc kubenswrapper[4813]: I1201 09:05:39.894726 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/2a681edc-eda3-4d33-a642-24898d63a97f-apiservice-cert\") pod \"2a681edc-eda3-4d33-a642-24898d63a97f\" (UID: \"2a681edc-eda3-4d33-a642-24898d63a97f\") " Dec 01 09:05:39 crc kubenswrapper[4813]: I1201 09:05:39.900317 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2a681edc-eda3-4d33-a642-24898d63a97f-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "2a681edc-eda3-4d33-a642-24898d63a97f" (UID: "2a681edc-eda3-4d33-a642-24898d63a97f"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:05:39 crc kubenswrapper[4813]: I1201 09:05:39.904660 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2a681edc-eda3-4d33-a642-24898d63a97f-kube-api-access-9fwkb" (OuterVolumeSpecName: "kube-api-access-9fwkb") pod "2a681edc-eda3-4d33-a642-24898d63a97f" (UID: "2a681edc-eda3-4d33-a642-24898d63a97f"). InnerVolumeSpecName "kube-api-access-9fwkb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:05:39 crc kubenswrapper[4813]: I1201 09:05:39.907153 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2a681edc-eda3-4d33-a642-24898d63a97f-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "2a681edc-eda3-4d33-a642-24898d63a97f" (UID: "2a681edc-eda3-4d33-a642-24898d63a97f"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:05:39 crc kubenswrapper[4813]: I1201 09:05:39.934231 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-index-jvhf5" Dec 01 09:05:40 crc kubenswrapper[4813]: I1201 09:05:39.995919 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hkctl\" (UniqueName: \"kubernetes.io/projected/6fee8f87-bd86-4153-81b5-fbfc619f609c-kube-api-access-hkctl\") pod \"6fee8f87-bd86-4153-81b5-fbfc619f609c\" (UID: \"6fee8f87-bd86-4153-81b5-fbfc619f609c\") " Dec 01 09:05:40 crc kubenswrapper[4813]: I1201 09:05:39.996529 4813 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/2a681edc-eda3-4d33-a642-24898d63a97f-webhook-cert\") on node \"crc\" DevicePath \"\"" Dec 01 09:05:40 crc kubenswrapper[4813]: I1201 09:05:39.996547 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9fwkb\" (UniqueName: \"kubernetes.io/projected/2a681edc-eda3-4d33-a642-24898d63a97f-kube-api-access-9fwkb\") on node \"crc\" DevicePath \"\"" Dec 01 09:05:40 crc kubenswrapper[4813]: I1201 09:05:39.996561 4813 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/2a681edc-eda3-4d33-a642-24898d63a97f-apiservice-cert\") on node \"crc\" DevicePath \"\"" Dec 01 09:05:40 crc kubenswrapper[4813]: I1201 09:05:39.999282 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6fee8f87-bd86-4153-81b5-fbfc619f609c-kube-api-access-hkctl" (OuterVolumeSpecName: "kube-api-access-hkctl") pod "6fee8f87-bd86-4153-81b5-fbfc619f609c" (UID: "6fee8f87-bd86-4153-81b5-fbfc619f609c"). InnerVolumeSpecName "kube-api-access-hkctl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:05:40 crc kubenswrapper[4813]: I1201 09:05:40.097928 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hkctl\" (UniqueName: \"kubernetes.io/projected/6fee8f87-bd86-4153-81b5-fbfc619f609c-kube-api-access-hkctl\") on node \"crc\" DevicePath \"\"" Dec 01 09:05:40 crc kubenswrapper[4813]: I1201 09:05:40.401009 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9f829ff8-1e05-44b9-85f2-570a3c0e6016" path="/var/lib/kubelet/pods/9f829ff8-1e05-44b9-85f2-570a3c0e6016/volumes" Dec 01 09:05:40 crc kubenswrapper[4813]: I1201 09:05:40.404835 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-g4dzh"] Dec 01 09:05:40 crc kubenswrapper[4813]: I1201 09:05:40.807012 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-index-jvhf5" event={"ID":"6fee8f87-bd86-4153-81b5-fbfc619f609c","Type":"ContainerDied","Data":"b65512b2d2468332cfba7275c8f8d3d5165263e2a50fa1296e0b99620592f164"} Dec 01 09:05:40 crc kubenswrapper[4813]: I1201 09:05:40.807076 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-index-jvhf5" Dec 01 09:05:40 crc kubenswrapper[4813]: I1201 09:05:40.807106 4813 scope.go:117] "RemoveContainer" containerID="7419e96d359c1e84aaf7eb33c841e99f68729bb2e4bbfb0c84284f2139f8559d" Dec 01 09:05:40 crc kubenswrapper[4813]: I1201 09:05:40.807180 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-g4dzh" podUID="07bbfadb-d0a3-478c-a2ed-36ddc6081a49" containerName="registry-server" containerID="cri-o://a3fbb9a3859ee92e312c0bd6fdf78ffd535f9a6a2df6bcbaad5f44c45e2cb309" gracePeriod=2 Dec 01 09:05:40 crc kubenswrapper[4813]: I1201 09:05:40.807863 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-7789cf9588-hv6bg" Dec 01 09:05:40 crc kubenswrapper[4813]: I1201 09:05:40.850543 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/keystone-operator-index-jvhf5"] Dec 01 09:05:40 crc kubenswrapper[4813]: I1201 09:05:40.857570 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/keystone-operator-index-jvhf5"] Dec 01 09:05:40 crc kubenswrapper[4813]: I1201 09:05:40.868151 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7789cf9588-hv6bg"] Dec 01 09:05:40 crc kubenswrapper[4813]: I1201 09:05:40.873426 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7789cf9588-hv6bg"] Dec 01 09:05:41 crc kubenswrapper[4813]: I1201 09:05:41.537798 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-779fc9694b-c7nbx"] Dec 01 09:05:41 crc kubenswrapper[4813]: I1201 09:05:41.538352 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-c7nbx" podUID="3820379b-32dc-46ea-a899-0990fadcab5b" containerName="operator" containerID="cri-o://da30da264478e8e370b386bfa87faaa9724dd0a68166bca2ac7eb0b45c18fd0a" gracePeriod=10 Dec 01 09:05:41 crc kubenswrapper[4813]: I1201 09:05:41.784875 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-index-ddpc8"] Dec 01 09:05:41 crc kubenswrapper[4813]: I1201 09:05:41.785166 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/rabbitmq-cluster-operator-index-ddpc8" podUID="3f8cb860-7fc8-490c-b5d9-18eed8d8db79" containerName="registry-server" containerID="cri-o://51afe760f18b00bc5ad4ad6d1aa2d5d888620e8de006174ea8a3610a0f990352" gracePeriod=30 Dec 01 09:05:41 crc kubenswrapper[4813]: I1201 09:05:41.819787 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-g4dzh" Dec 01 09:05:41 crc kubenswrapper[4813]: I1201 09:05:41.834368 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5908vlmf"] Dec 01 09:05:41 crc kubenswrapper[4813]: I1201 09:05:41.839208 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5908vlmf"] Dec 01 09:05:41 crc kubenswrapper[4813]: I1201 09:05:41.847788 4813 generic.go:334] "Generic (PLEG): container finished" podID="07bbfadb-d0a3-478c-a2ed-36ddc6081a49" containerID="a3fbb9a3859ee92e312c0bd6fdf78ffd535f9a6a2df6bcbaad5f44c45e2cb309" exitCode=0 Dec 01 09:05:41 crc kubenswrapper[4813]: I1201 09:05:41.847869 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-g4dzh" event={"ID":"07bbfadb-d0a3-478c-a2ed-36ddc6081a49","Type":"ContainerDied","Data":"a3fbb9a3859ee92e312c0bd6fdf78ffd535f9a6a2df6bcbaad5f44c45e2cb309"} Dec 01 09:05:41 crc kubenswrapper[4813]: I1201 09:05:41.847906 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-g4dzh" event={"ID":"07bbfadb-d0a3-478c-a2ed-36ddc6081a49","Type":"ContainerDied","Data":"5da0e0319711e8d1c40aa687a5875d18ae14d499d0d65c03d94a8e6ab8fc38e3"} Dec 01 09:05:41 crc kubenswrapper[4813]: I1201 09:05:41.847930 4813 scope.go:117] "RemoveContainer" containerID="a3fbb9a3859ee92e312c0bd6fdf78ffd535f9a6a2df6bcbaad5f44c45e2cb309" Dec 01 09:05:41 crc kubenswrapper[4813]: I1201 09:05:41.848090 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-g4dzh" Dec 01 09:05:41 crc kubenswrapper[4813]: I1201 09:05:41.850379 4813 generic.go:334] "Generic (PLEG): container finished" podID="3820379b-32dc-46ea-a899-0990fadcab5b" containerID="da30da264478e8e370b386bfa87faaa9724dd0a68166bca2ac7eb0b45c18fd0a" exitCode=0 Dec 01 09:05:41 crc kubenswrapper[4813]: I1201 09:05:41.850411 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-c7nbx" event={"ID":"3820379b-32dc-46ea-a899-0990fadcab5b","Type":"ContainerDied","Data":"da30da264478e8e370b386bfa87faaa9724dd0a68166bca2ac7eb0b45c18fd0a"} Dec 01 09:05:41 crc kubenswrapper[4813]: I1201 09:05:41.992700 4813 scope.go:117] "RemoveContainer" containerID="b0a66c5eb5f9de05c01d2a0ac255e1682f4a982e3742f4ccd7de893a7165976b" Dec 01 09:05:42 crc kubenswrapper[4813]: I1201 09:05:42.019597 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/07bbfadb-d0a3-478c-a2ed-36ddc6081a49-utilities\") pod \"07bbfadb-d0a3-478c-a2ed-36ddc6081a49\" (UID: \"07bbfadb-d0a3-478c-a2ed-36ddc6081a49\") " Dec 01 09:05:42 crc kubenswrapper[4813]: I1201 09:05:42.019780 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mq4l5\" (UniqueName: \"kubernetes.io/projected/07bbfadb-d0a3-478c-a2ed-36ddc6081a49-kube-api-access-mq4l5\") pod \"07bbfadb-d0a3-478c-a2ed-36ddc6081a49\" (UID: \"07bbfadb-d0a3-478c-a2ed-36ddc6081a49\") " Dec 01 09:05:42 crc kubenswrapper[4813]: I1201 09:05:42.019853 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/07bbfadb-d0a3-478c-a2ed-36ddc6081a49-catalog-content\") pod \"07bbfadb-d0a3-478c-a2ed-36ddc6081a49\" (UID: \"07bbfadb-d0a3-478c-a2ed-36ddc6081a49\") " Dec 01 09:05:42 crc kubenswrapper[4813]: I1201 09:05:42.020705 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/07bbfadb-d0a3-478c-a2ed-36ddc6081a49-utilities" (OuterVolumeSpecName: "utilities") pod "07bbfadb-d0a3-478c-a2ed-36ddc6081a49" (UID: "07bbfadb-d0a3-478c-a2ed-36ddc6081a49"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 09:05:42 crc kubenswrapper[4813]: I1201 09:05:42.027910 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/07bbfadb-d0a3-478c-a2ed-36ddc6081a49-kube-api-access-mq4l5" (OuterVolumeSpecName: "kube-api-access-mq4l5") pod "07bbfadb-d0a3-478c-a2ed-36ddc6081a49" (UID: "07bbfadb-d0a3-478c-a2ed-36ddc6081a49"). InnerVolumeSpecName "kube-api-access-mq4l5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:05:42 crc kubenswrapper[4813]: I1201 09:05:42.033189 4813 scope.go:117] "RemoveContainer" containerID="8933d3fb6c6148d55759aacf3a471c3b08aa2b10859ecd7a6d3421892e4018df" Dec 01 09:05:42 crc kubenswrapper[4813]: I1201 09:05:42.051609 4813 scope.go:117] "RemoveContainer" containerID="a3fbb9a3859ee92e312c0bd6fdf78ffd535f9a6a2df6bcbaad5f44c45e2cb309" Dec 01 09:05:42 crc kubenswrapper[4813]: E1201 09:05:42.055393 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a3fbb9a3859ee92e312c0bd6fdf78ffd535f9a6a2df6bcbaad5f44c45e2cb309\": container with ID starting with a3fbb9a3859ee92e312c0bd6fdf78ffd535f9a6a2df6bcbaad5f44c45e2cb309 not found: ID does not exist" containerID="a3fbb9a3859ee92e312c0bd6fdf78ffd535f9a6a2df6bcbaad5f44c45e2cb309" Dec 01 09:05:42 crc kubenswrapper[4813]: I1201 09:05:42.055455 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a3fbb9a3859ee92e312c0bd6fdf78ffd535f9a6a2df6bcbaad5f44c45e2cb309"} err="failed to get container status \"a3fbb9a3859ee92e312c0bd6fdf78ffd535f9a6a2df6bcbaad5f44c45e2cb309\": rpc error: code = NotFound desc = could not find container \"a3fbb9a3859ee92e312c0bd6fdf78ffd535f9a6a2df6bcbaad5f44c45e2cb309\": container with ID starting with a3fbb9a3859ee92e312c0bd6fdf78ffd535f9a6a2df6bcbaad5f44c45e2cb309 not found: ID does not exist" Dec 01 09:05:42 crc kubenswrapper[4813]: I1201 09:05:42.055493 4813 scope.go:117] "RemoveContainer" containerID="b0a66c5eb5f9de05c01d2a0ac255e1682f4a982e3742f4ccd7de893a7165976b" Dec 01 09:05:42 crc kubenswrapper[4813]: E1201 09:05:42.057362 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b0a66c5eb5f9de05c01d2a0ac255e1682f4a982e3742f4ccd7de893a7165976b\": container with ID starting with b0a66c5eb5f9de05c01d2a0ac255e1682f4a982e3742f4ccd7de893a7165976b not found: ID does not exist" containerID="b0a66c5eb5f9de05c01d2a0ac255e1682f4a982e3742f4ccd7de893a7165976b" Dec 01 09:05:42 crc kubenswrapper[4813]: I1201 09:05:42.057448 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b0a66c5eb5f9de05c01d2a0ac255e1682f4a982e3742f4ccd7de893a7165976b"} err="failed to get container status \"b0a66c5eb5f9de05c01d2a0ac255e1682f4a982e3742f4ccd7de893a7165976b\": rpc error: code = NotFound desc = could not find container \"b0a66c5eb5f9de05c01d2a0ac255e1682f4a982e3742f4ccd7de893a7165976b\": container with ID starting with b0a66c5eb5f9de05c01d2a0ac255e1682f4a982e3742f4ccd7de893a7165976b not found: ID does not exist" Dec 01 09:05:42 crc kubenswrapper[4813]: I1201 09:05:42.057485 4813 scope.go:117] "RemoveContainer" containerID="8933d3fb6c6148d55759aacf3a471c3b08aa2b10859ecd7a6d3421892e4018df" Dec 01 09:05:42 crc kubenswrapper[4813]: E1201 09:05:42.058022 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8933d3fb6c6148d55759aacf3a471c3b08aa2b10859ecd7a6d3421892e4018df\": container with ID starting with 8933d3fb6c6148d55759aacf3a471c3b08aa2b10859ecd7a6d3421892e4018df not found: ID does not exist" containerID="8933d3fb6c6148d55759aacf3a471c3b08aa2b10859ecd7a6d3421892e4018df" Dec 01 09:05:42 crc kubenswrapper[4813]: I1201 09:05:42.058067 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8933d3fb6c6148d55759aacf3a471c3b08aa2b10859ecd7a6d3421892e4018df"} err="failed to get container status \"8933d3fb6c6148d55759aacf3a471c3b08aa2b10859ecd7a6d3421892e4018df\": rpc error: code = NotFound desc = could not find container \"8933d3fb6c6148d55759aacf3a471c3b08aa2b10859ecd7a6d3421892e4018df\": container with ID starting with 8933d3fb6c6148d55759aacf3a471c3b08aa2b10859ecd7a6d3421892e4018df not found: ID does not exist" Dec 01 09:05:42 crc kubenswrapper[4813]: I1201 09:05:42.065758 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-c7nbx" Dec 01 09:05:42 crc kubenswrapper[4813]: I1201 09:05:42.102555 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/07bbfadb-d0a3-478c-a2ed-36ddc6081a49-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "07bbfadb-d0a3-478c-a2ed-36ddc6081a49" (UID: "07bbfadb-d0a3-478c-a2ed-36ddc6081a49"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 09:05:42 crc kubenswrapper[4813]: I1201 09:05:42.122013 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w994g\" (UniqueName: \"kubernetes.io/projected/3820379b-32dc-46ea-a899-0990fadcab5b-kube-api-access-w994g\") pod \"3820379b-32dc-46ea-a899-0990fadcab5b\" (UID: \"3820379b-32dc-46ea-a899-0990fadcab5b\") " Dec 01 09:05:42 crc kubenswrapper[4813]: I1201 09:05:42.123566 4813 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/07bbfadb-d0a3-478c-a2ed-36ddc6081a49-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 09:05:42 crc kubenswrapper[4813]: I1201 09:05:42.123585 4813 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/07bbfadb-d0a3-478c-a2ed-36ddc6081a49-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 09:05:42 crc kubenswrapper[4813]: I1201 09:05:42.123595 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mq4l5\" (UniqueName: \"kubernetes.io/projected/07bbfadb-d0a3-478c-a2ed-36ddc6081a49-kube-api-access-mq4l5\") on node \"crc\" DevicePath \"\"" Dec 01 09:05:42 crc kubenswrapper[4813]: I1201 09:05:42.136249 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3820379b-32dc-46ea-a899-0990fadcab5b-kube-api-access-w994g" (OuterVolumeSpecName: "kube-api-access-w994g") pod "3820379b-32dc-46ea-a899-0990fadcab5b" (UID: "3820379b-32dc-46ea-a899-0990fadcab5b"). InnerVolumeSpecName "kube-api-access-w994g". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:05:42 crc kubenswrapper[4813]: I1201 09:05:42.207412 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-g4dzh"] Dec 01 09:05:42 crc kubenswrapper[4813]: I1201 09:05:42.212102 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-g4dzh"] Dec 01 09:05:42 crc kubenswrapper[4813]: I1201 09:05:42.225676 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w994g\" (UniqueName: \"kubernetes.io/projected/3820379b-32dc-46ea-a899-0990fadcab5b-kube-api-access-w994g\") on node \"crc\" DevicePath \"\"" Dec 01 09:05:42 crc kubenswrapper[4813]: I1201 09:05:42.301774 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-index-ddpc8" Dec 01 09:05:42 crc kubenswrapper[4813]: I1201 09:05:42.401984 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="07bbfadb-d0a3-478c-a2ed-36ddc6081a49" path="/var/lib/kubelet/pods/07bbfadb-d0a3-478c-a2ed-36ddc6081a49/volumes" Dec 01 09:05:42 crc kubenswrapper[4813]: I1201 09:05:42.402603 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2a681edc-eda3-4d33-a642-24898d63a97f" path="/var/lib/kubelet/pods/2a681edc-eda3-4d33-a642-24898d63a97f/volumes" Dec 01 09:05:42 crc kubenswrapper[4813]: I1201 09:05:42.403079 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6fee8f87-bd86-4153-81b5-fbfc619f609c" path="/var/lib/kubelet/pods/6fee8f87-bd86-4153-81b5-fbfc619f609c/volumes" Dec 01 09:05:42 crc kubenswrapper[4813]: I1201 09:05:42.404257 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e1cf13b6-094e-4524-82e7-f165e3b24767" path="/var/lib/kubelet/pods/e1cf13b6-094e-4524-82e7-f165e3b24767/volumes" Dec 01 09:05:42 crc kubenswrapper[4813]: I1201 09:05:42.427725 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z7t4d\" (UniqueName: \"kubernetes.io/projected/3f8cb860-7fc8-490c-b5d9-18eed8d8db79-kube-api-access-z7t4d\") pod \"3f8cb860-7fc8-490c-b5d9-18eed8d8db79\" (UID: \"3f8cb860-7fc8-490c-b5d9-18eed8d8db79\") " Dec 01 09:05:42 crc kubenswrapper[4813]: I1201 09:05:42.431771 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3f8cb860-7fc8-490c-b5d9-18eed8d8db79-kube-api-access-z7t4d" (OuterVolumeSpecName: "kube-api-access-z7t4d") pod "3f8cb860-7fc8-490c-b5d9-18eed8d8db79" (UID: "3f8cb860-7fc8-490c-b5d9-18eed8d8db79"). InnerVolumeSpecName "kube-api-access-z7t4d". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:05:42 crc kubenswrapper[4813]: I1201 09:05:42.529542 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z7t4d\" (UniqueName: \"kubernetes.io/projected/3f8cb860-7fc8-490c-b5d9-18eed8d8db79-kube-api-access-z7t4d\") on node \"crc\" DevicePath \"\"" Dec 01 09:05:42 crc kubenswrapper[4813]: I1201 09:05:42.860934 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-c7nbx" event={"ID":"3820379b-32dc-46ea-a899-0990fadcab5b","Type":"ContainerDied","Data":"b79485e00949fafaeab9ccc86e5e1c800b67c9120cdb041d64a117d83303402c"} Dec 01 09:05:42 crc kubenswrapper[4813]: I1201 09:05:42.861016 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-c7nbx" Dec 01 09:05:42 crc kubenswrapper[4813]: I1201 09:05:42.861080 4813 scope.go:117] "RemoveContainer" containerID="da30da264478e8e370b386bfa87faaa9724dd0a68166bca2ac7eb0b45c18fd0a" Dec 01 09:05:42 crc kubenswrapper[4813]: I1201 09:05:42.863162 4813 generic.go:334] "Generic (PLEG): container finished" podID="3f8cb860-7fc8-490c-b5d9-18eed8d8db79" containerID="51afe760f18b00bc5ad4ad6d1aa2d5d888620e8de006174ea8a3610a0f990352" exitCode=0 Dec 01 09:05:42 crc kubenswrapper[4813]: I1201 09:05:42.863211 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-index-ddpc8" Dec 01 09:05:42 crc kubenswrapper[4813]: I1201 09:05:42.863206 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-index-ddpc8" event={"ID":"3f8cb860-7fc8-490c-b5d9-18eed8d8db79","Type":"ContainerDied","Data":"51afe760f18b00bc5ad4ad6d1aa2d5d888620e8de006174ea8a3610a0f990352"} Dec 01 09:05:42 crc kubenswrapper[4813]: I1201 09:05:42.863571 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-index-ddpc8" event={"ID":"3f8cb860-7fc8-490c-b5d9-18eed8d8db79","Type":"ContainerDied","Data":"f07d2f83e779093422c243c704ab50773bbc6fdcc0958ee65c257ccb2bfe3934"} Dec 01 09:05:42 crc kubenswrapper[4813]: I1201 09:05:42.889259 4813 scope.go:117] "RemoveContainer" containerID="51afe760f18b00bc5ad4ad6d1aa2d5d888620e8de006174ea8a3610a0f990352" Dec 01 09:05:42 crc kubenswrapper[4813]: I1201 09:05:42.900499 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-779fc9694b-c7nbx"] Dec 01 09:05:42 crc kubenswrapper[4813]: I1201 09:05:42.906583 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-779fc9694b-c7nbx"] Dec 01 09:05:42 crc kubenswrapper[4813]: I1201 09:05:42.910653 4813 scope.go:117] "RemoveContainer" containerID="51afe760f18b00bc5ad4ad6d1aa2d5d888620e8de006174ea8a3610a0f990352" Dec 01 09:05:42 crc kubenswrapper[4813]: E1201 09:05:42.912364 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"51afe760f18b00bc5ad4ad6d1aa2d5d888620e8de006174ea8a3610a0f990352\": container with ID starting with 51afe760f18b00bc5ad4ad6d1aa2d5d888620e8de006174ea8a3610a0f990352 not found: ID does not exist" containerID="51afe760f18b00bc5ad4ad6d1aa2d5d888620e8de006174ea8a3610a0f990352" Dec 01 09:05:42 crc kubenswrapper[4813]: I1201 09:05:42.912415 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"51afe760f18b00bc5ad4ad6d1aa2d5d888620e8de006174ea8a3610a0f990352"} err="failed to get container status \"51afe760f18b00bc5ad4ad6d1aa2d5d888620e8de006174ea8a3610a0f990352\": rpc error: code = NotFound desc = could not find container \"51afe760f18b00bc5ad4ad6d1aa2d5d888620e8de006174ea8a3610a0f990352\": container with ID starting with 51afe760f18b00bc5ad4ad6d1aa2d5d888620e8de006174ea8a3610a0f990352 not found: ID does not exist" Dec 01 09:05:42 crc kubenswrapper[4813]: I1201 09:05:42.913806 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-index-ddpc8"] Dec 01 09:05:42 crc kubenswrapper[4813]: I1201 09:05:42.922168 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-index-ddpc8"] Dec 01 09:05:44 crc kubenswrapper[4813]: I1201 09:05:44.406803 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3820379b-32dc-46ea-a899-0990fadcab5b" path="/var/lib/kubelet/pods/3820379b-32dc-46ea-a899-0990fadcab5b/volumes" Dec 01 09:05:44 crc kubenswrapper[4813]: I1201 09:05:44.408840 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3f8cb860-7fc8-490c-b5d9-18eed8d8db79" path="/var/lib/kubelet/pods/3f8cb860-7fc8-490c-b5d9-18eed8d8db79/volumes" Dec 01 09:05:46 crc kubenswrapper[4813]: I1201 09:05:46.347554 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/infra-operator-controller-manager-84d8d7c7c5-bcbsv"] Dec 01 09:05:46 crc kubenswrapper[4813]: I1201 09:05:46.348153 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/infra-operator-controller-manager-84d8d7c7c5-bcbsv" podUID="42e1cee7-cd3a-4194-af37-e29d5f04389a" containerName="manager" containerID="cri-o://01e4000fcaa813ab832d04bb6b4deffd5c3f3fcc49674f0acf6f665517a9cdd0" gracePeriod=10 Dec 01 09:05:46 crc kubenswrapper[4813]: I1201 09:05:46.348566 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/infra-operator-controller-manager-84d8d7c7c5-bcbsv" podUID="42e1cee7-cd3a-4194-af37-e29d5f04389a" containerName="kube-rbac-proxy" containerID="cri-o://ce31b4e09a2acc90072d380007f5b9a42af04d5db29e67e6451cfcba75bdf6e1" gracePeriod=10 Dec 01 09:05:46 crc kubenswrapper[4813]: I1201 09:05:46.591976 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/infra-operator-index-8b5ct"] Dec 01 09:05:46 crc kubenswrapper[4813]: I1201 09:05:46.592211 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/infra-operator-index-8b5ct" podUID="35db6389-d26f-45f7-919a-734a16107934" containerName="registry-server" containerID="cri-o://3750d324ce7a125845adb78d7ce4957cc4f47a0c9eec5a9e81edba3398829947" gracePeriod=30 Dec 01 09:05:46 crc kubenswrapper[4813]: I1201 09:05:46.639761 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5d4rn6x"] Dec 01 09:05:46 crc kubenswrapper[4813]: I1201 09:05:46.643035 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5d4rn6x"] Dec 01 09:05:46 crc kubenswrapper[4813]: I1201 09:05:46.883584 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-84d8d7c7c5-bcbsv" Dec 01 09:05:46 crc kubenswrapper[4813]: I1201 09:05:46.922829 4813 generic.go:334] "Generic (PLEG): container finished" podID="42e1cee7-cd3a-4194-af37-e29d5f04389a" containerID="ce31b4e09a2acc90072d380007f5b9a42af04d5db29e67e6451cfcba75bdf6e1" exitCode=0 Dec 01 09:05:46 crc kubenswrapper[4813]: I1201 09:05:46.922866 4813 generic.go:334] "Generic (PLEG): container finished" podID="42e1cee7-cd3a-4194-af37-e29d5f04389a" containerID="01e4000fcaa813ab832d04bb6b4deffd5c3f3fcc49674f0acf6f665517a9cdd0" exitCode=0 Dec 01 09:05:46 crc kubenswrapper[4813]: I1201 09:05:46.922868 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-84d8d7c7c5-bcbsv" event={"ID":"42e1cee7-cd3a-4194-af37-e29d5f04389a","Type":"ContainerDied","Data":"ce31b4e09a2acc90072d380007f5b9a42af04d5db29e67e6451cfcba75bdf6e1"} Dec 01 09:05:46 crc kubenswrapper[4813]: I1201 09:05:46.922920 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-84d8d7c7c5-bcbsv" event={"ID":"42e1cee7-cd3a-4194-af37-e29d5f04389a","Type":"ContainerDied","Data":"01e4000fcaa813ab832d04bb6b4deffd5c3f3fcc49674f0acf6f665517a9cdd0"} Dec 01 09:05:46 crc kubenswrapper[4813]: I1201 09:05:46.922922 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-84d8d7c7c5-bcbsv" Dec 01 09:05:46 crc kubenswrapper[4813]: I1201 09:05:46.922944 4813 scope.go:117] "RemoveContainer" containerID="ce31b4e09a2acc90072d380007f5b9a42af04d5db29e67e6451cfcba75bdf6e1" Dec 01 09:05:46 crc kubenswrapper[4813]: I1201 09:05:46.922932 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-84d8d7c7c5-bcbsv" event={"ID":"42e1cee7-cd3a-4194-af37-e29d5f04389a","Type":"ContainerDied","Data":"ac474539ad6a4f7cb7237f27e668c11f699497cde5edcd134b113b93c7dae48e"} Dec 01 09:05:46 crc kubenswrapper[4813]: I1201 09:05:46.930344 4813 generic.go:334] "Generic (PLEG): container finished" podID="35db6389-d26f-45f7-919a-734a16107934" containerID="3750d324ce7a125845adb78d7ce4957cc4f47a0c9eec5a9e81edba3398829947" exitCode=0 Dec 01 09:05:46 crc kubenswrapper[4813]: I1201 09:05:46.930388 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-index-8b5ct" event={"ID":"35db6389-d26f-45f7-919a-734a16107934","Type":"ContainerDied","Data":"3750d324ce7a125845adb78d7ce4957cc4f47a0c9eec5a9e81edba3398829947"} Dec 01 09:05:46 crc kubenswrapper[4813]: I1201 09:05:46.958017 4813 scope.go:117] "RemoveContainer" containerID="01e4000fcaa813ab832d04bb6b4deffd5c3f3fcc49674f0acf6f665517a9cdd0" Dec 01 09:05:46 crc kubenswrapper[4813]: I1201 09:05:46.977396 4813 scope.go:117] "RemoveContainer" containerID="ce31b4e09a2acc90072d380007f5b9a42af04d5db29e67e6451cfcba75bdf6e1" Dec 01 09:05:46 crc kubenswrapper[4813]: E1201 09:05:46.978005 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ce31b4e09a2acc90072d380007f5b9a42af04d5db29e67e6451cfcba75bdf6e1\": container with ID starting with ce31b4e09a2acc90072d380007f5b9a42af04d5db29e67e6451cfcba75bdf6e1 not found: ID does not exist" containerID="ce31b4e09a2acc90072d380007f5b9a42af04d5db29e67e6451cfcba75bdf6e1" Dec 01 09:05:46 crc kubenswrapper[4813]: I1201 09:05:46.978037 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ce31b4e09a2acc90072d380007f5b9a42af04d5db29e67e6451cfcba75bdf6e1"} err="failed to get container status \"ce31b4e09a2acc90072d380007f5b9a42af04d5db29e67e6451cfcba75bdf6e1\": rpc error: code = NotFound desc = could not find container \"ce31b4e09a2acc90072d380007f5b9a42af04d5db29e67e6451cfcba75bdf6e1\": container with ID starting with ce31b4e09a2acc90072d380007f5b9a42af04d5db29e67e6451cfcba75bdf6e1 not found: ID does not exist" Dec 01 09:05:46 crc kubenswrapper[4813]: I1201 09:05:46.978061 4813 scope.go:117] "RemoveContainer" containerID="01e4000fcaa813ab832d04bb6b4deffd5c3f3fcc49674f0acf6f665517a9cdd0" Dec 01 09:05:46 crc kubenswrapper[4813]: E1201 09:05:46.978613 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"01e4000fcaa813ab832d04bb6b4deffd5c3f3fcc49674f0acf6f665517a9cdd0\": container with ID starting with 01e4000fcaa813ab832d04bb6b4deffd5c3f3fcc49674f0acf6f665517a9cdd0 not found: ID does not exist" containerID="01e4000fcaa813ab832d04bb6b4deffd5c3f3fcc49674f0acf6f665517a9cdd0" Dec 01 09:05:46 crc kubenswrapper[4813]: I1201 09:05:46.978672 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"01e4000fcaa813ab832d04bb6b4deffd5c3f3fcc49674f0acf6f665517a9cdd0"} err="failed to get container status \"01e4000fcaa813ab832d04bb6b4deffd5c3f3fcc49674f0acf6f665517a9cdd0\": rpc error: code = NotFound desc = could not find container \"01e4000fcaa813ab832d04bb6b4deffd5c3f3fcc49674f0acf6f665517a9cdd0\": container with ID starting with 01e4000fcaa813ab832d04bb6b4deffd5c3f3fcc49674f0acf6f665517a9cdd0 not found: ID does not exist" Dec 01 09:05:46 crc kubenswrapper[4813]: I1201 09:05:46.978714 4813 scope.go:117] "RemoveContainer" containerID="ce31b4e09a2acc90072d380007f5b9a42af04d5db29e67e6451cfcba75bdf6e1" Dec 01 09:05:46 crc kubenswrapper[4813]: I1201 09:05:46.979066 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ce31b4e09a2acc90072d380007f5b9a42af04d5db29e67e6451cfcba75bdf6e1"} err="failed to get container status \"ce31b4e09a2acc90072d380007f5b9a42af04d5db29e67e6451cfcba75bdf6e1\": rpc error: code = NotFound desc = could not find container \"ce31b4e09a2acc90072d380007f5b9a42af04d5db29e67e6451cfcba75bdf6e1\": container with ID starting with ce31b4e09a2acc90072d380007f5b9a42af04d5db29e67e6451cfcba75bdf6e1 not found: ID does not exist" Dec 01 09:05:46 crc kubenswrapper[4813]: I1201 09:05:46.979110 4813 scope.go:117] "RemoveContainer" containerID="01e4000fcaa813ab832d04bb6b4deffd5c3f3fcc49674f0acf6f665517a9cdd0" Dec 01 09:05:46 crc kubenswrapper[4813]: I1201 09:05:46.979477 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"01e4000fcaa813ab832d04bb6b4deffd5c3f3fcc49674f0acf6f665517a9cdd0"} err="failed to get container status \"01e4000fcaa813ab832d04bb6b4deffd5c3f3fcc49674f0acf6f665517a9cdd0\": rpc error: code = NotFound desc = could not find container \"01e4000fcaa813ab832d04bb6b4deffd5c3f3fcc49674f0acf6f665517a9cdd0\": container with ID starting with 01e4000fcaa813ab832d04bb6b4deffd5c3f3fcc49674f0acf6f665517a9cdd0 not found: ID does not exist" Dec 01 09:05:47 crc kubenswrapper[4813]: I1201 09:05:47.001862 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-index-8b5ct" Dec 01 09:05:47 crc kubenswrapper[4813]: I1201 09:05:47.038279 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/42e1cee7-cd3a-4194-af37-e29d5f04389a-webhook-cert\") pod \"42e1cee7-cd3a-4194-af37-e29d5f04389a\" (UID: \"42e1cee7-cd3a-4194-af37-e29d5f04389a\") " Dec 01 09:05:47 crc kubenswrapper[4813]: I1201 09:05:47.038416 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-csgbg\" (UniqueName: \"kubernetes.io/projected/42e1cee7-cd3a-4194-af37-e29d5f04389a-kube-api-access-csgbg\") pod \"42e1cee7-cd3a-4194-af37-e29d5f04389a\" (UID: \"42e1cee7-cd3a-4194-af37-e29d5f04389a\") " Dec 01 09:05:47 crc kubenswrapper[4813]: I1201 09:05:47.038439 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/42e1cee7-cd3a-4194-af37-e29d5f04389a-apiservice-cert\") pod \"42e1cee7-cd3a-4194-af37-e29d5f04389a\" (UID: \"42e1cee7-cd3a-4194-af37-e29d5f04389a\") " Dec 01 09:05:47 crc kubenswrapper[4813]: I1201 09:05:47.043719 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/42e1cee7-cd3a-4194-af37-e29d5f04389a-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "42e1cee7-cd3a-4194-af37-e29d5f04389a" (UID: "42e1cee7-cd3a-4194-af37-e29d5f04389a"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:05:47 crc kubenswrapper[4813]: I1201 09:05:47.043763 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/42e1cee7-cd3a-4194-af37-e29d5f04389a-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "42e1cee7-cd3a-4194-af37-e29d5f04389a" (UID: "42e1cee7-cd3a-4194-af37-e29d5f04389a"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:05:47 crc kubenswrapper[4813]: I1201 09:05:47.043879 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/42e1cee7-cd3a-4194-af37-e29d5f04389a-kube-api-access-csgbg" (OuterVolumeSpecName: "kube-api-access-csgbg") pod "42e1cee7-cd3a-4194-af37-e29d5f04389a" (UID: "42e1cee7-cd3a-4194-af37-e29d5f04389a"). InnerVolumeSpecName "kube-api-access-csgbg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:05:47 crc kubenswrapper[4813]: I1201 09:05:47.139536 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m92t7\" (UniqueName: \"kubernetes.io/projected/35db6389-d26f-45f7-919a-734a16107934-kube-api-access-m92t7\") pod \"35db6389-d26f-45f7-919a-734a16107934\" (UID: \"35db6389-d26f-45f7-919a-734a16107934\") " Dec 01 09:05:47 crc kubenswrapper[4813]: I1201 09:05:47.139918 4813 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/42e1cee7-cd3a-4194-af37-e29d5f04389a-webhook-cert\") on node \"crc\" DevicePath \"\"" Dec 01 09:05:47 crc kubenswrapper[4813]: I1201 09:05:47.139936 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-csgbg\" (UniqueName: \"kubernetes.io/projected/42e1cee7-cd3a-4194-af37-e29d5f04389a-kube-api-access-csgbg\") on node \"crc\" DevicePath \"\"" Dec 01 09:05:47 crc kubenswrapper[4813]: I1201 09:05:47.139947 4813 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/42e1cee7-cd3a-4194-af37-e29d5f04389a-apiservice-cert\") on node \"crc\" DevicePath \"\"" Dec 01 09:05:47 crc kubenswrapper[4813]: I1201 09:05:47.142441 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/35db6389-d26f-45f7-919a-734a16107934-kube-api-access-m92t7" (OuterVolumeSpecName: "kube-api-access-m92t7") pod "35db6389-d26f-45f7-919a-734a16107934" (UID: "35db6389-d26f-45f7-919a-734a16107934"). InnerVolumeSpecName "kube-api-access-m92t7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:05:47 crc kubenswrapper[4813]: I1201 09:05:47.241846 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m92t7\" (UniqueName: \"kubernetes.io/projected/35db6389-d26f-45f7-919a-734a16107934-kube-api-access-m92t7\") on node \"crc\" DevicePath \"\"" Dec 01 09:05:47 crc kubenswrapper[4813]: I1201 09:05:47.261310 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/infra-operator-controller-manager-84d8d7c7c5-bcbsv"] Dec 01 09:05:47 crc kubenswrapper[4813]: I1201 09:05:47.267366 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/infra-operator-controller-manager-84d8d7c7c5-bcbsv"] Dec 01 09:05:47 crc kubenswrapper[4813]: I1201 09:05:47.394114 4813 scope.go:117] "RemoveContainer" containerID="09b49517239f156b4ac50c72e4d184fde70e2daaf7b602adcfc3b42ec90a71e0" Dec 01 09:05:47 crc kubenswrapper[4813]: E1201 09:05:47.394474 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9wn4q_openshift-machine-config-operator(3d423402-e08f-42e9-b60c-6eb9bf2f6d48)\"" pod="openshift-machine-config-operator/machine-config-daemon-9wn4q" podUID="3d423402-e08f-42e9-b60c-6eb9bf2f6d48" Dec 01 09:05:47 crc kubenswrapper[4813]: I1201 09:05:47.941790 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-index-8b5ct" event={"ID":"35db6389-d26f-45f7-919a-734a16107934","Type":"ContainerDied","Data":"5e8a1fb5f1aa201194f382989c965d75e3eb21fece89babc4e0a7fcf5f58d2fd"} Dec 01 09:05:47 crc kubenswrapper[4813]: I1201 09:05:47.942045 4813 scope.go:117] "RemoveContainer" containerID="3750d324ce7a125845adb78d7ce4957cc4f47a0c9eec5a9e81edba3398829947" Dec 01 09:05:47 crc kubenswrapper[4813]: I1201 09:05:47.941884 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-index-8b5ct" Dec 01 09:05:47 crc kubenswrapper[4813]: I1201 09:05:47.975686 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/infra-operator-index-8b5ct"] Dec 01 09:05:47 crc kubenswrapper[4813]: I1201 09:05:47.979891 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/infra-operator-index-8b5ct"] Dec 01 09:05:48 crc kubenswrapper[4813]: I1201 09:05:48.402798 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="35db6389-d26f-45f7-919a-734a16107934" path="/var/lib/kubelet/pods/35db6389-d26f-45f7-919a-734a16107934/volumes" Dec 01 09:05:48 crc kubenswrapper[4813]: I1201 09:05:48.403640 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="42e1cee7-cd3a-4194-af37-e29d5f04389a" path="/var/lib/kubelet/pods/42e1cee7-cd3a-4194-af37-e29d5f04389a/volumes" Dec 01 09:05:48 crc kubenswrapper[4813]: I1201 09:05:48.409215 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="db90917b-21c8-43d8-99c5-eb77590da02f" path="/var/lib/kubelet/pods/db90917b-21c8-43d8-99c5-eb77590da02f/volumes" Dec 01 09:05:48 crc kubenswrapper[4813]: I1201 09:05:48.422617 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-679c7f6ffd-mfx9v"] Dec 01 09:05:48 crc kubenswrapper[4813]: I1201 09:05:48.423405 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/mariadb-operator-controller-manager-679c7f6ffd-mfx9v" podUID="84350e15-9af9-4de1-bf0b-9e75c91b180a" containerName="manager" containerID="cri-o://4cbeee28f90ffdf966d5d0157b2f7373e15ff1820f16530504a35e976c1128d2" gracePeriod=10 Dec 01 09:05:48 crc kubenswrapper[4813]: I1201 09:05:48.578877 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/mariadb-operator-index-z8nm9"] Dec 01 09:05:48 crc kubenswrapper[4813]: I1201 09:05:48.579287 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/mariadb-operator-index-z8nm9" podUID="c4d16fd1-e8d3-4bb5-af1f-204e484c8530" containerName="registry-server" containerID="cri-o://c7d761c15931bdb12ada0b86c51712ab26b21bf2c891adcc8553e58602301b2f" gracePeriod=30 Dec 01 09:05:48 crc kubenswrapper[4813]: I1201 09:05:48.612456 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534fxznfm"] Dec 01 09:05:48 crc kubenswrapper[4813]: I1201 09:05:48.617890 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534fxznfm"] Dec 01 09:05:48 crc kubenswrapper[4813]: I1201 09:05:48.959189 4813 generic.go:334] "Generic (PLEG): container finished" podID="84350e15-9af9-4de1-bf0b-9e75c91b180a" containerID="4cbeee28f90ffdf966d5d0157b2f7373e15ff1820f16530504a35e976c1128d2" exitCode=0 Dec 01 09:05:48 crc kubenswrapper[4813]: I1201 09:05:48.959255 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-679c7f6ffd-mfx9v" event={"ID":"84350e15-9af9-4de1-bf0b-9e75c91b180a","Type":"ContainerDied","Data":"4cbeee28f90ffdf966d5d0157b2f7373e15ff1820f16530504a35e976c1128d2"} Dec 01 09:05:48 crc kubenswrapper[4813]: I1201 09:05:48.959284 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-679c7f6ffd-mfx9v" event={"ID":"84350e15-9af9-4de1-bf0b-9e75c91b180a","Type":"ContainerDied","Data":"a7271f59525fb78749cb368b151efb83eb5516aefb5053e5b02244a6bad89245"} Dec 01 09:05:48 crc kubenswrapper[4813]: I1201 09:05:48.959296 4813 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a7271f59525fb78749cb368b151efb83eb5516aefb5053e5b02244a6bad89245" Dec 01 09:05:48 crc kubenswrapper[4813]: I1201 09:05:48.964706 4813 generic.go:334] "Generic (PLEG): container finished" podID="c4d16fd1-e8d3-4bb5-af1f-204e484c8530" containerID="c7d761c15931bdb12ada0b86c51712ab26b21bf2c891adcc8553e58602301b2f" exitCode=0 Dec 01 09:05:48 crc kubenswrapper[4813]: I1201 09:05:48.964758 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-index-z8nm9" event={"ID":"c4d16fd1-e8d3-4bb5-af1f-204e484c8530","Type":"ContainerDied","Data":"c7d761c15931bdb12ada0b86c51712ab26b21bf2c891adcc8553e58602301b2f"} Dec 01 09:05:48 crc kubenswrapper[4813]: I1201 09:05:48.966321 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-679c7f6ffd-mfx9v" Dec 01 09:05:49 crc kubenswrapper[4813]: I1201 09:05:49.056192 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-index-z8nm9" Dec 01 09:05:49 crc kubenswrapper[4813]: I1201 09:05:49.065529 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hsfrr\" (UniqueName: \"kubernetes.io/projected/84350e15-9af9-4de1-bf0b-9e75c91b180a-kube-api-access-hsfrr\") pod \"84350e15-9af9-4de1-bf0b-9e75c91b180a\" (UID: \"84350e15-9af9-4de1-bf0b-9e75c91b180a\") " Dec 01 09:05:49 crc kubenswrapper[4813]: I1201 09:05:49.065607 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/84350e15-9af9-4de1-bf0b-9e75c91b180a-webhook-cert\") pod \"84350e15-9af9-4de1-bf0b-9e75c91b180a\" (UID: \"84350e15-9af9-4de1-bf0b-9e75c91b180a\") " Dec 01 09:05:49 crc kubenswrapper[4813]: I1201 09:05:49.065654 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/84350e15-9af9-4de1-bf0b-9e75c91b180a-apiservice-cert\") pod \"84350e15-9af9-4de1-bf0b-9e75c91b180a\" (UID: \"84350e15-9af9-4de1-bf0b-9e75c91b180a\") " Dec 01 09:05:49 crc kubenswrapper[4813]: I1201 09:05:49.071590 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/84350e15-9af9-4de1-bf0b-9e75c91b180a-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "84350e15-9af9-4de1-bf0b-9e75c91b180a" (UID: "84350e15-9af9-4de1-bf0b-9e75c91b180a"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:05:49 crc kubenswrapper[4813]: I1201 09:05:49.073342 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/84350e15-9af9-4de1-bf0b-9e75c91b180a-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "84350e15-9af9-4de1-bf0b-9e75c91b180a" (UID: "84350e15-9af9-4de1-bf0b-9e75c91b180a"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:05:49 crc kubenswrapper[4813]: I1201 09:05:49.073548 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/84350e15-9af9-4de1-bf0b-9e75c91b180a-kube-api-access-hsfrr" (OuterVolumeSpecName: "kube-api-access-hsfrr") pod "84350e15-9af9-4de1-bf0b-9e75c91b180a" (UID: "84350e15-9af9-4de1-bf0b-9e75c91b180a"). InnerVolumeSpecName "kube-api-access-hsfrr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:05:49 crc kubenswrapper[4813]: I1201 09:05:49.166929 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wlhbf\" (UniqueName: \"kubernetes.io/projected/c4d16fd1-e8d3-4bb5-af1f-204e484c8530-kube-api-access-wlhbf\") pod \"c4d16fd1-e8d3-4bb5-af1f-204e484c8530\" (UID: \"c4d16fd1-e8d3-4bb5-af1f-204e484c8530\") " Dec 01 09:05:49 crc kubenswrapper[4813]: I1201 09:05:49.167252 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hsfrr\" (UniqueName: \"kubernetes.io/projected/84350e15-9af9-4de1-bf0b-9e75c91b180a-kube-api-access-hsfrr\") on node \"crc\" DevicePath \"\"" Dec 01 09:05:49 crc kubenswrapper[4813]: I1201 09:05:49.167272 4813 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/84350e15-9af9-4de1-bf0b-9e75c91b180a-webhook-cert\") on node \"crc\" DevicePath \"\"" Dec 01 09:05:49 crc kubenswrapper[4813]: I1201 09:05:49.167287 4813 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/84350e15-9af9-4de1-bf0b-9e75c91b180a-apiservice-cert\") on node \"crc\" DevicePath \"\"" Dec 01 09:05:49 crc kubenswrapper[4813]: I1201 09:05:49.171424 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c4d16fd1-e8d3-4bb5-af1f-204e484c8530-kube-api-access-wlhbf" (OuterVolumeSpecName: "kube-api-access-wlhbf") pod "c4d16fd1-e8d3-4bb5-af1f-204e484c8530" (UID: "c4d16fd1-e8d3-4bb5-af1f-204e484c8530"). InnerVolumeSpecName "kube-api-access-wlhbf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:05:49 crc kubenswrapper[4813]: I1201 09:05:49.269251 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wlhbf\" (UniqueName: \"kubernetes.io/projected/c4d16fd1-e8d3-4bb5-af1f-204e484c8530-kube-api-access-wlhbf\") on node \"crc\" DevicePath \"\"" Dec 01 09:05:49 crc kubenswrapper[4813]: I1201 09:05:49.984041 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-index-z8nm9" event={"ID":"c4d16fd1-e8d3-4bb5-af1f-204e484c8530","Type":"ContainerDied","Data":"d96636e0a20ffd8ed70847e315e9ff215bb3c73fc947fae546d00aee27a50e41"} Dec 01 09:05:49 crc kubenswrapper[4813]: I1201 09:05:49.984104 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-679c7f6ffd-mfx9v" Dec 01 09:05:49 crc kubenswrapper[4813]: I1201 09:05:49.984130 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-index-z8nm9" Dec 01 09:05:49 crc kubenswrapper[4813]: I1201 09:05:49.984158 4813 scope.go:117] "RemoveContainer" containerID="c7d761c15931bdb12ada0b86c51712ab26b21bf2c891adcc8553e58602301b2f" Dec 01 09:05:50 crc kubenswrapper[4813]: I1201 09:05:50.026782 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-679c7f6ffd-mfx9v"] Dec 01 09:05:50 crc kubenswrapper[4813]: I1201 09:05:50.032738 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-679c7f6ffd-mfx9v"] Dec 01 09:05:50 crc kubenswrapper[4813]: I1201 09:05:50.039348 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/mariadb-operator-index-z8nm9"] Dec 01 09:05:50 crc kubenswrapper[4813]: I1201 09:05:50.043144 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/mariadb-operator-index-z8nm9"] Dec 01 09:05:50 crc kubenswrapper[4813]: I1201 09:05:50.402748 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="348e5b65-5859-4870-aaa1-85100418c446" path="/var/lib/kubelet/pods/348e5b65-5859-4870-aaa1-85100418c446/volumes" Dec 01 09:05:50 crc kubenswrapper[4813]: I1201 09:05:50.403619 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="84350e15-9af9-4de1-bf0b-9e75c91b180a" path="/var/lib/kubelet/pods/84350e15-9af9-4de1-bf0b-9e75c91b180a/volumes" Dec 01 09:05:50 crc kubenswrapper[4813]: I1201 09:05:50.404308 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c4d16fd1-e8d3-4bb5-af1f-204e484c8530" path="/var/lib/kubelet/pods/c4d16fd1-e8d3-4bb5-af1f-204e484c8530/volumes" Dec 01 09:05:55 crc kubenswrapper[4813]: I1201 09:05:55.039750 4813 generic.go:334] "Generic (PLEG): container finished" podID="995100a7-4c5a-409a-81f2-9dbaf2fa1d47" containerID="08b235b36c9b32c3a32345cc647c1bd488a16d47403f387ba92c8337ae7120e1" exitCode=137 Dec 01 09:05:55 crc kubenswrapper[4813]: I1201 09:05:55.040272 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"995100a7-4c5a-409a-81f2-9dbaf2fa1d47","Type":"ContainerDied","Data":"08b235b36c9b32c3a32345cc647c1bd488a16d47403f387ba92c8337ae7120e1"} Dec 01 09:05:55 crc kubenswrapper[4813]: I1201 09:05:55.327299 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/swift-storage-0" Dec 01 09:05:55 crc kubenswrapper[4813]: I1201 09:05:55.434745 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/995100a7-4c5a-409a-81f2-9dbaf2fa1d47-cache\") pod \"995100a7-4c5a-409a-81f2-9dbaf2fa1d47\" (UID: \"995100a7-4c5a-409a-81f2-9dbaf2fa1d47\") " Dec 01 09:05:55 crc kubenswrapper[4813]: I1201 09:05:55.434830 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/995100a7-4c5a-409a-81f2-9dbaf2fa1d47-etc-swift\") pod \"995100a7-4c5a-409a-81f2-9dbaf2fa1d47\" (UID: \"995100a7-4c5a-409a-81f2-9dbaf2fa1d47\") " Dec 01 09:05:55 crc kubenswrapper[4813]: I1201 09:05:55.434955 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/995100a7-4c5a-409a-81f2-9dbaf2fa1d47-lock\") pod \"995100a7-4c5a-409a-81f2-9dbaf2fa1d47\" (UID: \"995100a7-4c5a-409a-81f2-9dbaf2fa1d47\") " Dec 01 09:05:55 crc kubenswrapper[4813]: I1201 09:05:55.435659 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/995100a7-4c5a-409a-81f2-9dbaf2fa1d47-cache" (OuterVolumeSpecName: "cache") pod "995100a7-4c5a-409a-81f2-9dbaf2fa1d47" (UID: "995100a7-4c5a-409a-81f2-9dbaf2fa1d47"). InnerVolumeSpecName "cache". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 09:05:55 crc kubenswrapper[4813]: I1201 09:05:55.435688 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/995100a7-4c5a-409a-81f2-9dbaf2fa1d47-lock" (OuterVolumeSpecName: "lock") pod "995100a7-4c5a-409a-81f2-9dbaf2fa1d47" (UID: "995100a7-4c5a-409a-81f2-9dbaf2fa1d47"). InnerVolumeSpecName "lock". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 09:05:55 crc kubenswrapper[4813]: I1201 09:05:55.435925 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6jgvz\" (UniqueName: \"kubernetes.io/projected/995100a7-4c5a-409a-81f2-9dbaf2fa1d47-kube-api-access-6jgvz\") pod \"995100a7-4c5a-409a-81f2-9dbaf2fa1d47\" (UID: \"995100a7-4c5a-409a-81f2-9dbaf2fa1d47\") " Dec 01 09:05:55 crc kubenswrapper[4813]: I1201 09:05:55.436091 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swift\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"995100a7-4c5a-409a-81f2-9dbaf2fa1d47\" (UID: \"995100a7-4c5a-409a-81f2-9dbaf2fa1d47\") " Dec 01 09:05:55 crc kubenswrapper[4813]: I1201 09:05:55.439647 4813 reconciler_common.go:293] "Volume detached for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/995100a7-4c5a-409a-81f2-9dbaf2fa1d47-lock\") on node \"crc\" DevicePath \"\"" Dec 01 09:05:55 crc kubenswrapper[4813]: I1201 09:05:55.439685 4813 reconciler_common.go:293] "Volume detached for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/995100a7-4c5a-409a-81f2-9dbaf2fa1d47-cache\") on node \"crc\" DevicePath \"\"" Dec 01 09:05:55 crc kubenswrapper[4813]: I1201 09:05:55.440262 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage02-crc" (OuterVolumeSpecName: "swift") pod "995100a7-4c5a-409a-81f2-9dbaf2fa1d47" (UID: "995100a7-4c5a-409a-81f2-9dbaf2fa1d47"). InnerVolumeSpecName "local-storage02-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 01 09:05:55 crc kubenswrapper[4813]: I1201 09:05:55.440588 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/995100a7-4c5a-409a-81f2-9dbaf2fa1d47-kube-api-access-6jgvz" (OuterVolumeSpecName: "kube-api-access-6jgvz") pod "995100a7-4c5a-409a-81f2-9dbaf2fa1d47" (UID: "995100a7-4c5a-409a-81f2-9dbaf2fa1d47"). InnerVolumeSpecName "kube-api-access-6jgvz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:05:55 crc kubenswrapper[4813]: I1201 09:05:55.442415 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/995100a7-4c5a-409a-81f2-9dbaf2fa1d47-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "995100a7-4c5a-409a-81f2-9dbaf2fa1d47" (UID: "995100a7-4c5a-409a-81f2-9dbaf2fa1d47"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:05:55 crc kubenswrapper[4813]: I1201 09:05:55.541778 4813 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/995100a7-4c5a-409a-81f2-9dbaf2fa1d47-etc-swift\") on node \"crc\" DevicePath \"\"" Dec 01 09:05:55 crc kubenswrapper[4813]: I1201 09:05:55.542582 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6jgvz\" (UniqueName: \"kubernetes.io/projected/995100a7-4c5a-409a-81f2-9dbaf2fa1d47-kube-api-access-6jgvz\") on node \"crc\" DevicePath \"\"" Dec 01 09:05:55 crc kubenswrapper[4813]: I1201 09:05:55.542651 4813 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" " Dec 01 09:05:55 crc kubenswrapper[4813]: I1201 09:05:55.562900 4813 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage02-crc" (UniqueName: "kubernetes.io/local-volume/local-storage02-crc") on node "crc" Dec 01 09:05:55 crc kubenswrapper[4813]: I1201 09:05:55.643802 4813 reconciler_common.go:293] "Volume detached for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" DevicePath \"\"" Dec 01 09:05:56 crc kubenswrapper[4813]: I1201 09:05:56.062810 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"995100a7-4c5a-409a-81f2-9dbaf2fa1d47","Type":"ContainerDied","Data":"28acebd64f7f122629afcca41dc4293b68b2b5d7a6a7fb39197f92cfceb2c7e1"} Dec 01 09:05:56 crc kubenswrapper[4813]: I1201 09:05:56.062878 4813 scope.go:117] "RemoveContainer" containerID="08b235b36c9b32c3a32345cc647c1bd488a16d47403f387ba92c8337ae7120e1" Dec 01 09:05:56 crc kubenswrapper[4813]: I1201 09:05:56.062996 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/swift-storage-0" Dec 01 09:05:56 crc kubenswrapper[4813]: I1201 09:05:56.092622 4813 scope.go:117] "RemoveContainer" containerID="8a6edcb51ab7c3bb702c42f67b2432c8ace1cb2fa02e6a3db3fe9519ecf7f7f3" Dec 01 09:05:56 crc kubenswrapper[4813]: I1201 09:05:56.123079 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/swift-storage-0"] Dec 01 09:05:56 crc kubenswrapper[4813]: I1201 09:05:56.128805 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/swift-storage-0"] Dec 01 09:05:56 crc kubenswrapper[4813]: I1201 09:05:56.129022 4813 scope.go:117] "RemoveContainer" containerID="b5f856516db2a807cb93878232098a03c9c22a32b0221eba5c835097c972aca5" Dec 01 09:05:56 crc kubenswrapper[4813]: I1201 09:05:56.162027 4813 scope.go:117] "RemoveContainer" containerID="cfcbba573ff80c783a56d8b194b0bfee5c5bf61df702623a4752329024c48781" Dec 01 09:05:56 crc kubenswrapper[4813]: I1201 09:05:56.184432 4813 scope.go:117] "RemoveContainer" containerID="72d0f4ae012d4d6cd410f0e6e9f0041810afd974e12e68981d6f390c1ce1cb8a" Dec 01 09:05:56 crc kubenswrapper[4813]: I1201 09:05:56.200075 4813 scope.go:117] "RemoveContainer" containerID="4cd563e572aa4851bd7a2a6fdce31ea0962d58c89ab9c06f628d341281b7a914" Dec 01 09:05:56 crc kubenswrapper[4813]: I1201 09:05:56.219687 4813 scope.go:117] "RemoveContainer" containerID="90152b553c2c0696cb360f3d2b45d4e4f1d069d88a813084717ae44e3d7f007e" Dec 01 09:05:56 crc kubenswrapper[4813]: I1201 09:05:56.234034 4813 scope.go:117] "RemoveContainer" containerID="71ea72b9275533c0074b122f73f7beebac12af351e19c3c9c0e6c0b137e24f6f" Dec 01 09:05:56 crc kubenswrapper[4813]: I1201 09:05:56.249186 4813 scope.go:117] "RemoveContainer" containerID="6e3df495999a12b94a6ca78f0149ccd17c13cf04f3fc31edaef7a77490b259d1" Dec 01 09:05:56 crc kubenswrapper[4813]: I1201 09:05:56.266652 4813 scope.go:117] "RemoveContainer" containerID="d2d428af1b1d67ed1e1389a7fb3e3e1ea61d4ffd00a4526c2b45c9655f43d7ca" Dec 01 09:05:56 crc kubenswrapper[4813]: I1201 09:05:56.285889 4813 scope.go:117] "RemoveContainer" containerID="4a635b3fb091ce09c36d9181c2aa094b2a84e366b1bb19e73b4059e4af3205fa" Dec 01 09:05:56 crc kubenswrapper[4813]: I1201 09:05:56.303687 4813 scope.go:117] "RemoveContainer" containerID="2b9455a05903cfc3c8f0fe13912f2aeb0af2368d2f612165ece50ff474a9e935" Dec 01 09:05:56 crc kubenswrapper[4813]: I1201 09:05:56.327605 4813 scope.go:117] "RemoveContainer" containerID="0b01d236c0fd4b7e6153fcbb9c688913b9adc82fea28ce6cfa31368fb7072a52" Dec 01 09:05:56 crc kubenswrapper[4813]: I1201 09:05:56.358053 4813 scope.go:117] "RemoveContainer" containerID="532f0c05bdb2cee021285ec588aac0effe66a65c6f37749e25369d978d0515e4" Dec 01 09:05:56 crc kubenswrapper[4813]: I1201 09:05:56.380033 4813 scope.go:117] "RemoveContainer" containerID="c9f5cb19a620cfc7b9df0bad296f8f90075cf217fda34e7c9818acd5277ecc4b" Dec 01 09:05:56 crc kubenswrapper[4813]: I1201 09:05:56.412752 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="995100a7-4c5a-409a-81f2-9dbaf2fa1d47" path="/var/lib/kubelet/pods/995100a7-4c5a-409a-81f2-9dbaf2fa1d47/volumes" Dec 01 09:05:58 crc kubenswrapper[4813]: I1201 09:05:58.394085 4813 scope.go:117] "RemoveContainer" containerID="09b49517239f156b4ac50c72e4d184fde70e2daaf7b602adcfc3b42ec90a71e0" Dec 01 09:05:58 crc kubenswrapper[4813]: E1201 09:05:58.394475 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9wn4q_openshift-machine-config-operator(3d423402-e08f-42e9-b60c-6eb9bf2f6d48)\"" pod="openshift-machine-config-operator/machine-config-daemon-9wn4q" podUID="3d423402-e08f-42e9-b60c-6eb9bf2f6d48" Dec 01 09:06:02 crc kubenswrapper[4813]: I1201 09:06:02.182115 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-ffr6x/must-gather-2b9p6"] Dec 01 09:06:02 crc kubenswrapper[4813]: E1201 09:06:02.182428 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="73988763-722d-4652-8fac-053ba8217547" containerName="mysql-bootstrap" Dec 01 09:06:02 crc kubenswrapper[4813]: I1201 09:06:02.182451 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="73988763-722d-4652-8fac-053ba8217547" containerName="mysql-bootstrap" Dec 01 09:06:02 crc kubenswrapper[4813]: E1201 09:06:02.182460 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="07bbfadb-d0a3-478c-a2ed-36ddc6081a49" containerName="extract-content" Dec 01 09:06:02 crc kubenswrapper[4813]: I1201 09:06:02.182466 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="07bbfadb-d0a3-478c-a2ed-36ddc6081a49" containerName="extract-content" Dec 01 09:06:02 crc kubenswrapper[4813]: E1201 09:06:02.182476 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="995100a7-4c5a-409a-81f2-9dbaf2fa1d47" containerName="object-server" Dec 01 09:06:02 crc kubenswrapper[4813]: I1201 09:06:02.182486 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="995100a7-4c5a-409a-81f2-9dbaf2fa1d47" containerName="object-server" Dec 01 09:06:02 crc kubenswrapper[4813]: E1201 09:06:02.182498 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a274574f-1dfe-4afd-8dfa-661d14669e73" containerName="setup-container" Dec 01 09:06:02 crc kubenswrapper[4813]: I1201 09:06:02.182504 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="a274574f-1dfe-4afd-8dfa-661d14669e73" containerName="setup-container" Dec 01 09:06:02 crc kubenswrapper[4813]: E1201 09:06:02.182512 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="42e1cee7-cd3a-4194-af37-e29d5f04389a" containerName="kube-rbac-proxy" Dec 01 09:06:02 crc kubenswrapper[4813]: I1201 09:06:02.182517 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="42e1cee7-cd3a-4194-af37-e29d5f04389a" containerName="kube-rbac-proxy" Dec 01 09:06:02 crc kubenswrapper[4813]: E1201 09:06:02.182528 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="73988763-722d-4652-8fac-053ba8217547" containerName="galera" Dec 01 09:06:02 crc kubenswrapper[4813]: I1201 09:06:02.182533 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="73988763-722d-4652-8fac-053ba8217547" containerName="galera" Dec 01 09:06:02 crc kubenswrapper[4813]: E1201 09:06:02.182545 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="995100a7-4c5a-409a-81f2-9dbaf2fa1d47" containerName="account-replicator" Dec 01 09:06:02 crc kubenswrapper[4813]: I1201 09:06:02.182551 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="995100a7-4c5a-409a-81f2-9dbaf2fa1d47" containerName="account-replicator" Dec 01 09:06:02 crc kubenswrapper[4813]: E1201 09:06:02.182561 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="995100a7-4c5a-409a-81f2-9dbaf2fa1d47" containerName="container-server" Dec 01 09:06:02 crc kubenswrapper[4813]: I1201 09:06:02.182567 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="995100a7-4c5a-409a-81f2-9dbaf2fa1d47" containerName="container-server" Dec 01 09:06:02 crc kubenswrapper[4813]: E1201 09:06:02.182576 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a80abda8-962d-4b0a-bc3a-2e6d693d0393" containerName="keystone-api" Dec 01 09:06:02 crc kubenswrapper[4813]: I1201 09:06:02.182582 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="a80abda8-962d-4b0a-bc3a-2e6d693d0393" containerName="keystone-api" Dec 01 09:06:02 crc kubenswrapper[4813]: E1201 09:06:02.182588 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="995100a7-4c5a-409a-81f2-9dbaf2fa1d47" containerName="container-updater" Dec 01 09:06:02 crc kubenswrapper[4813]: I1201 09:06:02.182594 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="995100a7-4c5a-409a-81f2-9dbaf2fa1d47" containerName="container-updater" Dec 01 09:06:02 crc kubenswrapper[4813]: E1201 09:06:02.182603 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a274574f-1dfe-4afd-8dfa-661d14669e73" containerName="rabbitmq" Dec 01 09:06:02 crc kubenswrapper[4813]: I1201 09:06:02.182609 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="a274574f-1dfe-4afd-8dfa-661d14669e73" containerName="rabbitmq" Dec 01 09:06:02 crc kubenswrapper[4813]: E1201 09:06:02.182617 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3e52d3b1-2c61-4e2b-8222-35745063e838" containerName="manager" Dec 01 09:06:02 crc kubenswrapper[4813]: I1201 09:06:02.182623 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="3e52d3b1-2c61-4e2b-8222-35745063e838" containerName="manager" Dec 01 09:06:02 crc kubenswrapper[4813]: E1201 09:06:02.182630 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="995100a7-4c5a-409a-81f2-9dbaf2fa1d47" containerName="account-auditor" Dec 01 09:06:02 crc kubenswrapper[4813]: I1201 09:06:02.182636 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="995100a7-4c5a-409a-81f2-9dbaf2fa1d47" containerName="account-auditor" Dec 01 09:06:02 crc kubenswrapper[4813]: E1201 09:06:02.182645 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="88d8ade8-c257-464c-9dab-0052bed99037" containerName="memcached" Dec 01 09:06:02 crc kubenswrapper[4813]: I1201 09:06:02.182652 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="88d8ade8-c257-464c-9dab-0052bed99037" containerName="memcached" Dec 01 09:06:02 crc kubenswrapper[4813]: E1201 09:06:02.182673 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="abb50b08-d82d-4e6a-af37-5a99d2624339" containerName="manager" Dec 01 09:06:02 crc kubenswrapper[4813]: I1201 09:06:02.182679 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="abb50b08-d82d-4e6a-af37-5a99d2624339" containerName="manager" Dec 01 09:06:02 crc kubenswrapper[4813]: E1201 09:06:02.182686 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3820379b-32dc-46ea-a899-0990fadcab5b" containerName="operator" Dec 01 09:06:02 crc kubenswrapper[4813]: I1201 09:06:02.182692 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="3820379b-32dc-46ea-a899-0990fadcab5b" containerName="operator" Dec 01 09:06:02 crc kubenswrapper[4813]: E1201 09:06:02.182700 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="07bbfadb-d0a3-478c-a2ed-36ddc6081a49" containerName="extract-utilities" Dec 01 09:06:02 crc kubenswrapper[4813]: I1201 09:06:02.182707 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="07bbfadb-d0a3-478c-a2ed-36ddc6081a49" containerName="extract-utilities" Dec 01 09:06:02 crc kubenswrapper[4813]: E1201 09:06:02.182715 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="174d5e2e-9c14-48e0-af03-91a15c96ad66" containerName="registry-server" Dec 01 09:06:02 crc kubenswrapper[4813]: I1201 09:06:02.182721 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="174d5e2e-9c14-48e0-af03-91a15c96ad66" containerName="registry-server" Dec 01 09:06:02 crc kubenswrapper[4813]: E1201 09:06:02.182729 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="42e1cee7-cd3a-4194-af37-e29d5f04389a" containerName="manager" Dec 01 09:06:02 crc kubenswrapper[4813]: I1201 09:06:02.182734 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="42e1cee7-cd3a-4194-af37-e29d5f04389a" containerName="manager" Dec 01 09:06:02 crc kubenswrapper[4813]: E1201 09:06:02.182744 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="995100a7-4c5a-409a-81f2-9dbaf2fa1d47" containerName="container-auditor" Dec 01 09:06:02 crc kubenswrapper[4813]: I1201 09:06:02.182750 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="995100a7-4c5a-409a-81f2-9dbaf2fa1d47" containerName="container-auditor" Dec 01 09:06:02 crc kubenswrapper[4813]: E1201 09:06:02.182758 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b95238b9-4f2a-4ee8-8b18-44857550a8d1" containerName="registry-server" Dec 01 09:06:02 crc kubenswrapper[4813]: I1201 09:06:02.182764 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="b95238b9-4f2a-4ee8-8b18-44857550a8d1" containerName="registry-server" Dec 01 09:06:02 crc kubenswrapper[4813]: E1201 09:06:02.182772 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="995100a7-4c5a-409a-81f2-9dbaf2fa1d47" containerName="object-updater" Dec 01 09:06:02 crc kubenswrapper[4813]: I1201 09:06:02.182778 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="995100a7-4c5a-409a-81f2-9dbaf2fa1d47" containerName="object-updater" Dec 01 09:06:02 crc kubenswrapper[4813]: E1201 09:06:02.182784 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="995100a7-4c5a-409a-81f2-9dbaf2fa1d47" containerName="account-server" Dec 01 09:06:02 crc kubenswrapper[4813]: I1201 09:06:02.182790 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="995100a7-4c5a-409a-81f2-9dbaf2fa1d47" containerName="account-server" Dec 01 09:06:02 crc kubenswrapper[4813]: E1201 09:06:02.182796 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="07bbfadb-d0a3-478c-a2ed-36ddc6081a49" containerName="registry-server" Dec 01 09:06:02 crc kubenswrapper[4813]: I1201 09:06:02.182802 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="07bbfadb-d0a3-478c-a2ed-36ddc6081a49" containerName="registry-server" Dec 01 09:06:02 crc kubenswrapper[4813]: E1201 09:06:02.182810 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3f8cb860-7fc8-490c-b5d9-18eed8d8db79" containerName="registry-server" Dec 01 09:06:02 crc kubenswrapper[4813]: I1201 09:06:02.182815 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="3f8cb860-7fc8-490c-b5d9-18eed8d8db79" containerName="registry-server" Dec 01 09:06:02 crc kubenswrapper[4813]: E1201 09:06:02.182825 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="995100a7-4c5a-409a-81f2-9dbaf2fa1d47" containerName="rsync" Dec 01 09:06:02 crc kubenswrapper[4813]: I1201 09:06:02.182831 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="995100a7-4c5a-409a-81f2-9dbaf2fa1d47" containerName="rsync" Dec 01 09:06:02 crc kubenswrapper[4813]: E1201 09:06:02.182839 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="43c6449e-e9d4-4a5b-9231-d719d2a9e488" containerName="galera" Dec 01 09:06:02 crc kubenswrapper[4813]: I1201 09:06:02.182844 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="43c6449e-e9d4-4a5b-9231-d719d2a9e488" containerName="galera" Dec 01 09:06:02 crc kubenswrapper[4813]: E1201 09:06:02.182854 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="995100a7-4c5a-409a-81f2-9dbaf2fa1d47" containerName="object-auditor" Dec 01 09:06:02 crc kubenswrapper[4813]: I1201 09:06:02.182860 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="995100a7-4c5a-409a-81f2-9dbaf2fa1d47" containerName="object-auditor" Dec 01 09:06:02 crc kubenswrapper[4813]: E1201 09:06:02.182869 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="995100a7-4c5a-409a-81f2-9dbaf2fa1d47" containerName="account-reaper" Dec 01 09:06:02 crc kubenswrapper[4813]: I1201 09:06:02.182874 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="995100a7-4c5a-409a-81f2-9dbaf2fa1d47" containerName="account-reaper" Dec 01 09:06:02 crc kubenswrapper[4813]: E1201 09:06:02.182882 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="995100a7-4c5a-409a-81f2-9dbaf2fa1d47" containerName="object-expirer" Dec 01 09:06:02 crc kubenswrapper[4813]: I1201 09:06:02.182888 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="995100a7-4c5a-409a-81f2-9dbaf2fa1d47" containerName="object-expirer" Dec 01 09:06:02 crc kubenswrapper[4813]: E1201 09:06:02.182897 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="84350e15-9af9-4de1-bf0b-9e75c91b180a" containerName="manager" Dec 01 09:06:02 crc kubenswrapper[4813]: I1201 09:06:02.182904 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="84350e15-9af9-4de1-bf0b-9e75c91b180a" containerName="manager" Dec 01 09:06:02 crc kubenswrapper[4813]: E1201 09:06:02.182911 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="995100a7-4c5a-409a-81f2-9dbaf2fa1d47" containerName="object-replicator" Dec 01 09:06:02 crc kubenswrapper[4813]: I1201 09:06:02.182918 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="995100a7-4c5a-409a-81f2-9dbaf2fa1d47" containerName="object-replicator" Dec 01 09:06:02 crc kubenswrapper[4813]: E1201 09:06:02.182928 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="35db6389-d26f-45f7-919a-734a16107934" containerName="registry-server" Dec 01 09:06:02 crc kubenswrapper[4813]: I1201 09:06:02.182933 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="35db6389-d26f-45f7-919a-734a16107934" containerName="registry-server" Dec 01 09:06:02 crc kubenswrapper[4813]: E1201 09:06:02.182940 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="41360e5e-2620-4912-8035-0e1a9e0e715d" containerName="galera" Dec 01 09:06:02 crc kubenswrapper[4813]: I1201 09:06:02.182946 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="41360e5e-2620-4912-8035-0e1a9e0e715d" containerName="galera" Dec 01 09:06:02 crc kubenswrapper[4813]: E1201 09:06:02.182955 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="995100a7-4c5a-409a-81f2-9dbaf2fa1d47" containerName="swift-recon-cron" Dec 01 09:06:02 crc kubenswrapper[4813]: I1201 09:06:02.182973 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="995100a7-4c5a-409a-81f2-9dbaf2fa1d47" containerName="swift-recon-cron" Dec 01 09:06:02 crc kubenswrapper[4813]: E1201 09:06:02.182983 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c4d16fd1-e8d3-4bb5-af1f-204e484c8530" containerName="registry-server" Dec 01 09:06:02 crc kubenswrapper[4813]: I1201 09:06:02.182989 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="c4d16fd1-e8d3-4bb5-af1f-204e484c8530" containerName="registry-server" Dec 01 09:06:02 crc kubenswrapper[4813]: E1201 09:06:02.182997 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="43c6449e-e9d4-4a5b-9231-d719d2a9e488" containerName="mysql-bootstrap" Dec 01 09:06:02 crc kubenswrapper[4813]: I1201 09:06:02.183003 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="43c6449e-e9d4-4a5b-9231-d719d2a9e488" containerName="mysql-bootstrap" Dec 01 09:06:02 crc kubenswrapper[4813]: E1201 09:06:02.183012 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6fee8f87-bd86-4153-81b5-fbfc619f609c" containerName="registry-server" Dec 01 09:06:02 crc kubenswrapper[4813]: I1201 09:06:02.183018 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="6fee8f87-bd86-4153-81b5-fbfc619f609c" containerName="registry-server" Dec 01 09:06:02 crc kubenswrapper[4813]: E1201 09:06:02.183027 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="16ecbca6-953b-4f9f-981e-d9f011703629" containerName="mariadb-account-delete" Dec 01 09:06:02 crc kubenswrapper[4813]: I1201 09:06:02.183033 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="16ecbca6-953b-4f9f-981e-d9f011703629" containerName="mariadb-account-delete" Dec 01 09:06:02 crc kubenswrapper[4813]: E1201 09:06:02.183041 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="41360e5e-2620-4912-8035-0e1a9e0e715d" containerName="mysql-bootstrap" Dec 01 09:06:02 crc kubenswrapper[4813]: I1201 09:06:02.183047 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="41360e5e-2620-4912-8035-0e1a9e0e715d" containerName="mysql-bootstrap" Dec 01 09:06:02 crc kubenswrapper[4813]: E1201 09:06:02.183053 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2a681edc-eda3-4d33-a642-24898d63a97f" containerName="manager" Dec 01 09:06:02 crc kubenswrapper[4813]: I1201 09:06:02.183058 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="2a681edc-eda3-4d33-a642-24898d63a97f" containerName="manager" Dec 01 09:06:02 crc kubenswrapper[4813]: E1201 09:06:02.183066 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="995100a7-4c5a-409a-81f2-9dbaf2fa1d47" containerName="container-replicator" Dec 01 09:06:02 crc kubenswrapper[4813]: I1201 09:06:02.183071 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="995100a7-4c5a-409a-81f2-9dbaf2fa1d47" containerName="container-replicator" Dec 01 09:06:02 crc kubenswrapper[4813]: I1201 09:06:02.183210 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="c4d16fd1-e8d3-4bb5-af1f-204e484c8530" containerName="registry-server" Dec 01 09:06:02 crc kubenswrapper[4813]: I1201 09:06:02.183223 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="995100a7-4c5a-409a-81f2-9dbaf2fa1d47" containerName="account-server" Dec 01 09:06:02 crc kubenswrapper[4813]: I1201 09:06:02.183232 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="995100a7-4c5a-409a-81f2-9dbaf2fa1d47" containerName="account-reaper" Dec 01 09:06:02 crc kubenswrapper[4813]: I1201 09:06:02.183242 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="88d8ade8-c257-464c-9dab-0052bed99037" containerName="memcached" Dec 01 09:06:02 crc kubenswrapper[4813]: I1201 09:06:02.183249 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="a274574f-1dfe-4afd-8dfa-661d14669e73" containerName="rabbitmq" Dec 01 09:06:02 crc kubenswrapper[4813]: I1201 09:06:02.183257 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="41360e5e-2620-4912-8035-0e1a9e0e715d" containerName="galera" Dec 01 09:06:02 crc kubenswrapper[4813]: I1201 09:06:02.183266 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="995100a7-4c5a-409a-81f2-9dbaf2fa1d47" containerName="object-updater" Dec 01 09:06:02 crc kubenswrapper[4813]: I1201 09:06:02.183274 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="42e1cee7-cd3a-4194-af37-e29d5f04389a" containerName="kube-rbac-proxy" Dec 01 09:06:02 crc kubenswrapper[4813]: I1201 09:06:02.183279 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="3f8cb860-7fc8-490c-b5d9-18eed8d8db79" containerName="registry-server" Dec 01 09:06:02 crc kubenswrapper[4813]: I1201 09:06:02.183285 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="3e52d3b1-2c61-4e2b-8222-35745063e838" containerName="manager" Dec 01 09:06:02 crc kubenswrapper[4813]: I1201 09:06:02.183293 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="84350e15-9af9-4de1-bf0b-9e75c91b180a" containerName="manager" Dec 01 09:06:02 crc kubenswrapper[4813]: I1201 09:06:02.183300 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="35db6389-d26f-45f7-919a-734a16107934" containerName="registry-server" Dec 01 09:06:02 crc kubenswrapper[4813]: I1201 09:06:02.183306 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="995100a7-4c5a-409a-81f2-9dbaf2fa1d47" containerName="object-server" Dec 01 09:06:02 crc kubenswrapper[4813]: I1201 09:06:02.183313 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="a80abda8-962d-4b0a-bc3a-2e6d693d0393" containerName="keystone-api" Dec 01 09:06:02 crc kubenswrapper[4813]: I1201 09:06:02.183319 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="995100a7-4c5a-409a-81f2-9dbaf2fa1d47" containerName="object-expirer" Dec 01 09:06:02 crc kubenswrapper[4813]: I1201 09:06:02.183325 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="16ecbca6-953b-4f9f-981e-d9f011703629" containerName="mariadb-account-delete" Dec 01 09:06:02 crc kubenswrapper[4813]: I1201 09:06:02.183332 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="174d5e2e-9c14-48e0-af03-91a15c96ad66" containerName="registry-server" Dec 01 09:06:02 crc kubenswrapper[4813]: I1201 09:06:02.183338 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="b95238b9-4f2a-4ee8-8b18-44857550a8d1" containerName="registry-server" Dec 01 09:06:02 crc kubenswrapper[4813]: I1201 09:06:02.183345 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="995100a7-4c5a-409a-81f2-9dbaf2fa1d47" containerName="object-replicator" Dec 01 09:06:02 crc kubenswrapper[4813]: I1201 09:06:02.183353 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="995100a7-4c5a-409a-81f2-9dbaf2fa1d47" containerName="swift-recon-cron" Dec 01 09:06:02 crc kubenswrapper[4813]: I1201 09:06:02.183359 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="995100a7-4c5a-409a-81f2-9dbaf2fa1d47" containerName="container-server" Dec 01 09:06:02 crc kubenswrapper[4813]: I1201 09:06:02.183368 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="abb50b08-d82d-4e6a-af37-5a99d2624339" containerName="manager" Dec 01 09:06:02 crc kubenswrapper[4813]: I1201 09:06:02.183378 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="6fee8f87-bd86-4153-81b5-fbfc619f609c" containerName="registry-server" Dec 01 09:06:02 crc kubenswrapper[4813]: I1201 09:06:02.183384 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="995100a7-4c5a-409a-81f2-9dbaf2fa1d47" containerName="rsync" Dec 01 09:06:02 crc kubenswrapper[4813]: I1201 09:06:02.183423 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="42e1cee7-cd3a-4194-af37-e29d5f04389a" containerName="manager" Dec 01 09:06:02 crc kubenswrapper[4813]: I1201 09:06:02.183432 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="995100a7-4c5a-409a-81f2-9dbaf2fa1d47" containerName="account-replicator" Dec 01 09:06:02 crc kubenswrapper[4813]: I1201 09:06:02.183441 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="07bbfadb-d0a3-478c-a2ed-36ddc6081a49" containerName="registry-server" Dec 01 09:06:02 crc kubenswrapper[4813]: I1201 09:06:02.183448 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="995100a7-4c5a-409a-81f2-9dbaf2fa1d47" containerName="container-replicator" Dec 01 09:06:02 crc kubenswrapper[4813]: I1201 09:06:02.183454 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="995100a7-4c5a-409a-81f2-9dbaf2fa1d47" containerName="container-updater" Dec 01 09:06:02 crc kubenswrapper[4813]: I1201 09:06:02.183462 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="3820379b-32dc-46ea-a899-0990fadcab5b" containerName="operator" Dec 01 09:06:02 crc kubenswrapper[4813]: I1201 09:06:02.183468 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="2a681edc-eda3-4d33-a642-24898d63a97f" containerName="manager" Dec 01 09:06:02 crc kubenswrapper[4813]: I1201 09:06:02.183475 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="995100a7-4c5a-409a-81f2-9dbaf2fa1d47" containerName="account-auditor" Dec 01 09:06:02 crc kubenswrapper[4813]: I1201 09:06:02.183482 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="43c6449e-e9d4-4a5b-9231-d719d2a9e488" containerName="galera" Dec 01 09:06:02 crc kubenswrapper[4813]: I1201 09:06:02.183489 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="73988763-722d-4652-8fac-053ba8217547" containerName="galera" Dec 01 09:06:02 crc kubenswrapper[4813]: I1201 09:06:02.183495 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="995100a7-4c5a-409a-81f2-9dbaf2fa1d47" containerName="container-auditor" Dec 01 09:06:02 crc kubenswrapper[4813]: I1201 09:06:02.183502 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="995100a7-4c5a-409a-81f2-9dbaf2fa1d47" containerName="object-auditor" Dec 01 09:06:02 crc kubenswrapper[4813]: I1201 09:06:02.184243 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-ffr6x/must-gather-2b9p6" Dec 01 09:06:02 crc kubenswrapper[4813]: I1201 09:06:02.188489 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-ffr6x"/"default-dockercfg-5gtxz" Dec 01 09:06:02 crc kubenswrapper[4813]: I1201 09:06:02.188676 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-ffr6x"/"openshift-service-ca.crt" Dec 01 09:06:02 crc kubenswrapper[4813]: I1201 09:06:02.191380 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-ffr6x"/"kube-root-ca.crt" Dec 01 09:06:02 crc kubenswrapper[4813]: I1201 09:06:02.217521 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-ffr6x/must-gather-2b9p6"] Dec 01 09:06:02 crc kubenswrapper[4813]: I1201 09:06:02.345189 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/152332db-f02a-4365-9bc2-cf74ab20c6f2-must-gather-output\") pod \"must-gather-2b9p6\" (UID: \"152332db-f02a-4365-9bc2-cf74ab20c6f2\") " pod="openshift-must-gather-ffr6x/must-gather-2b9p6" Dec 01 09:06:02 crc kubenswrapper[4813]: I1201 09:06:02.345240 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w8nw9\" (UniqueName: \"kubernetes.io/projected/152332db-f02a-4365-9bc2-cf74ab20c6f2-kube-api-access-w8nw9\") pod \"must-gather-2b9p6\" (UID: \"152332db-f02a-4365-9bc2-cf74ab20c6f2\") " pod="openshift-must-gather-ffr6x/must-gather-2b9p6" Dec 01 09:06:02 crc kubenswrapper[4813]: I1201 09:06:02.446413 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/152332db-f02a-4365-9bc2-cf74ab20c6f2-must-gather-output\") pod \"must-gather-2b9p6\" (UID: \"152332db-f02a-4365-9bc2-cf74ab20c6f2\") " pod="openshift-must-gather-ffr6x/must-gather-2b9p6" Dec 01 09:06:02 crc kubenswrapper[4813]: I1201 09:06:02.446473 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w8nw9\" (UniqueName: \"kubernetes.io/projected/152332db-f02a-4365-9bc2-cf74ab20c6f2-kube-api-access-w8nw9\") pod \"must-gather-2b9p6\" (UID: \"152332db-f02a-4365-9bc2-cf74ab20c6f2\") " pod="openshift-must-gather-ffr6x/must-gather-2b9p6" Dec 01 09:06:02 crc kubenswrapper[4813]: I1201 09:06:02.447306 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/152332db-f02a-4365-9bc2-cf74ab20c6f2-must-gather-output\") pod \"must-gather-2b9p6\" (UID: \"152332db-f02a-4365-9bc2-cf74ab20c6f2\") " pod="openshift-must-gather-ffr6x/must-gather-2b9p6" Dec 01 09:06:02 crc kubenswrapper[4813]: I1201 09:06:02.466692 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w8nw9\" (UniqueName: \"kubernetes.io/projected/152332db-f02a-4365-9bc2-cf74ab20c6f2-kube-api-access-w8nw9\") pod \"must-gather-2b9p6\" (UID: \"152332db-f02a-4365-9bc2-cf74ab20c6f2\") " pod="openshift-must-gather-ffr6x/must-gather-2b9p6" Dec 01 09:06:02 crc kubenswrapper[4813]: I1201 09:06:02.503399 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-ffr6x/must-gather-2b9p6" Dec 01 09:06:02 crc kubenswrapper[4813]: I1201 09:06:02.715769 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-ffr6x/must-gather-2b9p6"] Dec 01 09:06:03 crc kubenswrapper[4813]: I1201 09:06:03.121993 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-ffr6x/must-gather-2b9p6" event={"ID":"152332db-f02a-4365-9bc2-cf74ab20c6f2","Type":"ContainerStarted","Data":"7f087a4277c525246b834d0f55a055fad8c3792839469235d9ccb9e7f5dc131a"} Dec 01 09:06:03 crc kubenswrapper[4813]: E1201 09:06:03.460398 4813 configmap.go:193] Couldn't get configMap glance-kuttl-tests/openstack-config: configmap "openstack-config" not found Dec 01 09:06:03 crc kubenswrapper[4813]: E1201 09:06:03.460568 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/4d0873f6-c7e7-4a9b-bd6e-a0263fd2bc5a-openstack-config podName:4d0873f6-c7e7-4a9b-bd6e-a0263fd2bc5a nodeName:}" failed. No retries permitted until 2025-12-01 09:06:03.960514745 +0000 UTC m=+1548.583436331 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "openstack-config" (UniqueName: "kubernetes.io/configmap/4d0873f6-c7e7-4a9b-bd6e-a0263fd2bc5a-openstack-config") pod "openstackclient" (UID: "4d0873f6-c7e7-4a9b-bd6e-a0263fd2bc5a") : configmap "openstack-config" not found Dec 01 09:06:03 crc kubenswrapper[4813]: E1201 09:06:03.461488 4813 secret.go:188] Couldn't get secret glance-kuttl-tests/openstack-config-secret: secret "openstack-config-secret" not found Dec 01 09:06:03 crc kubenswrapper[4813]: E1201 09:06:03.461533 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/4d0873f6-c7e7-4a9b-bd6e-a0263fd2bc5a-openstack-config-secret podName:4d0873f6-c7e7-4a9b-bd6e-a0263fd2bc5a nodeName:}" failed. No retries permitted until 2025-12-01 09:06:03.961520934 +0000 UTC m=+1548.584442520 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "openstack-config-secret" (UniqueName: "kubernetes.io/secret/4d0873f6-c7e7-4a9b-bd6e-a0263fd2bc5a-openstack-config-secret") pod "openstackclient" (UID: "4d0873f6-c7e7-4a9b-bd6e-a0263fd2bc5a") : secret "openstack-config-secret" not found Dec 01 09:06:03 crc kubenswrapper[4813]: E1201 09:06:03.969435 4813 configmap.go:193] Couldn't get configMap glance-kuttl-tests/openstack-config: configmap "openstack-config" not found Dec 01 09:06:03 crc kubenswrapper[4813]: E1201 09:06:03.969524 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/4d0873f6-c7e7-4a9b-bd6e-a0263fd2bc5a-openstack-config podName:4d0873f6-c7e7-4a9b-bd6e-a0263fd2bc5a nodeName:}" failed. No retries permitted until 2025-12-01 09:06:04.969505718 +0000 UTC m=+1549.592427304 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "openstack-config" (UniqueName: "kubernetes.io/configmap/4d0873f6-c7e7-4a9b-bd6e-a0263fd2bc5a-openstack-config") pod "openstackclient" (UID: "4d0873f6-c7e7-4a9b-bd6e-a0263fd2bc5a") : configmap "openstack-config" not found Dec 01 09:06:03 crc kubenswrapper[4813]: E1201 09:06:03.969444 4813 secret.go:188] Couldn't get secret glance-kuttl-tests/openstack-config-secret: secret "openstack-config-secret" not found Dec 01 09:06:03 crc kubenswrapper[4813]: E1201 09:06:03.969605 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/4d0873f6-c7e7-4a9b-bd6e-a0263fd2bc5a-openstack-config-secret podName:4d0873f6-c7e7-4a9b-bd6e-a0263fd2bc5a nodeName:}" failed. No retries permitted until 2025-12-01 09:06:04.969591791 +0000 UTC m=+1549.592513377 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "openstack-config-secret" (UniqueName: "kubernetes.io/secret/4d0873f6-c7e7-4a9b-bd6e-a0263fd2bc5a-openstack-config-secret") pod "openstackclient" (UID: "4d0873f6-c7e7-4a9b-bd6e-a0263fd2bc5a") : secret "openstack-config-secret" not found Dec 01 09:06:04 crc kubenswrapper[4813]: E1201 09:06:04.984951 4813 secret.go:188] Couldn't get secret glance-kuttl-tests/openstack-config-secret: secret "openstack-config-secret" not found Dec 01 09:06:04 crc kubenswrapper[4813]: E1201 09:06:04.985336 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/4d0873f6-c7e7-4a9b-bd6e-a0263fd2bc5a-openstack-config-secret podName:4d0873f6-c7e7-4a9b-bd6e-a0263fd2bc5a nodeName:}" failed. No retries permitted until 2025-12-01 09:06:06.985314521 +0000 UTC m=+1551.608236107 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "openstack-config-secret" (UniqueName: "kubernetes.io/secret/4d0873f6-c7e7-4a9b-bd6e-a0263fd2bc5a-openstack-config-secret") pod "openstackclient" (UID: "4d0873f6-c7e7-4a9b-bd6e-a0263fd2bc5a") : secret "openstack-config-secret" not found Dec 01 09:06:04 crc kubenswrapper[4813]: E1201 09:06:04.985021 4813 configmap.go:193] Couldn't get configMap glance-kuttl-tests/openstack-config: configmap "openstack-config" not found Dec 01 09:06:04 crc kubenswrapper[4813]: E1201 09:06:04.985464 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/4d0873f6-c7e7-4a9b-bd6e-a0263fd2bc5a-openstack-config podName:4d0873f6-c7e7-4a9b-bd6e-a0263fd2bc5a nodeName:}" failed. No retries permitted until 2025-12-01 09:06:06.985443424 +0000 UTC m=+1551.608365010 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "openstack-config" (UniqueName: "kubernetes.io/configmap/4d0873f6-c7e7-4a9b-bd6e-a0263fd2bc5a-openstack-config") pod "openstackclient" (UID: "4d0873f6-c7e7-4a9b-bd6e-a0263fd2bc5a") : configmap "openstack-config" not found Dec 01 09:06:06 crc kubenswrapper[4813]: I1201 09:06:06.952027 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-57xw2"] Dec 01 09:06:06 crc kubenswrapper[4813]: I1201 09:06:06.955235 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-57xw2" Dec 01 09:06:06 crc kubenswrapper[4813]: I1201 09:06:06.960335 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-57xw2"] Dec 01 09:06:07 crc kubenswrapper[4813]: I1201 09:06:07.018039 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vqp8v\" (UniqueName: \"kubernetes.io/projected/1a043261-4bae-46db-9b2f-20f6c9fc35aa-kube-api-access-vqp8v\") pod \"certified-operators-57xw2\" (UID: \"1a043261-4bae-46db-9b2f-20f6c9fc35aa\") " pod="openshift-marketplace/certified-operators-57xw2" Dec 01 09:06:07 crc kubenswrapper[4813]: I1201 09:06:07.018108 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1a043261-4bae-46db-9b2f-20f6c9fc35aa-catalog-content\") pod \"certified-operators-57xw2\" (UID: \"1a043261-4bae-46db-9b2f-20f6c9fc35aa\") " pod="openshift-marketplace/certified-operators-57xw2" Dec 01 09:06:07 crc kubenswrapper[4813]: I1201 09:06:07.018302 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1a043261-4bae-46db-9b2f-20f6c9fc35aa-utilities\") pod \"certified-operators-57xw2\" (UID: \"1a043261-4bae-46db-9b2f-20f6c9fc35aa\") " pod="openshift-marketplace/certified-operators-57xw2" Dec 01 09:06:07 crc kubenswrapper[4813]: E1201 09:06:07.018518 4813 secret.go:188] Couldn't get secret glance-kuttl-tests/openstack-config-secret: secret "openstack-config-secret" not found Dec 01 09:06:07 crc kubenswrapper[4813]: E1201 09:06:07.018599 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/4d0873f6-c7e7-4a9b-bd6e-a0263fd2bc5a-openstack-config-secret podName:4d0873f6-c7e7-4a9b-bd6e-a0263fd2bc5a nodeName:}" failed. No retries permitted until 2025-12-01 09:06:11.018576945 +0000 UTC m=+1555.641498621 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "openstack-config-secret" (UniqueName: "kubernetes.io/secret/4d0873f6-c7e7-4a9b-bd6e-a0263fd2bc5a-openstack-config-secret") pod "openstackclient" (UID: "4d0873f6-c7e7-4a9b-bd6e-a0263fd2bc5a") : secret "openstack-config-secret" not found Dec 01 09:06:07 crc kubenswrapper[4813]: E1201 09:06:07.018608 4813 configmap.go:193] Couldn't get configMap glance-kuttl-tests/openstack-config: configmap "openstack-config" not found Dec 01 09:06:07 crc kubenswrapper[4813]: E1201 09:06:07.018648 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/4d0873f6-c7e7-4a9b-bd6e-a0263fd2bc5a-openstack-config podName:4d0873f6-c7e7-4a9b-bd6e-a0263fd2bc5a nodeName:}" failed. No retries permitted until 2025-12-01 09:06:11.018638807 +0000 UTC m=+1555.641560393 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "openstack-config" (UniqueName: "kubernetes.io/configmap/4d0873f6-c7e7-4a9b-bd6e-a0263fd2bc5a-openstack-config") pod "openstackclient" (UID: "4d0873f6-c7e7-4a9b-bd6e-a0263fd2bc5a") : configmap "openstack-config" not found Dec 01 09:06:07 crc kubenswrapper[4813]: I1201 09:06:07.119490 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vqp8v\" (UniqueName: \"kubernetes.io/projected/1a043261-4bae-46db-9b2f-20f6c9fc35aa-kube-api-access-vqp8v\") pod \"certified-operators-57xw2\" (UID: \"1a043261-4bae-46db-9b2f-20f6c9fc35aa\") " pod="openshift-marketplace/certified-operators-57xw2" Dec 01 09:06:07 crc kubenswrapper[4813]: I1201 09:06:07.119559 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1a043261-4bae-46db-9b2f-20f6c9fc35aa-catalog-content\") pod \"certified-operators-57xw2\" (UID: \"1a043261-4bae-46db-9b2f-20f6c9fc35aa\") " pod="openshift-marketplace/certified-operators-57xw2" Dec 01 09:06:07 crc kubenswrapper[4813]: I1201 09:06:07.119600 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1a043261-4bae-46db-9b2f-20f6c9fc35aa-utilities\") pod \"certified-operators-57xw2\" (UID: \"1a043261-4bae-46db-9b2f-20f6c9fc35aa\") " pod="openshift-marketplace/certified-operators-57xw2" Dec 01 09:06:07 crc kubenswrapper[4813]: I1201 09:06:07.120200 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1a043261-4bae-46db-9b2f-20f6c9fc35aa-utilities\") pod \"certified-operators-57xw2\" (UID: \"1a043261-4bae-46db-9b2f-20f6c9fc35aa\") " pod="openshift-marketplace/certified-operators-57xw2" Dec 01 09:06:07 crc kubenswrapper[4813]: I1201 09:06:07.120286 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1a043261-4bae-46db-9b2f-20f6c9fc35aa-catalog-content\") pod \"certified-operators-57xw2\" (UID: \"1a043261-4bae-46db-9b2f-20f6c9fc35aa\") " pod="openshift-marketplace/certified-operators-57xw2" Dec 01 09:06:07 crc kubenswrapper[4813]: I1201 09:06:07.145441 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vqp8v\" (UniqueName: \"kubernetes.io/projected/1a043261-4bae-46db-9b2f-20f6c9fc35aa-kube-api-access-vqp8v\") pod \"certified-operators-57xw2\" (UID: \"1a043261-4bae-46db-9b2f-20f6c9fc35aa\") " pod="openshift-marketplace/certified-operators-57xw2" Dec 01 09:06:07 crc kubenswrapper[4813]: I1201 09:06:07.156778 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-ffr6x/must-gather-2b9p6" event={"ID":"152332db-f02a-4365-9bc2-cf74ab20c6f2","Type":"ContainerStarted","Data":"f167f42106c1e02babf66072ed2ece2eb0e7d9b99a5d90d4ebb96a87e5664cb5"} Dec 01 09:06:07 crc kubenswrapper[4813]: I1201 09:06:07.156831 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-ffr6x/must-gather-2b9p6" event={"ID":"152332db-f02a-4365-9bc2-cf74ab20c6f2","Type":"ContainerStarted","Data":"a0b97e73e8cb1a9533b45f04d39bb84d132e25c893ea551fe83995b9979892e7"} Dec 01 09:06:07 crc kubenswrapper[4813]: I1201 09:06:07.175670 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-ffr6x/must-gather-2b9p6" podStartSLOduration=1.267583472 podStartE2EDuration="5.175617576s" podCreationTimestamp="2025-12-01 09:06:02 +0000 UTC" firstStartedPulling="2025-12-01 09:06:02.741674648 +0000 UTC m=+1547.364596234" lastFinishedPulling="2025-12-01 09:06:06.649708752 +0000 UTC m=+1551.272630338" observedRunningTime="2025-12-01 09:06:07.171702765 +0000 UTC m=+1551.794624361" watchObservedRunningTime="2025-12-01 09:06:07.175617576 +0000 UTC m=+1551.798539172" Dec 01 09:06:07 crc kubenswrapper[4813]: I1201 09:06:07.272742 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-57xw2" Dec 01 09:06:07 crc kubenswrapper[4813]: I1201 09:06:07.749840 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-57xw2"] Dec 01 09:06:08 crc kubenswrapper[4813]: I1201 09:06:08.165754 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-57xw2" event={"ID":"1a043261-4bae-46db-9b2f-20f6c9fc35aa","Type":"ContainerStarted","Data":"81318cd7f5d4df110bb1a6faf651b5ec16110359f93d87788e8dff532a43f879"} Dec 01 09:06:09 crc kubenswrapper[4813]: I1201 09:06:09.175160 4813 generic.go:334] "Generic (PLEG): container finished" podID="1a043261-4bae-46db-9b2f-20f6c9fc35aa" containerID="1138372aa418b8350a77f894039b3c2e893facd0e1dae3b0c75c39c018f529e6" exitCode=0 Dec 01 09:06:09 crc kubenswrapper[4813]: I1201 09:06:09.175223 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-57xw2" event={"ID":"1a043261-4bae-46db-9b2f-20f6c9fc35aa","Type":"ContainerDied","Data":"1138372aa418b8350a77f894039b3c2e893facd0e1dae3b0c75c39c018f529e6"} Dec 01 09:06:10 crc kubenswrapper[4813]: I1201 09:06:10.185130 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-57xw2" event={"ID":"1a043261-4bae-46db-9b2f-20f6c9fc35aa","Type":"ContainerStarted","Data":"f1849564acaadab2451d37c635d59bf4d41c466a8e4d4456a8cff691f55eb93c"} Dec 01 09:06:11 crc kubenswrapper[4813]: E1201 09:06:11.030392 4813 configmap.go:193] Couldn't get configMap glance-kuttl-tests/openstack-config: configmap "openstack-config" not found Dec 01 09:06:11 crc kubenswrapper[4813]: E1201 09:06:11.030524 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/4d0873f6-c7e7-4a9b-bd6e-a0263fd2bc5a-openstack-config podName:4d0873f6-c7e7-4a9b-bd6e-a0263fd2bc5a nodeName:}" failed. No retries permitted until 2025-12-01 09:06:19.030495868 +0000 UTC m=+1563.653417494 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "openstack-config" (UniqueName: "kubernetes.io/configmap/4d0873f6-c7e7-4a9b-bd6e-a0263fd2bc5a-openstack-config") pod "openstackclient" (UID: "4d0873f6-c7e7-4a9b-bd6e-a0263fd2bc5a") : configmap "openstack-config" not found Dec 01 09:06:11 crc kubenswrapper[4813]: E1201 09:06:11.030541 4813 secret.go:188] Couldn't get secret glance-kuttl-tests/openstack-config-secret: secret "openstack-config-secret" not found Dec 01 09:06:11 crc kubenswrapper[4813]: E1201 09:06:11.030740 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/4d0873f6-c7e7-4a9b-bd6e-a0263fd2bc5a-openstack-config-secret podName:4d0873f6-c7e7-4a9b-bd6e-a0263fd2bc5a nodeName:}" failed. No retries permitted until 2025-12-01 09:06:19.030708174 +0000 UTC m=+1563.653629760 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "openstack-config-secret" (UniqueName: "kubernetes.io/secret/4d0873f6-c7e7-4a9b-bd6e-a0263fd2bc5a-openstack-config-secret") pod "openstackclient" (UID: "4d0873f6-c7e7-4a9b-bd6e-a0263fd2bc5a") : secret "openstack-config-secret" not found Dec 01 09:06:11 crc kubenswrapper[4813]: I1201 09:06:11.195905 4813 generic.go:334] "Generic (PLEG): container finished" podID="1a043261-4bae-46db-9b2f-20f6c9fc35aa" containerID="f1849564acaadab2451d37c635d59bf4d41c466a8e4d4456a8cff691f55eb93c" exitCode=0 Dec 01 09:06:11 crc kubenswrapper[4813]: I1201 09:06:11.195990 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-57xw2" event={"ID":"1a043261-4bae-46db-9b2f-20f6c9fc35aa","Type":"ContainerDied","Data":"f1849564acaadab2451d37c635d59bf4d41c466a8e4d4456a8cff691f55eb93c"} Dec 01 09:06:12 crc kubenswrapper[4813]: I1201 09:06:12.206290 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-57xw2" event={"ID":"1a043261-4bae-46db-9b2f-20f6c9fc35aa","Type":"ContainerStarted","Data":"e732eb3ddd2c9f81862379cca1e2982206ca5a2987fa73e71b6a70592e2ff2f0"} Dec 01 09:06:12 crc kubenswrapper[4813]: I1201 09:06:12.228047 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-57xw2" podStartSLOduration=3.755350517 podStartE2EDuration="6.228023956s" podCreationTimestamp="2025-12-01 09:06:06 +0000 UTC" firstStartedPulling="2025-12-01 09:06:09.17693916 +0000 UTC m=+1553.799860746" lastFinishedPulling="2025-12-01 09:06:11.649612599 +0000 UTC m=+1556.272534185" observedRunningTime="2025-12-01 09:06:12.223172588 +0000 UTC m=+1556.846094244" watchObservedRunningTime="2025-12-01 09:06:12.228023956 +0000 UTC m=+1556.850945542" Dec 01 09:06:13 crc kubenswrapper[4813]: I1201 09:06:13.393482 4813 scope.go:117] "RemoveContainer" containerID="09b49517239f156b4ac50c72e4d184fde70e2daaf7b602adcfc3b42ec90a71e0" Dec 01 09:06:13 crc kubenswrapper[4813]: E1201 09:06:13.393705 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9wn4q_openshift-machine-config-operator(3d423402-e08f-42e9-b60c-6eb9bf2f6d48)\"" pod="openshift-machine-config-operator/machine-config-daemon-9wn4q" podUID="3d423402-e08f-42e9-b60c-6eb9bf2f6d48" Dec 01 09:06:17 crc kubenswrapper[4813]: I1201 09:06:17.273736 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-57xw2" Dec 01 09:06:17 crc kubenswrapper[4813]: I1201 09:06:17.274897 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-57xw2" Dec 01 09:06:17 crc kubenswrapper[4813]: I1201 09:06:17.311740 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-57xw2" Dec 01 09:06:17 crc kubenswrapper[4813]: I1201 09:06:17.784717 4813 scope.go:117] "RemoveContainer" containerID="f134099a523e27c706e1cb041a4e91b3cdd27b3f9d19bf30f9407a8691909167" Dec 01 09:06:17 crc kubenswrapper[4813]: I1201 09:06:17.812024 4813 scope.go:117] "RemoveContainer" containerID="4e11dedd29ab57c4ae5dcca4bc4950418b6887646e4a1a2186fa20a3abb3c157" Dec 01 09:06:17 crc kubenswrapper[4813]: I1201 09:06:17.900539 4813 scope.go:117] "RemoveContainer" containerID="3087b1b199743282b6c6e5936792080167b630ca83056686c61c81117d0593c0" Dec 01 09:06:17 crc kubenswrapper[4813]: I1201 09:06:17.923173 4813 scope.go:117] "RemoveContainer" containerID="4cbeee28f90ffdf966d5d0157b2f7373e15ff1820f16530504a35e976c1128d2" Dec 01 09:06:17 crc kubenswrapper[4813]: I1201 09:06:17.942884 4813 scope.go:117] "RemoveContainer" containerID="be43a5c128772057e5349a736cc9bad664c0f0c15e38b3b70ab423e136bca998" Dec 01 09:06:17 crc kubenswrapper[4813]: I1201 09:06:17.961264 4813 scope.go:117] "RemoveContainer" containerID="f85162b0555d9137364b7ea0770528a5d3d405ae6a72a4ed39d2524bc0b97042" Dec 01 09:06:18 crc kubenswrapper[4813]: I1201 09:06:18.009957 4813 scope.go:117] "RemoveContainer" containerID="6c9d4fd47e98ed28fd8994c1016abb1bd81eb6ea18f494e4bf22aeeb2ec24c2d" Dec 01 09:06:18 crc kubenswrapper[4813]: I1201 09:06:18.040253 4813 scope.go:117] "RemoveContainer" containerID="6eae32d4e0a5b9b7e4f9d222515b1e4685e708027fc970d304e51a7d82dc0ea7" Dec 01 09:06:18 crc kubenswrapper[4813]: I1201 09:06:18.059277 4813 scope.go:117] "RemoveContainer" containerID="06a6da92472edd0d069f498ec01bb66b4764863b11ad60a1218b0c6576ea89de" Dec 01 09:06:18 crc kubenswrapper[4813]: I1201 09:06:18.077204 4813 scope.go:117] "RemoveContainer" containerID="25b3a5735cd7eaa24c2967139820a3df74e032619e2a68233879342cdf48fa2c" Dec 01 09:06:18 crc kubenswrapper[4813]: I1201 09:06:18.099875 4813 scope.go:117] "RemoveContainer" containerID="b3764de15eb2105efb977ba179c0526333622e4ab41c6c22974988ef16451ed7" Dec 01 09:06:18 crc kubenswrapper[4813]: I1201 09:06:18.120881 4813 scope.go:117] "RemoveContainer" containerID="cc8ed80f25c94a3910eb49f3e867bbe55dc9a7d59f8c1d3d1c1505e51538c778" Dec 01 09:06:18 crc kubenswrapper[4813]: I1201 09:06:18.135709 4813 scope.go:117] "RemoveContainer" containerID="6b8b612de5a1e9011c313e31e4406c95924b8b9af3a343a761a6a6d23864b416" Dec 01 09:06:18 crc kubenswrapper[4813]: I1201 09:06:18.151380 4813 scope.go:117] "RemoveContainer" containerID="0b87db48124c7f7feff803119917724c19e455832f19c3f2431969766e9d77a7" Dec 01 09:06:18 crc kubenswrapper[4813]: I1201 09:06:18.169206 4813 scope.go:117] "RemoveContainer" containerID="290ea1b79b6f4b663007a194b70dbae1e7de36d13ce1e6540f824831f816f892" Dec 01 09:06:18 crc kubenswrapper[4813]: I1201 09:06:18.188873 4813 scope.go:117] "RemoveContainer" containerID="0143f1788b1191f5c1d35c6e1a0986aa9fa3172f2a269ba224e731080c7c7a10" Dec 01 09:06:18 crc kubenswrapper[4813]: I1201 09:06:18.204687 4813 scope.go:117] "RemoveContainer" containerID="eac1c2f1c7e233613b531497d81712a4b61aed431ac7cad1204910e5e2633b87" Dec 01 09:06:18 crc kubenswrapper[4813]: I1201 09:06:18.224581 4813 scope.go:117] "RemoveContainer" containerID="ea57ebe323034f474f161dd5537a994724462fbe4ec60285cf2934c80c409ec0" Dec 01 09:06:18 crc kubenswrapper[4813]: I1201 09:06:18.240372 4813 scope.go:117] "RemoveContainer" containerID="6728f57e2f7744cf7ec174d5a9f8e49de494f20365bc3d286390a66b714260e0" Dec 01 09:06:18 crc kubenswrapper[4813]: I1201 09:06:18.262342 4813 scope.go:117] "RemoveContainer" containerID="48dbd3ab6f02167f321a00621dc1663966530d13f29261a0304ea32c0e6001c9" Dec 01 09:06:18 crc kubenswrapper[4813]: I1201 09:06:18.280986 4813 scope.go:117] "RemoveContainer" containerID="69c747470314cd28ca10e4f24bc635fc4aec16e2a4fc9c19015931af4655e644" Dec 01 09:06:18 crc kubenswrapper[4813]: I1201 09:06:18.299777 4813 scope.go:117] "RemoveContainer" containerID="9c0ec35f5d5d3797790bee4bbb0abe2d40e46be0e232315ab4892b4a6bbf2d84" Dec 01 09:06:18 crc kubenswrapper[4813]: I1201 09:06:18.321386 4813 scope.go:117] "RemoveContainer" containerID="6ded26056932a862dac56217fb260faaea05eba2290dfb9fffd9893408dd638c" Dec 01 09:06:18 crc kubenswrapper[4813]: I1201 09:06:18.322177 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-57xw2" Dec 01 09:06:18 crc kubenswrapper[4813]: I1201 09:06:18.342372 4813 scope.go:117] "RemoveContainer" containerID="c05742026abe1865b48f1118ca08065c5b800894aeeaacc582124c7505e4a7ef" Dec 01 09:06:18 crc kubenswrapper[4813]: I1201 09:06:18.360931 4813 scope.go:117] "RemoveContainer" containerID="2776f225b4fec97833b5b31c52c31da4f313fc6edf4b20c00fc9f06c1e82ad5b" Dec 01 09:06:18 crc kubenswrapper[4813]: I1201 09:06:18.385845 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-57xw2"] Dec 01 09:06:18 crc kubenswrapper[4813]: I1201 09:06:18.398142 4813 scope.go:117] "RemoveContainer" containerID="b1134778a50ac07e3bd0a14620a57aaf591e9bb521245d9851ad3a1f57aea8b8" Dec 01 09:06:18 crc kubenswrapper[4813]: I1201 09:06:18.417110 4813 scope.go:117] "RemoveContainer" containerID="1919ffcab60a2c9578329d6abb86defb8b5f6220be98ab8d091ef47b8e44420f" Dec 01 09:06:18 crc kubenswrapper[4813]: I1201 09:06:18.434891 4813 scope.go:117] "RemoveContainer" containerID="45a618c9430ab1f255c16e35c6c18bd5d07f8286d9619ee72c3673628cc81da7" Dec 01 09:06:18 crc kubenswrapper[4813]: I1201 09:06:18.462479 4813 scope.go:117] "RemoveContainer" containerID="c77365be1f5b03ab4bf99c3ffbe4da8a4e301ddcea7a11b9f74f9f10847510e6" Dec 01 09:06:19 crc kubenswrapper[4813]: E1201 09:06:19.091541 4813 configmap.go:193] Couldn't get configMap glance-kuttl-tests/openstack-config: configmap "openstack-config" not found Dec 01 09:06:19 crc kubenswrapper[4813]: E1201 09:06:19.091628 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/4d0873f6-c7e7-4a9b-bd6e-a0263fd2bc5a-openstack-config podName:4d0873f6-c7e7-4a9b-bd6e-a0263fd2bc5a nodeName:}" failed. No retries permitted until 2025-12-01 09:06:35.091606229 +0000 UTC m=+1579.714527815 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "openstack-config" (UniqueName: "kubernetes.io/configmap/4d0873f6-c7e7-4a9b-bd6e-a0263fd2bc5a-openstack-config") pod "openstackclient" (UID: "4d0873f6-c7e7-4a9b-bd6e-a0263fd2bc5a") : configmap "openstack-config" not found Dec 01 09:06:19 crc kubenswrapper[4813]: E1201 09:06:19.091557 4813 secret.go:188] Couldn't get secret glance-kuttl-tests/openstack-config-secret: secret "openstack-config-secret" not found Dec 01 09:06:19 crc kubenswrapper[4813]: E1201 09:06:19.092038 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/4d0873f6-c7e7-4a9b-bd6e-a0263fd2bc5a-openstack-config-secret podName:4d0873f6-c7e7-4a9b-bd6e-a0263fd2bc5a nodeName:}" failed. No retries permitted until 2025-12-01 09:06:35.092023581 +0000 UTC m=+1579.714945167 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "openstack-config-secret" (UniqueName: "kubernetes.io/secret/4d0873f6-c7e7-4a9b-bd6e-a0263fd2bc5a-openstack-config-secret") pod "openstackclient" (UID: "4d0873f6-c7e7-4a9b-bd6e-a0263fd2bc5a") : secret "openstack-config-secret" not found Dec 01 09:06:20 crc kubenswrapper[4813]: I1201 09:06:20.289767 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-57xw2" podUID="1a043261-4bae-46db-9b2f-20f6c9fc35aa" containerName="registry-server" containerID="cri-o://e732eb3ddd2c9f81862379cca1e2982206ca5a2987fa73e71b6a70592e2ff2f0" gracePeriod=2 Dec 01 09:06:21 crc kubenswrapper[4813]: I1201 09:06:21.297385 4813 generic.go:334] "Generic (PLEG): container finished" podID="1a043261-4bae-46db-9b2f-20f6c9fc35aa" containerID="e732eb3ddd2c9f81862379cca1e2982206ca5a2987fa73e71b6a70592e2ff2f0" exitCode=0 Dec 01 09:06:21 crc kubenswrapper[4813]: I1201 09:06:21.297502 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-57xw2" event={"ID":"1a043261-4bae-46db-9b2f-20f6c9fc35aa","Type":"ContainerDied","Data":"e732eb3ddd2c9f81862379cca1e2982206ca5a2987fa73e71b6a70592e2ff2f0"} Dec 01 09:06:21 crc kubenswrapper[4813]: I1201 09:06:21.748549 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-57xw2" Dec 01 09:06:21 crc kubenswrapper[4813]: I1201 09:06:21.933126 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vqp8v\" (UniqueName: \"kubernetes.io/projected/1a043261-4bae-46db-9b2f-20f6c9fc35aa-kube-api-access-vqp8v\") pod \"1a043261-4bae-46db-9b2f-20f6c9fc35aa\" (UID: \"1a043261-4bae-46db-9b2f-20f6c9fc35aa\") " Dec 01 09:06:21 crc kubenswrapper[4813]: I1201 09:06:21.933238 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1a043261-4bae-46db-9b2f-20f6c9fc35aa-utilities\") pod \"1a043261-4bae-46db-9b2f-20f6c9fc35aa\" (UID: \"1a043261-4bae-46db-9b2f-20f6c9fc35aa\") " Dec 01 09:06:21 crc kubenswrapper[4813]: I1201 09:06:21.933281 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1a043261-4bae-46db-9b2f-20f6c9fc35aa-catalog-content\") pod \"1a043261-4bae-46db-9b2f-20f6c9fc35aa\" (UID: \"1a043261-4bae-46db-9b2f-20f6c9fc35aa\") " Dec 01 09:06:21 crc kubenswrapper[4813]: I1201 09:06:21.934132 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1a043261-4bae-46db-9b2f-20f6c9fc35aa-utilities" (OuterVolumeSpecName: "utilities") pod "1a043261-4bae-46db-9b2f-20f6c9fc35aa" (UID: "1a043261-4bae-46db-9b2f-20f6c9fc35aa"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 09:06:21 crc kubenswrapper[4813]: I1201 09:06:21.940300 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1a043261-4bae-46db-9b2f-20f6c9fc35aa-kube-api-access-vqp8v" (OuterVolumeSpecName: "kube-api-access-vqp8v") pod "1a043261-4bae-46db-9b2f-20f6c9fc35aa" (UID: "1a043261-4bae-46db-9b2f-20f6c9fc35aa"). InnerVolumeSpecName "kube-api-access-vqp8v". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:06:21 crc kubenswrapper[4813]: I1201 09:06:21.987448 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1a043261-4bae-46db-9b2f-20f6c9fc35aa-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1a043261-4bae-46db-9b2f-20f6c9fc35aa" (UID: "1a043261-4bae-46db-9b2f-20f6c9fc35aa"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 09:06:22 crc kubenswrapper[4813]: I1201 09:06:22.034685 4813 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1a043261-4bae-46db-9b2f-20f6c9fc35aa-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 09:06:22 crc kubenswrapper[4813]: I1201 09:06:22.034741 4813 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1a043261-4bae-46db-9b2f-20f6c9fc35aa-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 09:06:22 crc kubenswrapper[4813]: I1201 09:06:22.034769 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vqp8v\" (UniqueName: \"kubernetes.io/projected/1a043261-4bae-46db-9b2f-20f6c9fc35aa-kube-api-access-vqp8v\") on node \"crc\" DevicePath \"\"" Dec 01 09:06:22 crc kubenswrapper[4813]: I1201 09:06:22.317306 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-57xw2" event={"ID":"1a043261-4bae-46db-9b2f-20f6c9fc35aa","Type":"ContainerDied","Data":"81318cd7f5d4df110bb1a6faf651b5ec16110359f93d87788e8dff532a43f879"} Dec 01 09:06:22 crc kubenswrapper[4813]: I1201 09:06:22.317387 4813 scope.go:117] "RemoveContainer" containerID="e732eb3ddd2c9f81862379cca1e2982206ca5a2987fa73e71b6a70592e2ff2f0" Dec 01 09:06:22 crc kubenswrapper[4813]: I1201 09:06:22.317397 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-57xw2" Dec 01 09:06:22 crc kubenswrapper[4813]: I1201 09:06:22.335371 4813 scope.go:117] "RemoveContainer" containerID="f1849564acaadab2451d37c635d59bf4d41c466a8e4d4456a8cff691f55eb93c" Dec 01 09:06:22 crc kubenswrapper[4813]: I1201 09:06:22.351159 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-57xw2"] Dec 01 09:06:22 crc kubenswrapper[4813]: I1201 09:06:22.363261 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-57xw2"] Dec 01 09:06:22 crc kubenswrapper[4813]: I1201 09:06:22.368183 4813 scope.go:117] "RemoveContainer" containerID="1138372aa418b8350a77f894039b3c2e893facd0e1dae3b0c75c39c018f529e6" Dec 01 09:06:22 crc kubenswrapper[4813]: I1201 09:06:22.402645 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1a043261-4bae-46db-9b2f-20f6c9fc35aa" path="/var/lib/kubelet/pods/1a043261-4bae-46db-9b2f-20f6c9fc35aa/volumes" Dec 01 09:06:26 crc kubenswrapper[4813]: I1201 09:06:26.398217 4813 scope.go:117] "RemoveContainer" containerID="09b49517239f156b4ac50c72e4d184fde70e2daaf7b602adcfc3b42ec90a71e0" Dec 01 09:06:26 crc kubenswrapper[4813]: E1201 09:06:26.399072 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9wn4q_openshift-machine-config-operator(3d423402-e08f-42e9-b60c-6eb9bf2f6d48)\"" pod="openshift-machine-config-operator/machine-config-daemon-9wn4q" podUID="3d423402-e08f-42e9-b60c-6eb9bf2f6d48" Dec 01 09:06:32 crc kubenswrapper[4813]: I1201 09:06:32.626519 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-7c4v7"] Dec 01 09:06:32 crc kubenswrapper[4813]: E1201 09:06:32.627530 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1a043261-4bae-46db-9b2f-20f6c9fc35aa" containerName="extract-content" Dec 01 09:06:32 crc kubenswrapper[4813]: I1201 09:06:32.627550 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="1a043261-4bae-46db-9b2f-20f6c9fc35aa" containerName="extract-content" Dec 01 09:06:32 crc kubenswrapper[4813]: E1201 09:06:32.627581 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1a043261-4bae-46db-9b2f-20f6c9fc35aa" containerName="extract-utilities" Dec 01 09:06:32 crc kubenswrapper[4813]: I1201 09:06:32.627588 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="1a043261-4bae-46db-9b2f-20f6c9fc35aa" containerName="extract-utilities" Dec 01 09:06:32 crc kubenswrapper[4813]: E1201 09:06:32.627603 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1a043261-4bae-46db-9b2f-20f6c9fc35aa" containerName="registry-server" Dec 01 09:06:32 crc kubenswrapper[4813]: I1201 09:06:32.627610 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="1a043261-4bae-46db-9b2f-20f6c9fc35aa" containerName="registry-server" Dec 01 09:06:32 crc kubenswrapper[4813]: I1201 09:06:32.627739 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="1a043261-4bae-46db-9b2f-20f6c9fc35aa" containerName="registry-server" Dec 01 09:06:32 crc kubenswrapper[4813]: I1201 09:06:32.628550 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7c4v7" Dec 01 09:06:32 crc kubenswrapper[4813]: I1201 09:06:32.638616 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-7c4v7"] Dec 01 09:06:32 crc kubenswrapper[4813]: I1201 09:06:32.779268 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-79hx5\" (UniqueName: \"kubernetes.io/projected/9aa179fa-17ce-4b1d-8016-1906584cd152-kube-api-access-79hx5\") pod \"redhat-operators-7c4v7\" (UID: \"9aa179fa-17ce-4b1d-8016-1906584cd152\") " pod="openshift-marketplace/redhat-operators-7c4v7" Dec 01 09:06:32 crc kubenswrapper[4813]: I1201 09:06:32.779367 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9aa179fa-17ce-4b1d-8016-1906584cd152-catalog-content\") pod \"redhat-operators-7c4v7\" (UID: \"9aa179fa-17ce-4b1d-8016-1906584cd152\") " pod="openshift-marketplace/redhat-operators-7c4v7" Dec 01 09:06:32 crc kubenswrapper[4813]: I1201 09:06:32.779392 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9aa179fa-17ce-4b1d-8016-1906584cd152-utilities\") pod \"redhat-operators-7c4v7\" (UID: \"9aa179fa-17ce-4b1d-8016-1906584cd152\") " pod="openshift-marketplace/redhat-operators-7c4v7" Dec 01 09:06:32 crc kubenswrapper[4813]: I1201 09:06:32.880378 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-79hx5\" (UniqueName: \"kubernetes.io/projected/9aa179fa-17ce-4b1d-8016-1906584cd152-kube-api-access-79hx5\") pod \"redhat-operators-7c4v7\" (UID: \"9aa179fa-17ce-4b1d-8016-1906584cd152\") " pod="openshift-marketplace/redhat-operators-7c4v7" Dec 01 09:06:32 crc kubenswrapper[4813]: I1201 09:06:32.880465 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9aa179fa-17ce-4b1d-8016-1906584cd152-catalog-content\") pod \"redhat-operators-7c4v7\" (UID: \"9aa179fa-17ce-4b1d-8016-1906584cd152\") " pod="openshift-marketplace/redhat-operators-7c4v7" Dec 01 09:06:32 crc kubenswrapper[4813]: I1201 09:06:32.880485 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9aa179fa-17ce-4b1d-8016-1906584cd152-utilities\") pod \"redhat-operators-7c4v7\" (UID: \"9aa179fa-17ce-4b1d-8016-1906584cd152\") " pod="openshift-marketplace/redhat-operators-7c4v7" Dec 01 09:06:32 crc kubenswrapper[4813]: I1201 09:06:32.881002 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9aa179fa-17ce-4b1d-8016-1906584cd152-catalog-content\") pod \"redhat-operators-7c4v7\" (UID: \"9aa179fa-17ce-4b1d-8016-1906584cd152\") " pod="openshift-marketplace/redhat-operators-7c4v7" Dec 01 09:06:32 crc kubenswrapper[4813]: I1201 09:06:32.881047 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9aa179fa-17ce-4b1d-8016-1906584cd152-utilities\") pod \"redhat-operators-7c4v7\" (UID: \"9aa179fa-17ce-4b1d-8016-1906584cd152\") " pod="openshift-marketplace/redhat-operators-7c4v7" Dec 01 09:06:32 crc kubenswrapper[4813]: I1201 09:06:32.905906 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-79hx5\" (UniqueName: \"kubernetes.io/projected/9aa179fa-17ce-4b1d-8016-1906584cd152-kube-api-access-79hx5\") pod \"redhat-operators-7c4v7\" (UID: \"9aa179fa-17ce-4b1d-8016-1906584cd152\") " pod="openshift-marketplace/redhat-operators-7c4v7" Dec 01 09:06:32 crc kubenswrapper[4813]: I1201 09:06:32.943719 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7c4v7" Dec 01 09:06:33 crc kubenswrapper[4813]: I1201 09:06:33.180615 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-7c4v7"] Dec 01 09:06:33 crc kubenswrapper[4813]: I1201 09:06:33.400242 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7c4v7" event={"ID":"9aa179fa-17ce-4b1d-8016-1906584cd152","Type":"ContainerStarted","Data":"1ee4c84740f068cc7bcdf91b40e5826f00f562441c3bff9e863dc1c9beebb153"} Dec 01 09:06:34 crc kubenswrapper[4813]: I1201 09:06:34.407508 4813 generic.go:334] "Generic (PLEG): container finished" podID="9aa179fa-17ce-4b1d-8016-1906584cd152" containerID="06843aa08687156c47a6bbed8673a02fefbb32cdad6dff80a0ed867bae78ad6f" exitCode=0 Dec 01 09:06:34 crc kubenswrapper[4813]: I1201 09:06:34.407558 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7c4v7" event={"ID":"9aa179fa-17ce-4b1d-8016-1906584cd152","Type":"ContainerDied","Data":"06843aa08687156c47a6bbed8673a02fefbb32cdad6dff80a0ed867bae78ad6f"} Dec 01 09:06:35 crc kubenswrapper[4813]: E1201 09:06:35.125504 4813 secret.go:188] Couldn't get secret glance-kuttl-tests/openstack-config-secret: secret "openstack-config-secret" not found Dec 01 09:06:35 crc kubenswrapper[4813]: E1201 09:06:35.125521 4813 configmap.go:193] Couldn't get configMap glance-kuttl-tests/openstack-config: configmap "openstack-config" not found Dec 01 09:06:35 crc kubenswrapper[4813]: E1201 09:06:35.126189 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/4d0873f6-c7e7-4a9b-bd6e-a0263fd2bc5a-openstack-config-secret podName:4d0873f6-c7e7-4a9b-bd6e-a0263fd2bc5a nodeName:}" failed. No retries permitted until 2025-12-01 09:07:07.126165156 +0000 UTC m=+1611.749086742 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "openstack-config-secret" (UniqueName: "kubernetes.io/secret/4d0873f6-c7e7-4a9b-bd6e-a0263fd2bc5a-openstack-config-secret") pod "openstackclient" (UID: "4d0873f6-c7e7-4a9b-bd6e-a0263fd2bc5a") : secret "openstack-config-secret" not found Dec 01 09:06:35 crc kubenswrapper[4813]: E1201 09:06:35.126322 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/4d0873f6-c7e7-4a9b-bd6e-a0263fd2bc5a-openstack-config podName:4d0873f6-c7e7-4a9b-bd6e-a0263fd2bc5a nodeName:}" failed. No retries permitted until 2025-12-01 09:07:07.12627985 +0000 UTC m=+1611.749201436 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "openstack-config" (UniqueName: "kubernetes.io/configmap/4d0873f6-c7e7-4a9b-bd6e-a0263fd2bc5a-openstack-config") pod "openstackclient" (UID: "4d0873f6-c7e7-4a9b-bd6e-a0263fd2bc5a") : configmap "openstack-config" not found Dec 01 09:06:36 crc kubenswrapper[4813]: I1201 09:06:36.422415 4813 generic.go:334] "Generic (PLEG): container finished" podID="9aa179fa-17ce-4b1d-8016-1906584cd152" containerID="1a9f9ad72327bbcfb0ee5573a26b90b7fe762acce84c1fe4b03fd010ff056d22" exitCode=0 Dec 01 09:06:36 crc kubenswrapper[4813]: I1201 09:06:36.422461 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7c4v7" event={"ID":"9aa179fa-17ce-4b1d-8016-1906584cd152","Type":"ContainerDied","Data":"1a9f9ad72327bbcfb0ee5573a26b90b7fe762acce84c1fe4b03fd010ff056d22"} Dec 01 09:06:38 crc kubenswrapper[4813]: I1201 09:06:38.394864 4813 scope.go:117] "RemoveContainer" containerID="09b49517239f156b4ac50c72e4d184fde70e2daaf7b602adcfc3b42ec90a71e0" Dec 01 09:06:38 crc kubenswrapper[4813]: E1201 09:06:38.395760 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9wn4q_openshift-machine-config-operator(3d423402-e08f-42e9-b60c-6eb9bf2f6d48)\"" pod="openshift-machine-config-operator/machine-config-daemon-9wn4q" podUID="3d423402-e08f-42e9-b60c-6eb9bf2f6d48" Dec 01 09:06:38 crc kubenswrapper[4813]: I1201 09:06:38.457253 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7c4v7" event={"ID":"9aa179fa-17ce-4b1d-8016-1906584cd152","Type":"ContainerStarted","Data":"3b7f589da9863b90b56183c01d49ff06b343ecebc5e0864a0da3916b660d89f6"} Dec 01 09:06:38 crc kubenswrapper[4813]: I1201 09:06:38.486091 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-7c4v7" podStartSLOduration=3.628511084 podStartE2EDuration="6.48604189s" podCreationTimestamp="2025-12-01 09:06:32 +0000 UTC" firstStartedPulling="2025-12-01 09:06:34.409673567 +0000 UTC m=+1579.032595153" lastFinishedPulling="2025-12-01 09:06:37.267204373 +0000 UTC m=+1581.890125959" observedRunningTime="2025-12-01 09:06:38.478884505 +0000 UTC m=+1583.101806101" watchObservedRunningTime="2025-12-01 09:06:38.48604189 +0000 UTC m=+1583.108963476" Dec 01 09:06:39 crc kubenswrapper[4813]: I1201 09:06:39.524865 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368cx94s5_071eb401-2f12-41e4-b52e-312ca4152da5/util/0.log" Dec 01 09:06:39 crc kubenswrapper[4813]: I1201 09:06:39.699802 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368cx94s5_071eb401-2f12-41e4-b52e-312ca4152da5/pull/0.log" Dec 01 09:06:39 crc kubenswrapper[4813]: I1201 09:06:39.718725 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368cx94s5_071eb401-2f12-41e4-b52e-312ca4152da5/util/0.log" Dec 01 09:06:39 crc kubenswrapper[4813]: I1201 09:06:39.726396 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368cx94s5_071eb401-2f12-41e4-b52e-312ca4152da5/pull/0.log" Dec 01 09:06:39 crc kubenswrapper[4813]: I1201 09:06:39.924269 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368cx94s5_071eb401-2f12-41e4-b52e-312ca4152da5/pull/0.log" Dec 01 09:06:39 crc kubenswrapper[4813]: I1201 09:06:39.928173 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368cx94s5_071eb401-2f12-41e4-b52e-312ca4152da5/extract/0.log" Dec 01 09:06:39 crc kubenswrapper[4813]: I1201 09:06:39.969220 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368cx94s5_071eb401-2f12-41e4-b52e-312ca4152da5/util/0.log" Dec 01 09:06:40 crc kubenswrapper[4813]: I1201 09:06:40.301389 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-5998794db6-9lz55_d9af9c23-0392-4caf-b10f-afc96b109c49/manager/0.log" Dec 01 09:06:40 crc kubenswrapper[4813]: I1201 09:06:40.309779 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-index-jn2z5_488d9561-dccc-4b1f-b62f-7af825416efb/registry-server/0.log" Dec 01 09:06:42 crc kubenswrapper[4813]: I1201 09:06:42.944197 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-7c4v7" Dec 01 09:06:42 crc kubenswrapper[4813]: I1201 09:06:42.944276 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-7c4v7" Dec 01 09:06:43 crc kubenswrapper[4813]: I1201 09:06:43.985102 4813 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-7c4v7" podUID="9aa179fa-17ce-4b1d-8016-1906584cd152" containerName="registry-server" probeResult="failure" output=< Dec 01 09:06:43 crc kubenswrapper[4813]: timeout: failed to connect service ":50051" within 1s Dec 01 09:06:43 crc kubenswrapper[4813]: > Dec 01 09:06:52 crc kubenswrapper[4813]: I1201 09:06:52.394094 4813 scope.go:117] "RemoveContainer" containerID="09b49517239f156b4ac50c72e4d184fde70e2daaf7b602adcfc3b42ec90a71e0" Dec 01 09:06:52 crc kubenswrapper[4813]: E1201 09:06:52.395029 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9wn4q_openshift-machine-config-operator(3d423402-e08f-42e9-b60c-6eb9bf2f6d48)\"" pod="openshift-machine-config-operator/machine-config-daemon-9wn4q" podUID="3d423402-e08f-42e9-b60c-6eb9bf2f6d48" Dec 01 09:06:53 crc kubenswrapper[4813]: I1201 09:06:53.014195 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-7c4v7" Dec 01 09:06:53 crc kubenswrapper[4813]: I1201 09:06:53.067748 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-7c4v7" Dec 01 09:06:53 crc kubenswrapper[4813]: I1201 09:06:53.219863 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-fxgbc_2fd313fd-c0b1-4aae-8b3b-2d078a7fc7d0/control-plane-machine-set-operator/0.log" Dec 01 09:06:53 crc kubenswrapper[4813]: I1201 09:06:53.270911 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-7c4v7"] Dec 01 09:06:53 crc kubenswrapper[4813]: I1201 09:06:53.385507 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-n72r5_b35b7145-810b-4945-9819-6d9bb650ec73/kube-rbac-proxy/0.log" Dec 01 09:06:53 crc kubenswrapper[4813]: I1201 09:06:53.485061 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-n72r5_b35b7145-810b-4945-9819-6d9bb650ec73/machine-api-operator/0.log" Dec 01 09:06:54 crc kubenswrapper[4813]: I1201 09:06:54.613081 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-7c4v7" podUID="9aa179fa-17ce-4b1d-8016-1906584cd152" containerName="registry-server" containerID="cri-o://3b7f589da9863b90b56183c01d49ff06b343ecebc5e0864a0da3916b660d89f6" gracePeriod=2 Dec 01 09:06:56 crc kubenswrapper[4813]: I1201 09:06:56.634449 4813 generic.go:334] "Generic (PLEG): container finished" podID="9aa179fa-17ce-4b1d-8016-1906584cd152" containerID="3b7f589da9863b90b56183c01d49ff06b343ecebc5e0864a0da3916b660d89f6" exitCode=0 Dec 01 09:06:56 crc kubenswrapper[4813]: I1201 09:06:56.634516 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7c4v7" event={"ID":"9aa179fa-17ce-4b1d-8016-1906584cd152","Type":"ContainerDied","Data":"3b7f589da9863b90b56183c01d49ff06b343ecebc5e0864a0da3916b660d89f6"} Dec 01 09:06:56 crc kubenswrapper[4813]: I1201 09:06:56.916849 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7c4v7" Dec 01 09:06:57 crc kubenswrapper[4813]: I1201 09:06:57.091487 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9aa179fa-17ce-4b1d-8016-1906584cd152-utilities\") pod \"9aa179fa-17ce-4b1d-8016-1906584cd152\" (UID: \"9aa179fa-17ce-4b1d-8016-1906584cd152\") " Dec 01 09:06:57 crc kubenswrapper[4813]: I1201 09:06:57.091594 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9aa179fa-17ce-4b1d-8016-1906584cd152-catalog-content\") pod \"9aa179fa-17ce-4b1d-8016-1906584cd152\" (UID: \"9aa179fa-17ce-4b1d-8016-1906584cd152\") " Dec 01 09:06:57 crc kubenswrapper[4813]: I1201 09:06:57.091659 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-79hx5\" (UniqueName: \"kubernetes.io/projected/9aa179fa-17ce-4b1d-8016-1906584cd152-kube-api-access-79hx5\") pod \"9aa179fa-17ce-4b1d-8016-1906584cd152\" (UID: \"9aa179fa-17ce-4b1d-8016-1906584cd152\") " Dec 01 09:06:57 crc kubenswrapper[4813]: I1201 09:06:57.092829 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9aa179fa-17ce-4b1d-8016-1906584cd152-utilities" (OuterVolumeSpecName: "utilities") pod "9aa179fa-17ce-4b1d-8016-1906584cd152" (UID: "9aa179fa-17ce-4b1d-8016-1906584cd152"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 09:06:57 crc kubenswrapper[4813]: I1201 09:06:57.098229 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9aa179fa-17ce-4b1d-8016-1906584cd152-kube-api-access-79hx5" (OuterVolumeSpecName: "kube-api-access-79hx5") pod "9aa179fa-17ce-4b1d-8016-1906584cd152" (UID: "9aa179fa-17ce-4b1d-8016-1906584cd152"). InnerVolumeSpecName "kube-api-access-79hx5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:06:57 crc kubenswrapper[4813]: I1201 09:06:57.192923 4813 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9aa179fa-17ce-4b1d-8016-1906584cd152-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 09:06:57 crc kubenswrapper[4813]: I1201 09:06:57.193291 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-79hx5\" (UniqueName: \"kubernetes.io/projected/9aa179fa-17ce-4b1d-8016-1906584cd152-kube-api-access-79hx5\") on node \"crc\" DevicePath \"\"" Dec 01 09:06:57 crc kubenswrapper[4813]: I1201 09:06:57.214793 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9aa179fa-17ce-4b1d-8016-1906584cd152-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9aa179fa-17ce-4b1d-8016-1906584cd152" (UID: "9aa179fa-17ce-4b1d-8016-1906584cd152"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 09:06:57 crc kubenswrapper[4813]: I1201 09:06:57.294398 4813 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9aa179fa-17ce-4b1d-8016-1906584cd152-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 09:06:57 crc kubenswrapper[4813]: I1201 09:06:57.642545 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7c4v7" event={"ID":"9aa179fa-17ce-4b1d-8016-1906584cd152","Type":"ContainerDied","Data":"1ee4c84740f068cc7bcdf91b40e5826f00f562441c3bff9e863dc1c9beebb153"} Dec 01 09:06:57 crc kubenswrapper[4813]: I1201 09:06:57.642610 4813 scope.go:117] "RemoveContainer" containerID="3b7f589da9863b90b56183c01d49ff06b343ecebc5e0864a0da3916b660d89f6" Dec 01 09:06:57 crc kubenswrapper[4813]: I1201 09:06:57.642714 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7c4v7" Dec 01 09:06:57 crc kubenswrapper[4813]: I1201 09:06:57.668885 4813 scope.go:117] "RemoveContainer" containerID="1a9f9ad72327bbcfb0ee5573a26b90b7fe762acce84c1fe4b03fd010ff056d22" Dec 01 09:06:57 crc kubenswrapper[4813]: I1201 09:06:57.676061 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-7c4v7"] Dec 01 09:06:57 crc kubenswrapper[4813]: I1201 09:06:57.683007 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-7c4v7"] Dec 01 09:06:57 crc kubenswrapper[4813]: I1201 09:06:57.694248 4813 scope.go:117] "RemoveContainer" containerID="06843aa08687156c47a6bbed8673a02fefbb32cdad6dff80a0ed867bae78ad6f" Dec 01 09:06:58 crc kubenswrapper[4813]: I1201 09:06:58.401821 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9aa179fa-17ce-4b1d-8016-1906584cd152" path="/var/lib/kubelet/pods/9aa179fa-17ce-4b1d-8016-1906584cd152/volumes" Dec 01 09:07:03 crc kubenswrapper[4813]: I1201 09:07:03.393649 4813 scope.go:117] "RemoveContainer" containerID="09b49517239f156b4ac50c72e4d184fde70e2daaf7b602adcfc3b42ec90a71e0" Dec 01 09:07:03 crc kubenswrapper[4813]: E1201 09:07:03.394548 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9wn4q_openshift-machine-config-operator(3d423402-e08f-42e9-b60c-6eb9bf2f6d48)\"" pod="openshift-machine-config-operator/machine-config-daemon-9wn4q" podUID="3d423402-e08f-42e9-b60c-6eb9bf2f6d48" Dec 01 09:07:07 crc kubenswrapper[4813]: E1201 09:07:07.198714 4813 configmap.go:193] Couldn't get configMap glance-kuttl-tests/openstack-config: configmap "openstack-config" not found Dec 01 09:07:07 crc kubenswrapper[4813]: E1201 09:07:07.199118 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/4d0873f6-c7e7-4a9b-bd6e-a0263fd2bc5a-openstack-config podName:4d0873f6-c7e7-4a9b-bd6e-a0263fd2bc5a nodeName:}" failed. No retries permitted until 2025-12-01 09:08:11.199090581 +0000 UTC m=+1675.822012167 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "openstack-config" (UniqueName: "kubernetes.io/configmap/4d0873f6-c7e7-4a9b-bd6e-a0263fd2bc5a-openstack-config") pod "openstackclient" (UID: "4d0873f6-c7e7-4a9b-bd6e-a0263fd2bc5a") : configmap "openstack-config" not found Dec 01 09:07:07 crc kubenswrapper[4813]: E1201 09:07:07.198755 4813 secret.go:188] Couldn't get secret glance-kuttl-tests/openstack-config-secret: secret "openstack-config-secret" not found Dec 01 09:07:07 crc kubenswrapper[4813]: E1201 09:07:07.199224 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/4d0873f6-c7e7-4a9b-bd6e-a0263fd2bc5a-openstack-config-secret podName:4d0873f6-c7e7-4a9b-bd6e-a0263fd2bc5a nodeName:}" failed. No retries permitted until 2025-12-01 09:08:11.199199264 +0000 UTC m=+1675.822120870 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "openstack-config-secret" (UniqueName: "kubernetes.io/secret/4d0873f6-c7e7-4a9b-bd6e-a0263fd2bc5a-openstack-config-secret") pod "openstackclient" (UID: "4d0873f6-c7e7-4a9b-bd6e-a0263fd2bc5a") : secret "openstack-config-secret" not found Dec 01 09:07:08 crc kubenswrapper[4813]: I1201 09:07:08.960638 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-f8648f98b-m47sh_ce2e13a7-5901-440b-8a7c-6516b77aea04/kube-rbac-proxy/0.log" Dec 01 09:07:08 crc kubenswrapper[4813]: I1201 09:07:08.968214 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-f8648f98b-m47sh_ce2e13a7-5901-440b-8a7c-6516b77aea04/controller/0.log" Dec 01 09:07:09 crc kubenswrapper[4813]: I1201 09:07:09.109370 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-7fcb986d4-rgj8f_97457a58-aba0-4e0a-b812-ef4fd0912116/frr-k8s-webhook-server/0.log" Dec 01 09:07:09 crc kubenswrapper[4813]: I1201 09:07:09.164284 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zplkx_f1cda8d0-a6b1-478a-9f8e-83d1fa070bf5/cp-frr-files/0.log" Dec 01 09:07:09 crc kubenswrapper[4813]: I1201 09:07:09.284928 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zplkx_f1cda8d0-a6b1-478a-9f8e-83d1fa070bf5/cp-reloader/0.log" Dec 01 09:07:09 crc kubenswrapper[4813]: I1201 09:07:09.292278 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zplkx_f1cda8d0-a6b1-478a-9f8e-83d1fa070bf5/cp-frr-files/0.log" Dec 01 09:07:09 crc kubenswrapper[4813]: I1201 09:07:09.323461 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zplkx_f1cda8d0-a6b1-478a-9f8e-83d1fa070bf5/cp-metrics/0.log" Dec 01 09:07:09 crc kubenswrapper[4813]: I1201 09:07:09.380467 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zplkx_f1cda8d0-a6b1-478a-9f8e-83d1fa070bf5/cp-reloader/0.log" Dec 01 09:07:09 crc kubenswrapper[4813]: I1201 09:07:09.500007 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zplkx_f1cda8d0-a6b1-478a-9f8e-83d1fa070bf5/cp-frr-files/0.log" Dec 01 09:07:09 crc kubenswrapper[4813]: I1201 09:07:09.547514 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zplkx_f1cda8d0-a6b1-478a-9f8e-83d1fa070bf5/cp-metrics/0.log" Dec 01 09:07:09 crc kubenswrapper[4813]: I1201 09:07:09.559499 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zplkx_f1cda8d0-a6b1-478a-9f8e-83d1fa070bf5/cp-reloader/0.log" Dec 01 09:07:09 crc kubenswrapper[4813]: I1201 09:07:09.580136 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zplkx_f1cda8d0-a6b1-478a-9f8e-83d1fa070bf5/cp-metrics/0.log" Dec 01 09:07:09 crc kubenswrapper[4813]: I1201 09:07:09.703262 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zplkx_f1cda8d0-a6b1-478a-9f8e-83d1fa070bf5/cp-frr-files/0.log" Dec 01 09:07:09 crc kubenswrapper[4813]: I1201 09:07:09.703426 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zplkx_f1cda8d0-a6b1-478a-9f8e-83d1fa070bf5/cp-metrics/0.log" Dec 01 09:07:09 crc kubenswrapper[4813]: I1201 09:07:09.719688 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zplkx_f1cda8d0-a6b1-478a-9f8e-83d1fa070bf5/cp-reloader/0.log" Dec 01 09:07:09 crc kubenswrapper[4813]: I1201 09:07:09.764490 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zplkx_f1cda8d0-a6b1-478a-9f8e-83d1fa070bf5/controller/0.log" Dec 01 09:07:09 crc kubenswrapper[4813]: I1201 09:07:09.890659 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zplkx_f1cda8d0-a6b1-478a-9f8e-83d1fa070bf5/frr-metrics/0.log" Dec 01 09:07:09 crc kubenswrapper[4813]: I1201 09:07:09.902797 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zplkx_f1cda8d0-a6b1-478a-9f8e-83d1fa070bf5/kube-rbac-proxy/0.log" Dec 01 09:07:09 crc kubenswrapper[4813]: I1201 09:07:09.987775 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zplkx_f1cda8d0-a6b1-478a-9f8e-83d1fa070bf5/kube-rbac-proxy-frr/0.log" Dec 01 09:07:10 crc kubenswrapper[4813]: I1201 09:07:10.100465 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zplkx_f1cda8d0-a6b1-478a-9f8e-83d1fa070bf5/reloader/0.log" Dec 01 09:07:10 crc kubenswrapper[4813]: I1201 09:07:10.179326 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-754cc864cb-4d2tv_4017da37-9dca-4e2d-bab0-a7896bdf3cd5/manager/0.log" Dec 01 09:07:10 crc kubenswrapper[4813]: I1201 09:07:10.356434 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-79d58979df-jnh5k_ed86505b-1475-4ca2-9c22-73ce93b80ab9/webhook-server/0.log" Dec 01 09:07:10 crc kubenswrapper[4813]: I1201 09:07:10.457146 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-227b6_1bf9a820-0d04-43e4-a59b-e6168a6e11fe/kube-rbac-proxy/0.log" Dec 01 09:07:10 crc kubenswrapper[4813]: I1201 09:07:10.548197 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zplkx_f1cda8d0-a6b1-478a-9f8e-83d1fa070bf5/frr/0.log" Dec 01 09:07:10 crc kubenswrapper[4813]: I1201 09:07:10.721268 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-227b6_1bf9a820-0d04-43e4-a59b-e6168a6e11fe/speaker/0.log" Dec 01 09:07:17 crc kubenswrapper[4813]: I1201 09:07:17.394350 4813 scope.go:117] "RemoveContainer" containerID="09b49517239f156b4ac50c72e4d184fde70e2daaf7b602adcfc3b42ec90a71e0" Dec 01 09:07:17 crc kubenswrapper[4813]: E1201 09:07:17.395665 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9wn4q_openshift-machine-config-operator(3d423402-e08f-42e9-b60c-6eb9bf2f6d48)\"" pod="openshift-machine-config-operator/machine-config-daemon-9wn4q" podUID="3d423402-e08f-42e9-b60c-6eb9bf2f6d48" Dec 01 09:07:18 crc kubenswrapper[4813]: I1201 09:07:18.916530 4813 scope.go:117] "RemoveContainer" containerID="6e67bbc17ad4e709ff5bed996102f6ebc16522e36b77fd9d2ecf70d4d72df3a6" Dec 01 09:07:18 crc kubenswrapper[4813]: I1201 09:07:18.940287 4813 scope.go:117] "RemoveContainer" containerID="90721cbc8882119c577ead880faf6ef839f46c442f67c31345fe35b86e07362f" Dec 01 09:07:18 crc kubenswrapper[4813]: I1201 09:07:18.997818 4813 scope.go:117] "RemoveContainer" containerID="999cf5344afaf180b4fdc0b4a14b75b0a4918a25cb64e82b2fda77658ac894f7" Dec 01 09:07:19 crc kubenswrapper[4813]: I1201 09:07:19.022002 4813 scope.go:117] "RemoveContainer" containerID="209e365f6e86b127b33bad81a1c8713bf80558613377028d1df9285b7a9ee8e4" Dec 01 09:07:19 crc kubenswrapper[4813]: I1201 09:07:19.047768 4813 scope.go:117] "RemoveContainer" containerID="65fc01f686054552418e0984b87e8e822fc48eddd50873431e757f3f28f510f4" Dec 01 09:07:19 crc kubenswrapper[4813]: I1201 09:07:19.066619 4813 scope.go:117] "RemoveContainer" containerID="34f7e76fca6a7ec2ea3bd96b3cde58af3d3c87d1a190335eb7651ffc30db835e" Dec 01 09:07:22 crc kubenswrapper[4813]: I1201 09:07:22.581633 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_openstackclient_4d0873f6-c7e7-4a9b-bd6e-a0263fd2bc5a/openstackclient/0.log" Dec 01 09:07:29 crc kubenswrapper[4813]: I1201 09:07:29.394678 4813 scope.go:117] "RemoveContainer" containerID="09b49517239f156b4ac50c72e4d184fde70e2daaf7b602adcfc3b42ec90a71e0" Dec 01 09:07:29 crc kubenswrapper[4813]: E1201 09:07:29.395792 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9wn4q_openshift-machine-config-operator(3d423402-e08f-42e9-b60c-6eb9bf2f6d48)\"" pod="openshift-machine-config-operator/machine-config-daemon-9wn4q" podUID="3d423402-e08f-42e9-b60c-6eb9bf2f6d48" Dec 01 09:07:34 crc kubenswrapper[4813]: I1201 09:07:34.440050 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83qc9k9_d7e5aa34-61f6-476a-a08e-39ce74c1f82a/util/0.log" Dec 01 09:07:34 crc kubenswrapper[4813]: I1201 09:07:34.597795 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83qc9k9_d7e5aa34-61f6-476a-a08e-39ce74c1f82a/pull/0.log" Dec 01 09:07:34 crc kubenswrapper[4813]: I1201 09:07:34.644546 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83qc9k9_d7e5aa34-61f6-476a-a08e-39ce74c1f82a/pull/0.log" Dec 01 09:07:34 crc kubenswrapper[4813]: I1201 09:07:34.798828 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83qc9k9_d7e5aa34-61f6-476a-a08e-39ce74c1f82a/util/0.log" Dec 01 09:07:34 crc kubenswrapper[4813]: I1201 09:07:34.817588 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83qc9k9_d7e5aa34-61f6-476a-a08e-39ce74c1f82a/util/0.log" Dec 01 09:07:34 crc kubenswrapper[4813]: I1201 09:07:34.820790 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83qc9k9_d7e5aa34-61f6-476a-a08e-39ce74c1f82a/extract/0.log" Dec 01 09:07:34 crc kubenswrapper[4813]: I1201 09:07:34.855630 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83qc9k9_d7e5aa34-61f6-476a-a08e-39ce74c1f82a/pull/0.log" Dec 01 09:07:34 crc kubenswrapper[4813]: I1201 09:07:34.999909 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-gwz7r_03720b75-74be-44af-85fd-6b5a55578ca6/extract-utilities/0.log" Dec 01 09:07:35 crc kubenswrapper[4813]: I1201 09:07:35.149747 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-gwz7r_03720b75-74be-44af-85fd-6b5a55578ca6/extract-content/0.log" Dec 01 09:07:35 crc kubenswrapper[4813]: I1201 09:07:35.178854 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-gwz7r_03720b75-74be-44af-85fd-6b5a55578ca6/extract-utilities/0.log" Dec 01 09:07:35 crc kubenswrapper[4813]: I1201 09:07:35.180115 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-gwz7r_03720b75-74be-44af-85fd-6b5a55578ca6/extract-content/0.log" Dec 01 09:07:35 crc kubenswrapper[4813]: I1201 09:07:35.356676 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-gwz7r_03720b75-74be-44af-85fd-6b5a55578ca6/extract-utilities/0.log" Dec 01 09:07:35 crc kubenswrapper[4813]: I1201 09:07:35.430517 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-gwz7r_03720b75-74be-44af-85fd-6b5a55578ca6/extract-content/0.log" Dec 01 09:07:35 crc kubenswrapper[4813]: I1201 09:07:35.732849 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-f87fz_95c21bda-6191-4d26-a3d4-57056be08207/extract-utilities/0.log" Dec 01 09:07:35 crc kubenswrapper[4813]: I1201 09:07:35.741796 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-gwz7r_03720b75-74be-44af-85fd-6b5a55578ca6/registry-server/0.log" Dec 01 09:07:35 crc kubenswrapper[4813]: I1201 09:07:35.846739 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-f87fz_95c21bda-6191-4d26-a3d4-57056be08207/extract-utilities/0.log" Dec 01 09:07:35 crc kubenswrapper[4813]: I1201 09:07:35.880482 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-f87fz_95c21bda-6191-4d26-a3d4-57056be08207/extract-content/0.log" Dec 01 09:07:35 crc kubenswrapper[4813]: I1201 09:07:35.913757 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-f87fz_95c21bda-6191-4d26-a3d4-57056be08207/extract-content/0.log" Dec 01 09:07:36 crc kubenswrapper[4813]: I1201 09:07:36.105098 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-f87fz_95c21bda-6191-4d26-a3d4-57056be08207/extract-content/0.log" Dec 01 09:07:36 crc kubenswrapper[4813]: I1201 09:07:36.109584 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-f87fz_95c21bda-6191-4d26-a3d4-57056be08207/extract-utilities/0.log" Dec 01 09:07:36 crc kubenswrapper[4813]: I1201 09:07:36.338302 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-nmwnn_8d57deaa-d1bd-421f-84c3-8682472f1491/marketplace-operator/0.log" Dec 01 09:07:36 crc kubenswrapper[4813]: I1201 09:07:36.421256 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-dclft_5b4cbc77-2764-4d42-aa64-e040598935b7/extract-utilities/0.log" Dec 01 09:07:36 crc kubenswrapper[4813]: I1201 09:07:36.615259 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-dclft_5b4cbc77-2764-4d42-aa64-e040598935b7/extract-content/0.log" Dec 01 09:07:36 crc kubenswrapper[4813]: I1201 09:07:36.653196 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-dclft_5b4cbc77-2764-4d42-aa64-e040598935b7/extract-utilities/0.log" Dec 01 09:07:36 crc kubenswrapper[4813]: I1201 09:07:36.663691 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-f87fz_95c21bda-6191-4d26-a3d4-57056be08207/registry-server/0.log" Dec 01 09:07:36 crc kubenswrapper[4813]: I1201 09:07:36.719198 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-dclft_5b4cbc77-2764-4d42-aa64-e040598935b7/extract-content/0.log" Dec 01 09:07:36 crc kubenswrapper[4813]: I1201 09:07:36.877094 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-dclft_5b4cbc77-2764-4d42-aa64-e040598935b7/extract-utilities/0.log" Dec 01 09:07:36 crc kubenswrapper[4813]: I1201 09:07:36.886989 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-dclft_5b4cbc77-2764-4d42-aa64-e040598935b7/extract-content/0.log" Dec 01 09:07:37 crc kubenswrapper[4813]: I1201 09:07:37.006901 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-dclft_5b4cbc77-2764-4d42-aa64-e040598935b7/registry-server/0.log" Dec 01 09:07:37 crc kubenswrapper[4813]: I1201 09:07:37.052446 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-h5ffj_9f704b3a-26b9-43b8-8598-02e769dfbb87/extract-utilities/0.log" Dec 01 09:07:37 crc kubenswrapper[4813]: I1201 09:07:37.231833 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-h5ffj_9f704b3a-26b9-43b8-8598-02e769dfbb87/extract-utilities/0.log" Dec 01 09:07:37 crc kubenswrapper[4813]: I1201 09:07:37.259781 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-h5ffj_9f704b3a-26b9-43b8-8598-02e769dfbb87/extract-content/0.log" Dec 01 09:07:37 crc kubenswrapper[4813]: I1201 09:07:37.294352 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-h5ffj_9f704b3a-26b9-43b8-8598-02e769dfbb87/extract-content/0.log" Dec 01 09:07:37 crc kubenswrapper[4813]: I1201 09:07:37.609653 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-h5ffj_9f704b3a-26b9-43b8-8598-02e769dfbb87/extract-content/0.log" Dec 01 09:07:37 crc kubenswrapper[4813]: I1201 09:07:37.627825 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-h5ffj_9f704b3a-26b9-43b8-8598-02e769dfbb87/extract-utilities/0.log" Dec 01 09:07:37 crc kubenswrapper[4813]: I1201 09:07:37.819832 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-h5ffj_9f704b3a-26b9-43b8-8598-02e769dfbb87/registry-server/0.log" Dec 01 09:07:40 crc kubenswrapper[4813]: I1201 09:07:40.394074 4813 scope.go:117] "RemoveContainer" containerID="09b49517239f156b4ac50c72e4d184fde70e2daaf7b602adcfc3b42ec90a71e0" Dec 01 09:07:40 crc kubenswrapper[4813]: E1201 09:07:40.394429 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9wn4q_openshift-machine-config-operator(3d423402-e08f-42e9-b60c-6eb9bf2f6d48)\"" pod="openshift-machine-config-operator/machine-config-daemon-9wn4q" podUID="3d423402-e08f-42e9-b60c-6eb9bf2f6d48" Dec 01 09:07:54 crc kubenswrapper[4813]: I1201 09:07:54.393279 4813 scope.go:117] "RemoveContainer" containerID="09b49517239f156b4ac50c72e4d184fde70e2daaf7b602adcfc3b42ec90a71e0" Dec 01 09:07:54 crc kubenswrapper[4813]: E1201 09:07:54.394267 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9wn4q_openshift-machine-config-operator(3d423402-e08f-42e9-b60c-6eb9bf2f6d48)\"" pod="openshift-machine-config-operator/machine-config-daemon-9wn4q" podUID="3d423402-e08f-42e9-b60c-6eb9bf2f6d48" Dec 01 09:08:07 crc kubenswrapper[4813]: I1201 09:08:07.393475 4813 scope.go:117] "RemoveContainer" containerID="09b49517239f156b4ac50c72e4d184fde70e2daaf7b602adcfc3b42ec90a71e0" Dec 01 09:08:07 crc kubenswrapper[4813]: E1201 09:08:07.396933 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9wn4q_openshift-machine-config-operator(3d423402-e08f-42e9-b60c-6eb9bf2f6d48)\"" pod="openshift-machine-config-operator/machine-config-daemon-9wn4q" podUID="3d423402-e08f-42e9-b60c-6eb9bf2f6d48" Dec 01 09:08:11 crc kubenswrapper[4813]: E1201 09:08:11.257827 4813 secret.go:188] Couldn't get secret glance-kuttl-tests/openstack-config-secret: secret "openstack-config-secret" not found Dec 01 09:08:11 crc kubenswrapper[4813]: E1201 09:08:11.258313 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/4d0873f6-c7e7-4a9b-bd6e-a0263fd2bc5a-openstack-config-secret podName:4d0873f6-c7e7-4a9b-bd6e-a0263fd2bc5a nodeName:}" failed. No retries permitted until 2025-12-01 09:10:13.258261467 +0000 UTC m=+1797.881183053 (durationBeforeRetry 2m2s). Error: MountVolume.SetUp failed for volume "openstack-config-secret" (UniqueName: "kubernetes.io/secret/4d0873f6-c7e7-4a9b-bd6e-a0263fd2bc5a-openstack-config-secret") pod "openstackclient" (UID: "4d0873f6-c7e7-4a9b-bd6e-a0263fd2bc5a") : secret "openstack-config-secret" not found Dec 01 09:08:11 crc kubenswrapper[4813]: E1201 09:08:11.258049 4813 configmap.go:193] Couldn't get configMap glance-kuttl-tests/openstack-config: configmap "openstack-config" not found Dec 01 09:08:11 crc kubenswrapper[4813]: E1201 09:08:11.258487 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/4d0873f6-c7e7-4a9b-bd6e-a0263fd2bc5a-openstack-config podName:4d0873f6-c7e7-4a9b-bd6e-a0263fd2bc5a nodeName:}" failed. No retries permitted until 2025-12-01 09:10:13.258459663 +0000 UTC m=+1797.881381289 (durationBeforeRetry 2m2s). Error: MountVolume.SetUp failed for volume "openstack-config" (UniqueName: "kubernetes.io/configmap/4d0873f6-c7e7-4a9b-bd6e-a0263fd2bc5a-openstack-config") pod "openstackclient" (UID: "4d0873f6-c7e7-4a9b-bd6e-a0263fd2bc5a") : configmap "openstack-config" not found Dec 01 09:08:18 crc kubenswrapper[4813]: I1201 09:08:18.398562 4813 scope.go:117] "RemoveContainer" containerID="09b49517239f156b4ac50c72e4d184fde70e2daaf7b602adcfc3b42ec90a71e0" Dec 01 09:08:18 crc kubenswrapper[4813]: E1201 09:08:18.399613 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9wn4q_openshift-machine-config-operator(3d423402-e08f-42e9-b60c-6eb9bf2f6d48)\"" pod="openshift-machine-config-operator/machine-config-daemon-9wn4q" podUID="3d423402-e08f-42e9-b60c-6eb9bf2f6d48" Dec 01 09:08:19 crc kubenswrapper[4813]: I1201 09:08:19.207801 4813 scope.go:117] "RemoveContainer" containerID="39ba0adee13915c1669441022a9cf3d86928a608ad457ed24bc0ab7ca527f640" Dec 01 09:08:19 crc kubenswrapper[4813]: I1201 09:08:19.256907 4813 scope.go:117] "RemoveContainer" containerID="23b69e3281f74625e3c7674bf4a4b8ced6ff8e237e045d429a4b21c032a4c42c" Dec 01 09:08:19 crc kubenswrapper[4813]: I1201 09:08:19.281457 4813 scope.go:117] "RemoveContainer" containerID="e52f3b09c24264a2aaf36a533f3da9ad43d8f094026fa71de0903de56e5e8ef6" Dec 01 09:08:19 crc kubenswrapper[4813]: I1201 09:08:19.336707 4813 scope.go:117] "RemoveContainer" containerID="cd655f585ece329a36be0668cdd4a160b0949ccdd5582302eda8d26fa6c29547" Dec 01 09:08:19 crc kubenswrapper[4813]: I1201 09:08:19.442224 4813 scope.go:117] "RemoveContainer" containerID="a2ff5f4871bec4b409097a90d5929c75feb2f96c17f8cf1b75cbfa57f68a9374" Dec 01 09:08:30 crc kubenswrapper[4813]: I1201 09:08:30.396575 4813 scope.go:117] "RemoveContainer" containerID="09b49517239f156b4ac50c72e4d184fde70e2daaf7b602adcfc3b42ec90a71e0" Dec 01 09:08:30 crc kubenswrapper[4813]: E1201 09:08:30.398605 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9wn4q_openshift-machine-config-operator(3d423402-e08f-42e9-b60c-6eb9bf2f6d48)\"" pod="openshift-machine-config-operator/machine-config-daemon-9wn4q" podUID="3d423402-e08f-42e9-b60c-6eb9bf2f6d48" Dec 01 09:08:43 crc kubenswrapper[4813]: I1201 09:08:43.461406 4813 generic.go:334] "Generic (PLEG): container finished" podID="152332db-f02a-4365-9bc2-cf74ab20c6f2" containerID="a0b97e73e8cb1a9533b45f04d39bb84d132e25c893ea551fe83995b9979892e7" exitCode=0 Dec 01 09:08:43 crc kubenswrapper[4813]: I1201 09:08:43.461493 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-ffr6x/must-gather-2b9p6" event={"ID":"152332db-f02a-4365-9bc2-cf74ab20c6f2","Type":"ContainerDied","Data":"a0b97e73e8cb1a9533b45f04d39bb84d132e25c893ea551fe83995b9979892e7"} Dec 01 09:08:43 crc kubenswrapper[4813]: I1201 09:08:43.463466 4813 scope.go:117] "RemoveContainer" containerID="a0b97e73e8cb1a9533b45f04d39bb84d132e25c893ea551fe83995b9979892e7" Dec 01 09:08:44 crc kubenswrapper[4813]: I1201 09:08:44.148209 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-ffr6x_must-gather-2b9p6_152332db-f02a-4365-9bc2-cf74ab20c6f2/gather/0.log" Dec 01 09:08:45 crc kubenswrapper[4813]: I1201 09:08:45.393670 4813 scope.go:117] "RemoveContainer" containerID="09b49517239f156b4ac50c72e4d184fde70e2daaf7b602adcfc3b42ec90a71e0" Dec 01 09:08:45 crc kubenswrapper[4813]: E1201 09:08:45.394296 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9wn4q_openshift-machine-config-operator(3d423402-e08f-42e9-b60c-6eb9bf2f6d48)\"" pod="openshift-machine-config-operator/machine-config-daemon-9wn4q" podUID="3d423402-e08f-42e9-b60c-6eb9bf2f6d48" Dec 01 09:08:51 crc kubenswrapper[4813]: I1201 09:08:51.083269 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-ffr6x/must-gather-2b9p6"] Dec 01 09:08:51 crc kubenswrapper[4813]: I1201 09:08:51.084078 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-ffr6x/must-gather-2b9p6" podUID="152332db-f02a-4365-9bc2-cf74ab20c6f2" containerName="copy" containerID="cri-o://f167f42106c1e02babf66072ed2ece2eb0e7d9b99a5d90d4ebb96a87e5664cb5" gracePeriod=2 Dec 01 09:08:51 crc kubenswrapper[4813]: I1201 09:08:51.092546 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-ffr6x/must-gather-2b9p6"] Dec 01 09:08:51 crc kubenswrapper[4813]: I1201 09:08:51.422902 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-ffr6x_must-gather-2b9p6_152332db-f02a-4365-9bc2-cf74ab20c6f2/copy/0.log" Dec 01 09:08:51 crc kubenswrapper[4813]: I1201 09:08:51.423571 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-ffr6x/must-gather-2b9p6" Dec 01 09:08:51 crc kubenswrapper[4813]: I1201 09:08:51.514346 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-ffr6x_must-gather-2b9p6_152332db-f02a-4365-9bc2-cf74ab20c6f2/copy/0.log" Dec 01 09:08:51 crc kubenswrapper[4813]: I1201 09:08:51.514838 4813 generic.go:334] "Generic (PLEG): container finished" podID="152332db-f02a-4365-9bc2-cf74ab20c6f2" containerID="f167f42106c1e02babf66072ed2ece2eb0e7d9b99a5d90d4ebb96a87e5664cb5" exitCode=143 Dec 01 09:08:51 crc kubenswrapper[4813]: I1201 09:08:51.514900 4813 scope.go:117] "RemoveContainer" containerID="f167f42106c1e02babf66072ed2ece2eb0e7d9b99a5d90d4ebb96a87e5664cb5" Dec 01 09:08:51 crc kubenswrapper[4813]: I1201 09:08:51.515127 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-ffr6x/must-gather-2b9p6" Dec 01 09:08:51 crc kubenswrapper[4813]: I1201 09:08:51.523437 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w8nw9\" (UniqueName: \"kubernetes.io/projected/152332db-f02a-4365-9bc2-cf74ab20c6f2-kube-api-access-w8nw9\") pod \"152332db-f02a-4365-9bc2-cf74ab20c6f2\" (UID: \"152332db-f02a-4365-9bc2-cf74ab20c6f2\") " Dec 01 09:08:51 crc kubenswrapper[4813]: I1201 09:08:51.523484 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/152332db-f02a-4365-9bc2-cf74ab20c6f2-must-gather-output\") pod \"152332db-f02a-4365-9bc2-cf74ab20c6f2\" (UID: \"152332db-f02a-4365-9bc2-cf74ab20c6f2\") " Dec 01 09:08:51 crc kubenswrapper[4813]: I1201 09:08:51.533658 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/152332db-f02a-4365-9bc2-cf74ab20c6f2-kube-api-access-w8nw9" (OuterVolumeSpecName: "kube-api-access-w8nw9") pod "152332db-f02a-4365-9bc2-cf74ab20c6f2" (UID: "152332db-f02a-4365-9bc2-cf74ab20c6f2"). InnerVolumeSpecName "kube-api-access-w8nw9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:08:51 crc kubenswrapper[4813]: I1201 09:08:51.536146 4813 scope.go:117] "RemoveContainer" containerID="a0b97e73e8cb1a9533b45f04d39bb84d132e25c893ea551fe83995b9979892e7" Dec 01 09:08:51 crc kubenswrapper[4813]: I1201 09:08:51.615861 4813 scope.go:117] "RemoveContainer" containerID="f167f42106c1e02babf66072ed2ece2eb0e7d9b99a5d90d4ebb96a87e5664cb5" Dec 01 09:08:51 crc kubenswrapper[4813]: E1201 09:08:51.618213 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f167f42106c1e02babf66072ed2ece2eb0e7d9b99a5d90d4ebb96a87e5664cb5\": container with ID starting with f167f42106c1e02babf66072ed2ece2eb0e7d9b99a5d90d4ebb96a87e5664cb5 not found: ID does not exist" containerID="f167f42106c1e02babf66072ed2ece2eb0e7d9b99a5d90d4ebb96a87e5664cb5" Dec 01 09:08:51 crc kubenswrapper[4813]: I1201 09:08:51.618290 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f167f42106c1e02babf66072ed2ece2eb0e7d9b99a5d90d4ebb96a87e5664cb5"} err="failed to get container status \"f167f42106c1e02babf66072ed2ece2eb0e7d9b99a5d90d4ebb96a87e5664cb5\": rpc error: code = NotFound desc = could not find container \"f167f42106c1e02babf66072ed2ece2eb0e7d9b99a5d90d4ebb96a87e5664cb5\": container with ID starting with f167f42106c1e02babf66072ed2ece2eb0e7d9b99a5d90d4ebb96a87e5664cb5 not found: ID does not exist" Dec 01 09:08:51 crc kubenswrapper[4813]: I1201 09:08:51.618325 4813 scope.go:117] "RemoveContainer" containerID="a0b97e73e8cb1a9533b45f04d39bb84d132e25c893ea551fe83995b9979892e7" Dec 01 09:08:51 crc kubenswrapper[4813]: E1201 09:08:51.620164 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a0b97e73e8cb1a9533b45f04d39bb84d132e25c893ea551fe83995b9979892e7\": container with ID starting with a0b97e73e8cb1a9533b45f04d39bb84d132e25c893ea551fe83995b9979892e7 not found: ID does not exist" containerID="a0b97e73e8cb1a9533b45f04d39bb84d132e25c893ea551fe83995b9979892e7" Dec 01 09:08:51 crc kubenswrapper[4813]: I1201 09:08:51.620209 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a0b97e73e8cb1a9533b45f04d39bb84d132e25c893ea551fe83995b9979892e7"} err="failed to get container status \"a0b97e73e8cb1a9533b45f04d39bb84d132e25c893ea551fe83995b9979892e7\": rpc error: code = NotFound desc = could not find container \"a0b97e73e8cb1a9533b45f04d39bb84d132e25c893ea551fe83995b9979892e7\": container with ID starting with a0b97e73e8cb1a9533b45f04d39bb84d132e25c893ea551fe83995b9979892e7 not found: ID does not exist" Dec 01 09:08:51 crc kubenswrapper[4813]: I1201 09:08:51.624999 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w8nw9\" (UniqueName: \"kubernetes.io/projected/152332db-f02a-4365-9bc2-cf74ab20c6f2-kube-api-access-w8nw9\") on node \"crc\" DevicePath \"\"" Dec 01 09:08:51 crc kubenswrapper[4813]: I1201 09:08:51.630163 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/152332db-f02a-4365-9bc2-cf74ab20c6f2-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "152332db-f02a-4365-9bc2-cf74ab20c6f2" (UID: "152332db-f02a-4365-9bc2-cf74ab20c6f2"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 09:08:51 crc kubenswrapper[4813]: I1201 09:08:51.726238 4813 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/152332db-f02a-4365-9bc2-cf74ab20c6f2-must-gather-output\") on node \"crc\" DevicePath \"\"" Dec 01 09:08:52 crc kubenswrapper[4813]: I1201 09:08:52.401619 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="152332db-f02a-4365-9bc2-cf74ab20c6f2" path="/var/lib/kubelet/pods/152332db-f02a-4365-9bc2-cf74ab20c6f2/volumes" Dec 01 09:08:57 crc kubenswrapper[4813]: I1201 09:08:57.393664 4813 scope.go:117] "RemoveContainer" containerID="09b49517239f156b4ac50c72e4d184fde70e2daaf7b602adcfc3b42ec90a71e0" Dec 01 09:08:57 crc kubenswrapper[4813]: E1201 09:08:57.394244 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9wn4q_openshift-machine-config-operator(3d423402-e08f-42e9-b60c-6eb9bf2f6d48)\"" pod="openshift-machine-config-operator/machine-config-daemon-9wn4q" podUID="3d423402-e08f-42e9-b60c-6eb9bf2f6d48" Dec 01 09:09:10 crc kubenswrapper[4813]: I1201 09:09:10.394777 4813 scope.go:117] "RemoveContainer" containerID="09b49517239f156b4ac50c72e4d184fde70e2daaf7b602adcfc3b42ec90a71e0" Dec 01 09:09:10 crc kubenswrapper[4813]: E1201 09:09:10.396079 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9wn4q_openshift-machine-config-operator(3d423402-e08f-42e9-b60c-6eb9bf2f6d48)\"" pod="openshift-machine-config-operator/machine-config-daemon-9wn4q" podUID="3d423402-e08f-42e9-b60c-6eb9bf2f6d48" Dec 01 09:09:19 crc kubenswrapper[4813]: I1201 09:09:19.509803 4813 scope.go:117] "RemoveContainer" containerID="b1f1e4333cca2a8d91aa2d3416981ba00a00c8bd2ea576ab629d745b8760ca3f" Dec 01 09:09:19 crc kubenswrapper[4813]: I1201 09:09:19.538188 4813 scope.go:117] "RemoveContainer" containerID="b13bd94f99ec83c1cf6a08e3272145e5d16fcbaa439ac316c8433d38d61ccf67" Dec 01 09:09:19 crc kubenswrapper[4813]: I1201 09:09:19.594151 4813 scope.go:117] "RemoveContainer" containerID="b4e5ecf79bdec70cdef568e7ab234ac0de6101613bf4e190c96d980bed839811" Dec 01 09:09:19 crc kubenswrapper[4813]: I1201 09:09:19.611925 4813 scope.go:117] "RemoveContainer" containerID="8250ba203c774f9ed961312dac84a6a807f187353d50fe2580fe5ee85b4bd6f1" Dec 01 09:09:25 crc kubenswrapper[4813]: I1201 09:09:25.393870 4813 scope.go:117] "RemoveContainer" containerID="09b49517239f156b4ac50c72e4d184fde70e2daaf7b602adcfc3b42ec90a71e0" Dec 01 09:09:25 crc kubenswrapper[4813]: E1201 09:09:25.394649 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9wn4q_openshift-machine-config-operator(3d423402-e08f-42e9-b60c-6eb9bf2f6d48)\"" pod="openshift-machine-config-operator/machine-config-daemon-9wn4q" podUID="3d423402-e08f-42e9-b60c-6eb9bf2f6d48" Dec 01 09:09:38 crc kubenswrapper[4813]: I1201 09:09:38.393683 4813 scope.go:117] "RemoveContainer" containerID="09b49517239f156b4ac50c72e4d184fde70e2daaf7b602adcfc3b42ec90a71e0" Dec 01 09:09:38 crc kubenswrapper[4813]: E1201 09:09:38.394812 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9wn4q_openshift-machine-config-operator(3d423402-e08f-42e9-b60c-6eb9bf2f6d48)\"" pod="openshift-machine-config-operator/machine-config-daemon-9wn4q" podUID="3d423402-e08f-42e9-b60c-6eb9bf2f6d48" Dec 01 09:09:51 crc kubenswrapper[4813]: I1201 09:09:51.394291 4813 scope.go:117] "RemoveContainer" containerID="09b49517239f156b4ac50c72e4d184fde70e2daaf7b602adcfc3b42ec90a71e0" Dec 01 09:09:51 crc kubenswrapper[4813]: E1201 09:09:51.395136 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9wn4q_openshift-machine-config-operator(3d423402-e08f-42e9-b60c-6eb9bf2f6d48)\"" pod="openshift-machine-config-operator/machine-config-daemon-9wn4q" podUID="3d423402-e08f-42e9-b60c-6eb9bf2f6d48" Dec 01 09:10:06 crc kubenswrapper[4813]: I1201 09:10:06.397544 4813 scope.go:117] "RemoveContainer" containerID="09b49517239f156b4ac50c72e4d184fde70e2daaf7b602adcfc3b42ec90a71e0" Dec 01 09:10:06 crc kubenswrapper[4813]: E1201 09:10:06.398371 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9wn4q_openshift-machine-config-operator(3d423402-e08f-42e9-b60c-6eb9bf2f6d48)\"" pod="openshift-machine-config-operator/machine-config-daemon-9wn4q" podUID="3d423402-e08f-42e9-b60c-6eb9bf2f6d48" Dec 01 09:10:13 crc kubenswrapper[4813]: E1201 09:10:13.275495 4813 configmap.go:193] Couldn't get configMap glance-kuttl-tests/openstack-config: configmap "openstack-config" not found Dec 01 09:10:13 crc kubenswrapper[4813]: E1201 09:10:13.275527 4813 secret.go:188] Couldn't get secret glance-kuttl-tests/openstack-config-secret: secret "openstack-config-secret" not found Dec 01 09:10:13 crc kubenswrapper[4813]: E1201 09:10:13.276273 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/4d0873f6-c7e7-4a9b-bd6e-a0263fd2bc5a-openstack-config podName:4d0873f6-c7e7-4a9b-bd6e-a0263fd2bc5a nodeName:}" failed. No retries permitted until 2025-12-01 09:12:15.276191162 +0000 UTC m=+1919.899112798 (durationBeforeRetry 2m2s). Error: MountVolume.SetUp failed for volume "openstack-config" (UniqueName: "kubernetes.io/configmap/4d0873f6-c7e7-4a9b-bd6e-a0263fd2bc5a-openstack-config") pod "openstackclient" (UID: "4d0873f6-c7e7-4a9b-bd6e-a0263fd2bc5a") : configmap "openstack-config" not found Dec 01 09:10:13 crc kubenswrapper[4813]: E1201 09:10:13.276357 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/4d0873f6-c7e7-4a9b-bd6e-a0263fd2bc5a-openstack-config-secret podName:4d0873f6-c7e7-4a9b-bd6e-a0263fd2bc5a nodeName:}" failed. No retries permitted until 2025-12-01 09:12:15.276324866 +0000 UTC m=+1919.899246482 (durationBeforeRetry 2m2s). Error: MountVolume.SetUp failed for volume "openstack-config-secret" (UniqueName: "kubernetes.io/secret/4d0873f6-c7e7-4a9b-bd6e-a0263fd2bc5a-openstack-config-secret") pod "openstackclient" (UID: "4d0873f6-c7e7-4a9b-bd6e-a0263fd2bc5a") : secret "openstack-config-secret" not found Dec 01 09:10:19 crc kubenswrapper[4813]: I1201 09:10:19.688460 4813 scope.go:117] "RemoveContainer" containerID="bdc319325263c4f53d0f49b476c0d63f087082dd8eda5e1e4f5f291bdaaffad5" Dec 01 09:10:19 crc kubenswrapper[4813]: I1201 09:10:19.720937 4813 scope.go:117] "RemoveContainer" containerID="19c723c9091ce35ec9683cf2c52fb493beafe751cf1fe8498cf83f234b727873" Dec 01 09:10:19 crc kubenswrapper[4813]: I1201 09:10:19.787751 4813 scope.go:117] "RemoveContainer" containerID="e560ea1af90413a541ed0e779818533d49870856d37b3a88dfb8b16e0583f993" Dec 01 09:10:19 crc kubenswrapper[4813]: I1201 09:10:19.807805 4813 scope.go:117] "RemoveContainer" containerID="86514afa63c5089fdc14f9b082809d402f7390c825756c51d1a5199647657b3d" Dec 01 09:10:19 crc kubenswrapper[4813]: I1201 09:10:19.833175 4813 scope.go:117] "RemoveContainer" containerID="1c35642330f47379d43b8119514c1a7daa3e3e9f0dd54bef1d33b46a9d211985" Dec 01 09:10:19 crc kubenswrapper[4813]: I1201 09:10:19.851469 4813 scope.go:117] "RemoveContainer" containerID="158ce31cbd9e2c3e30d880e6f28ca453fb84f4568471c7d8b694016d66c31583" Dec 01 09:10:20 crc kubenswrapper[4813]: I1201 09:10:20.394128 4813 scope.go:117] "RemoveContainer" containerID="09b49517239f156b4ac50c72e4d184fde70e2daaf7b602adcfc3b42ec90a71e0" Dec 01 09:10:21 crc kubenswrapper[4813]: I1201 09:10:21.320506 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-9wn4q" event={"ID":"3d423402-e08f-42e9-b60c-6eb9bf2f6d48","Type":"ContainerStarted","Data":"47c278315d64e3056cb6f536af9317d9d67616d501c03a3c61eb1b5cb51b2405"} Dec 01 09:11:12 crc kubenswrapper[4813]: I1201 09:11:12.320446 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-sd6gm/must-gather-mpdwz"] Dec 01 09:11:12 crc kubenswrapper[4813]: E1201 09:11:12.321589 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9aa179fa-17ce-4b1d-8016-1906584cd152" containerName="extract-utilities" Dec 01 09:11:12 crc kubenswrapper[4813]: I1201 09:11:12.321630 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="9aa179fa-17ce-4b1d-8016-1906584cd152" containerName="extract-utilities" Dec 01 09:11:12 crc kubenswrapper[4813]: E1201 09:11:12.321662 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="152332db-f02a-4365-9bc2-cf74ab20c6f2" containerName="copy" Dec 01 09:11:12 crc kubenswrapper[4813]: I1201 09:11:12.321670 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="152332db-f02a-4365-9bc2-cf74ab20c6f2" containerName="copy" Dec 01 09:11:12 crc kubenswrapper[4813]: E1201 09:11:12.321693 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="152332db-f02a-4365-9bc2-cf74ab20c6f2" containerName="gather" Dec 01 09:11:12 crc kubenswrapper[4813]: I1201 09:11:12.321702 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="152332db-f02a-4365-9bc2-cf74ab20c6f2" containerName="gather" Dec 01 09:11:12 crc kubenswrapper[4813]: E1201 09:11:12.321718 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9aa179fa-17ce-4b1d-8016-1906584cd152" containerName="extract-content" Dec 01 09:11:12 crc kubenswrapper[4813]: I1201 09:11:12.321726 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="9aa179fa-17ce-4b1d-8016-1906584cd152" containerName="extract-content" Dec 01 09:11:12 crc kubenswrapper[4813]: E1201 09:11:12.321741 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9aa179fa-17ce-4b1d-8016-1906584cd152" containerName="registry-server" Dec 01 09:11:12 crc kubenswrapper[4813]: I1201 09:11:12.321748 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="9aa179fa-17ce-4b1d-8016-1906584cd152" containerName="registry-server" Dec 01 09:11:12 crc kubenswrapper[4813]: I1201 09:11:12.321937 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="152332db-f02a-4365-9bc2-cf74ab20c6f2" containerName="gather" Dec 01 09:11:12 crc kubenswrapper[4813]: I1201 09:11:12.321986 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="152332db-f02a-4365-9bc2-cf74ab20c6f2" containerName="copy" Dec 01 09:11:12 crc kubenswrapper[4813]: I1201 09:11:12.322000 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="9aa179fa-17ce-4b1d-8016-1906584cd152" containerName="registry-server" Dec 01 09:11:12 crc kubenswrapper[4813]: I1201 09:11:12.323482 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-sd6gm/must-gather-mpdwz" Dec 01 09:11:12 crc kubenswrapper[4813]: I1201 09:11:12.330353 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-sd6gm"/"default-dockercfg-shqh5" Dec 01 09:11:12 crc kubenswrapper[4813]: I1201 09:11:12.330844 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-sd6gm"/"kube-root-ca.crt" Dec 01 09:11:12 crc kubenswrapper[4813]: I1201 09:11:12.330844 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-sd6gm"/"openshift-service-ca.crt" Dec 01 09:11:12 crc kubenswrapper[4813]: I1201 09:11:12.338508 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-sd6gm/must-gather-mpdwz"] Dec 01 09:11:12 crc kubenswrapper[4813]: I1201 09:11:12.518450 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/fa60e61e-2a74-4649-9956-eee9c0bb15cc-must-gather-output\") pod \"must-gather-mpdwz\" (UID: \"fa60e61e-2a74-4649-9956-eee9c0bb15cc\") " pod="openshift-must-gather-sd6gm/must-gather-mpdwz" Dec 01 09:11:12 crc kubenswrapper[4813]: I1201 09:11:12.519043 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bl9dc\" (UniqueName: \"kubernetes.io/projected/fa60e61e-2a74-4649-9956-eee9c0bb15cc-kube-api-access-bl9dc\") pod \"must-gather-mpdwz\" (UID: \"fa60e61e-2a74-4649-9956-eee9c0bb15cc\") " pod="openshift-must-gather-sd6gm/must-gather-mpdwz" Dec 01 09:11:12 crc kubenswrapper[4813]: I1201 09:11:12.621217 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/fa60e61e-2a74-4649-9956-eee9c0bb15cc-must-gather-output\") pod \"must-gather-mpdwz\" (UID: \"fa60e61e-2a74-4649-9956-eee9c0bb15cc\") " pod="openshift-must-gather-sd6gm/must-gather-mpdwz" Dec 01 09:11:12 crc kubenswrapper[4813]: I1201 09:11:12.621429 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bl9dc\" (UniqueName: \"kubernetes.io/projected/fa60e61e-2a74-4649-9956-eee9c0bb15cc-kube-api-access-bl9dc\") pod \"must-gather-mpdwz\" (UID: \"fa60e61e-2a74-4649-9956-eee9c0bb15cc\") " pod="openshift-must-gather-sd6gm/must-gather-mpdwz" Dec 01 09:11:12 crc kubenswrapper[4813]: I1201 09:11:12.621813 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/fa60e61e-2a74-4649-9956-eee9c0bb15cc-must-gather-output\") pod \"must-gather-mpdwz\" (UID: \"fa60e61e-2a74-4649-9956-eee9c0bb15cc\") " pod="openshift-must-gather-sd6gm/must-gather-mpdwz" Dec 01 09:11:12 crc kubenswrapper[4813]: I1201 09:11:12.641257 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bl9dc\" (UniqueName: \"kubernetes.io/projected/fa60e61e-2a74-4649-9956-eee9c0bb15cc-kube-api-access-bl9dc\") pod \"must-gather-mpdwz\" (UID: \"fa60e61e-2a74-4649-9956-eee9c0bb15cc\") " pod="openshift-must-gather-sd6gm/must-gather-mpdwz" Dec 01 09:11:12 crc kubenswrapper[4813]: I1201 09:11:12.643883 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-sd6gm/must-gather-mpdwz" Dec 01 09:11:12 crc kubenswrapper[4813]: I1201 09:11:12.960662 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-sd6gm/must-gather-mpdwz"] Dec 01 09:11:13 crc kubenswrapper[4813]: I1201 09:11:13.800823 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-sd6gm/must-gather-mpdwz" event={"ID":"fa60e61e-2a74-4649-9956-eee9c0bb15cc","Type":"ContainerStarted","Data":"8772a396422cc9a7ceafad873cd441791ce54573e060515d4029cbc636dfa146"} Dec 01 09:11:13 crc kubenswrapper[4813]: I1201 09:11:13.801196 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-sd6gm/must-gather-mpdwz" event={"ID":"fa60e61e-2a74-4649-9956-eee9c0bb15cc","Type":"ContainerStarted","Data":"72b7a4f9f471289de5a4eb203f9d31da96ee153933bc45286a303d28a09fd295"} Dec 01 09:11:13 crc kubenswrapper[4813]: I1201 09:11:13.801217 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-sd6gm/must-gather-mpdwz" event={"ID":"fa60e61e-2a74-4649-9956-eee9c0bb15cc","Type":"ContainerStarted","Data":"24632b9d0366005d6e396d7a2386896a81295d76bfe0e915567cee2dc4f5aba2"} Dec 01 09:11:13 crc kubenswrapper[4813]: I1201 09:11:13.824773 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-sd6gm/must-gather-mpdwz" podStartSLOduration=1.824731425 podStartE2EDuration="1.824731425s" podCreationTimestamp="2025-12-01 09:11:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 09:11:13.820706151 +0000 UTC m=+1858.443627757" watchObservedRunningTime="2025-12-01 09:11:13.824731425 +0000 UTC m=+1858.447653011" Dec 01 09:11:19 crc kubenswrapper[4813]: I1201 09:11:19.917739 4813 scope.go:117] "RemoveContainer" containerID="1e92e4c9dd814667919f1d0dd4ef78e7e090b2558af069309ed4e0894bcc82e4" Dec 01 09:11:46 crc kubenswrapper[4813]: I1201 09:11:46.023102 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368cx94s5_071eb401-2f12-41e4-b52e-312ca4152da5/util/0.log" Dec 01 09:11:46 crc kubenswrapper[4813]: I1201 09:11:46.171897 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368cx94s5_071eb401-2f12-41e4-b52e-312ca4152da5/util/0.log" Dec 01 09:11:46 crc kubenswrapper[4813]: I1201 09:11:46.194394 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368cx94s5_071eb401-2f12-41e4-b52e-312ca4152da5/pull/0.log" Dec 01 09:11:46 crc kubenswrapper[4813]: I1201 09:11:46.196946 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368cx94s5_071eb401-2f12-41e4-b52e-312ca4152da5/pull/0.log" Dec 01 09:11:46 crc kubenswrapper[4813]: I1201 09:11:46.356369 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368cx94s5_071eb401-2f12-41e4-b52e-312ca4152da5/util/0.log" Dec 01 09:11:46 crc kubenswrapper[4813]: I1201 09:11:46.384547 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368cx94s5_071eb401-2f12-41e4-b52e-312ca4152da5/pull/0.log" Dec 01 09:11:46 crc kubenswrapper[4813]: I1201 09:11:46.426099 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368cx94s5_071eb401-2f12-41e4-b52e-312ca4152da5/extract/0.log" Dec 01 09:11:46 crc kubenswrapper[4813]: I1201 09:11:46.531695 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-5998794db6-9lz55_d9af9c23-0392-4caf-b10f-afc96b109c49/manager/0.log" Dec 01 09:11:46 crc kubenswrapper[4813]: I1201 09:11:46.589672 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-index-jn2z5_488d9561-dccc-4b1f-b62f-7af825416efb/registry-server/0.log" Dec 01 09:11:58 crc kubenswrapper[4813]: I1201 09:11:58.527475 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-fxgbc_2fd313fd-c0b1-4aae-8b3b-2d078a7fc7d0/control-plane-machine-set-operator/0.log" Dec 01 09:11:58 crc kubenswrapper[4813]: I1201 09:11:58.692498 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-n72r5_b35b7145-810b-4945-9819-6d9bb650ec73/kube-rbac-proxy/0.log" Dec 01 09:11:58 crc kubenswrapper[4813]: I1201 09:11:58.693702 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-n72r5_b35b7145-810b-4945-9819-6d9bb650ec73/machine-api-operator/0.log" Dec 01 09:12:14 crc kubenswrapper[4813]: I1201 09:12:14.365166 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-f8648f98b-m47sh_ce2e13a7-5901-440b-8a7c-6516b77aea04/kube-rbac-proxy/0.log" Dec 01 09:12:14 crc kubenswrapper[4813]: I1201 09:12:14.366169 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-f8648f98b-m47sh_ce2e13a7-5901-440b-8a7c-6516b77aea04/controller/0.log" Dec 01 09:12:14 crc kubenswrapper[4813]: I1201 09:12:14.531198 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-7fcb986d4-rgj8f_97457a58-aba0-4e0a-b812-ef4fd0912116/frr-k8s-webhook-server/0.log" Dec 01 09:12:14 crc kubenswrapper[4813]: I1201 09:12:14.550453 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zplkx_f1cda8d0-a6b1-478a-9f8e-83d1fa070bf5/cp-frr-files/0.log" Dec 01 09:12:14 crc kubenswrapper[4813]: I1201 09:12:14.727388 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zplkx_f1cda8d0-a6b1-478a-9f8e-83d1fa070bf5/cp-reloader/0.log" Dec 01 09:12:14 crc kubenswrapper[4813]: I1201 09:12:14.753656 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zplkx_f1cda8d0-a6b1-478a-9f8e-83d1fa070bf5/cp-metrics/0.log" Dec 01 09:12:14 crc kubenswrapper[4813]: I1201 09:12:14.755881 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zplkx_f1cda8d0-a6b1-478a-9f8e-83d1fa070bf5/cp-frr-files/0.log" Dec 01 09:12:14 crc kubenswrapper[4813]: I1201 09:12:14.774000 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zplkx_f1cda8d0-a6b1-478a-9f8e-83d1fa070bf5/cp-reloader/0.log" Dec 01 09:12:14 crc kubenswrapper[4813]: I1201 09:12:14.918990 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zplkx_f1cda8d0-a6b1-478a-9f8e-83d1fa070bf5/cp-metrics/0.log" Dec 01 09:12:14 crc kubenswrapper[4813]: I1201 09:12:14.926928 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zplkx_f1cda8d0-a6b1-478a-9f8e-83d1fa070bf5/cp-metrics/0.log" Dec 01 09:12:14 crc kubenswrapper[4813]: I1201 09:12:14.930548 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zplkx_f1cda8d0-a6b1-478a-9f8e-83d1fa070bf5/cp-frr-files/0.log" Dec 01 09:12:14 crc kubenswrapper[4813]: I1201 09:12:14.936763 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zplkx_f1cda8d0-a6b1-478a-9f8e-83d1fa070bf5/cp-reloader/0.log" Dec 01 09:12:15 crc kubenswrapper[4813]: I1201 09:12:15.115811 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zplkx_f1cda8d0-a6b1-478a-9f8e-83d1fa070bf5/cp-frr-files/0.log" Dec 01 09:12:15 crc kubenswrapper[4813]: I1201 09:12:15.131084 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zplkx_f1cda8d0-a6b1-478a-9f8e-83d1fa070bf5/cp-metrics/0.log" Dec 01 09:12:15 crc kubenswrapper[4813]: I1201 09:12:15.140457 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zplkx_f1cda8d0-a6b1-478a-9f8e-83d1fa070bf5/cp-reloader/0.log" Dec 01 09:12:15 crc kubenswrapper[4813]: I1201 09:12:15.152054 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zplkx_f1cda8d0-a6b1-478a-9f8e-83d1fa070bf5/controller/0.log" Dec 01 09:12:15 crc kubenswrapper[4813]: I1201 09:12:15.315222 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zplkx_f1cda8d0-a6b1-478a-9f8e-83d1fa070bf5/kube-rbac-proxy/0.log" Dec 01 09:12:15 crc kubenswrapper[4813]: I1201 09:12:15.319556 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zplkx_f1cda8d0-a6b1-478a-9f8e-83d1fa070bf5/frr-metrics/0.log" Dec 01 09:12:15 crc kubenswrapper[4813]: I1201 09:12:15.346796 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zplkx_f1cda8d0-a6b1-478a-9f8e-83d1fa070bf5/kube-rbac-proxy-frr/0.log" Dec 01 09:12:15 crc kubenswrapper[4813]: E1201 09:12:15.366102 4813 configmap.go:193] Couldn't get configMap glance-kuttl-tests/openstack-config: configmap "openstack-config" not found Dec 01 09:12:15 crc kubenswrapper[4813]: E1201 09:12:15.366287 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/4d0873f6-c7e7-4a9b-bd6e-a0263fd2bc5a-openstack-config podName:4d0873f6-c7e7-4a9b-bd6e-a0263fd2bc5a nodeName:}" failed. No retries permitted until 2025-12-01 09:14:17.366234034 +0000 UTC m=+2041.989155620 (durationBeforeRetry 2m2s). Error: MountVolume.SetUp failed for volume "openstack-config" (UniqueName: "kubernetes.io/configmap/4d0873f6-c7e7-4a9b-bd6e-a0263fd2bc5a-openstack-config") pod "openstackclient" (UID: "4d0873f6-c7e7-4a9b-bd6e-a0263fd2bc5a") : configmap "openstack-config" not found Dec 01 09:12:15 crc kubenswrapper[4813]: E1201 09:12:15.366929 4813 secret.go:188] Couldn't get secret glance-kuttl-tests/openstack-config-secret: secret "openstack-config-secret" not found Dec 01 09:12:15 crc kubenswrapper[4813]: E1201 09:12:15.367001 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/4d0873f6-c7e7-4a9b-bd6e-a0263fd2bc5a-openstack-config-secret podName:4d0873f6-c7e7-4a9b-bd6e-a0263fd2bc5a nodeName:}" failed. No retries permitted until 2025-12-01 09:14:17.366989096 +0000 UTC m=+2041.989910682 (durationBeforeRetry 2m2s). Error: MountVolume.SetUp failed for volume "openstack-config-secret" (UniqueName: "kubernetes.io/secret/4d0873f6-c7e7-4a9b-bd6e-a0263fd2bc5a-openstack-config-secret") pod "openstackclient" (UID: "4d0873f6-c7e7-4a9b-bd6e-a0263fd2bc5a") : secret "openstack-config-secret" not found Dec 01 09:12:15 crc kubenswrapper[4813]: I1201 09:12:15.504351 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zplkx_f1cda8d0-a6b1-478a-9f8e-83d1fa070bf5/reloader/0.log" Dec 01 09:12:15 crc kubenswrapper[4813]: I1201 09:12:15.621742 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-754cc864cb-4d2tv_4017da37-9dca-4e2d-bab0-a7896bdf3cd5/manager/0.log" Dec 01 09:12:15 crc kubenswrapper[4813]: I1201 09:12:15.775839 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-79d58979df-jnh5k_ed86505b-1475-4ca2-9c22-73ce93b80ab9/webhook-server/0.log" Dec 01 09:12:15 crc kubenswrapper[4813]: I1201 09:12:15.937045 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-227b6_1bf9a820-0d04-43e4-a59b-e6168a6e11fe/kube-rbac-proxy/0.log" Dec 01 09:12:15 crc kubenswrapper[4813]: I1201 09:12:15.998759 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zplkx_f1cda8d0-a6b1-478a-9f8e-83d1fa070bf5/frr/0.log" Dec 01 09:12:16 crc kubenswrapper[4813]: I1201 09:12:16.200984 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-227b6_1bf9a820-0d04-43e4-a59b-e6168a6e11fe/speaker/0.log" Dec 01 09:12:19 crc kubenswrapper[4813]: I1201 09:12:19.966418 4813 scope.go:117] "RemoveContainer" containerID="65dc583e88fe1abbbaf461f2f189e99d2cefde16fd6c99f281b62ca2548dc02a" Dec 01 09:12:27 crc kubenswrapper[4813]: I1201 09:12:27.567725 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_openstackclient_4d0873f6-c7e7-4a9b-bd6e-a0263fd2bc5a/openstackclient/0.log" Dec 01 09:12:38 crc kubenswrapper[4813]: I1201 09:12:38.360201 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83qc9k9_d7e5aa34-61f6-476a-a08e-39ce74c1f82a/util/0.log" Dec 01 09:12:38 crc kubenswrapper[4813]: I1201 09:12:38.593882 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83qc9k9_d7e5aa34-61f6-476a-a08e-39ce74c1f82a/util/0.log" Dec 01 09:12:38 crc kubenswrapper[4813]: I1201 09:12:38.597903 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83qc9k9_d7e5aa34-61f6-476a-a08e-39ce74c1f82a/pull/0.log" Dec 01 09:12:38 crc kubenswrapper[4813]: I1201 09:12:38.619002 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83qc9k9_d7e5aa34-61f6-476a-a08e-39ce74c1f82a/pull/0.log" Dec 01 09:12:38 crc kubenswrapper[4813]: I1201 09:12:38.762862 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83qc9k9_d7e5aa34-61f6-476a-a08e-39ce74c1f82a/pull/0.log" Dec 01 09:12:38 crc kubenswrapper[4813]: I1201 09:12:38.788176 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83qc9k9_d7e5aa34-61f6-476a-a08e-39ce74c1f82a/extract/0.log" Dec 01 09:12:38 crc kubenswrapper[4813]: I1201 09:12:38.794270 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83qc9k9_d7e5aa34-61f6-476a-a08e-39ce74c1f82a/util/0.log" Dec 01 09:12:38 crc kubenswrapper[4813]: I1201 09:12:38.925683 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-gwz7r_03720b75-74be-44af-85fd-6b5a55578ca6/extract-utilities/0.log" Dec 01 09:12:39 crc kubenswrapper[4813]: I1201 09:12:39.074652 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-gwz7r_03720b75-74be-44af-85fd-6b5a55578ca6/extract-content/0.log" Dec 01 09:12:39 crc kubenswrapper[4813]: I1201 09:12:39.083330 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-gwz7r_03720b75-74be-44af-85fd-6b5a55578ca6/extract-content/0.log" Dec 01 09:12:39 crc kubenswrapper[4813]: I1201 09:12:39.085256 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-gwz7r_03720b75-74be-44af-85fd-6b5a55578ca6/extract-utilities/0.log" Dec 01 09:12:39 crc kubenswrapper[4813]: I1201 09:12:39.238227 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-gwz7r_03720b75-74be-44af-85fd-6b5a55578ca6/extract-utilities/0.log" Dec 01 09:12:39 crc kubenswrapper[4813]: I1201 09:12:39.251722 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-gwz7r_03720b75-74be-44af-85fd-6b5a55578ca6/extract-content/0.log" Dec 01 09:12:39 crc kubenswrapper[4813]: I1201 09:12:39.479835 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-f87fz_95c21bda-6191-4d26-a3d4-57056be08207/extract-utilities/0.log" Dec 01 09:12:39 crc kubenswrapper[4813]: I1201 09:12:39.594229 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-f87fz_95c21bda-6191-4d26-a3d4-57056be08207/extract-utilities/0.log" Dec 01 09:12:39 crc kubenswrapper[4813]: I1201 09:12:39.623791 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-f87fz_95c21bda-6191-4d26-a3d4-57056be08207/extract-content/0.log" Dec 01 09:12:39 crc kubenswrapper[4813]: I1201 09:12:39.631489 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-gwz7r_03720b75-74be-44af-85fd-6b5a55578ca6/registry-server/0.log" Dec 01 09:12:39 crc kubenswrapper[4813]: I1201 09:12:39.672814 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-f87fz_95c21bda-6191-4d26-a3d4-57056be08207/extract-content/0.log" Dec 01 09:12:39 crc kubenswrapper[4813]: I1201 09:12:39.795768 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-f87fz_95c21bda-6191-4d26-a3d4-57056be08207/extract-utilities/0.log" Dec 01 09:12:39 crc kubenswrapper[4813]: I1201 09:12:39.815431 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-f87fz_95c21bda-6191-4d26-a3d4-57056be08207/extract-content/0.log" Dec 01 09:12:40 crc kubenswrapper[4813]: I1201 09:12:40.014234 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-nmwnn_8d57deaa-d1bd-421f-84c3-8682472f1491/marketplace-operator/0.log" Dec 01 09:12:40 crc kubenswrapper[4813]: I1201 09:12:40.082885 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-dclft_5b4cbc77-2764-4d42-aa64-e040598935b7/extract-utilities/0.log" Dec 01 09:12:40 crc kubenswrapper[4813]: I1201 09:12:40.315517 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-dclft_5b4cbc77-2764-4d42-aa64-e040598935b7/extract-content/0.log" Dec 01 09:12:40 crc kubenswrapper[4813]: I1201 09:12:40.315789 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-dclft_5b4cbc77-2764-4d42-aa64-e040598935b7/extract-utilities/0.log" Dec 01 09:12:40 crc kubenswrapper[4813]: I1201 09:12:40.354560 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-dclft_5b4cbc77-2764-4d42-aa64-e040598935b7/extract-content/0.log" Dec 01 09:12:40 crc kubenswrapper[4813]: I1201 09:12:40.404589 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-f87fz_95c21bda-6191-4d26-a3d4-57056be08207/registry-server/0.log" Dec 01 09:12:40 crc kubenswrapper[4813]: I1201 09:12:40.550607 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-dclft_5b4cbc77-2764-4d42-aa64-e040598935b7/extract-utilities/0.log" Dec 01 09:12:40 crc kubenswrapper[4813]: I1201 09:12:40.558471 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-dclft_5b4cbc77-2764-4d42-aa64-e040598935b7/extract-content/0.log" Dec 01 09:12:40 crc kubenswrapper[4813]: I1201 09:12:40.641865 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-dclft_5b4cbc77-2764-4d42-aa64-e040598935b7/registry-server/0.log" Dec 01 09:12:40 crc kubenswrapper[4813]: I1201 09:12:40.751061 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-h5ffj_9f704b3a-26b9-43b8-8598-02e769dfbb87/extract-utilities/0.log" Dec 01 09:12:40 crc kubenswrapper[4813]: I1201 09:12:40.873656 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-h5ffj_9f704b3a-26b9-43b8-8598-02e769dfbb87/extract-utilities/0.log" Dec 01 09:12:40 crc kubenswrapper[4813]: I1201 09:12:40.877887 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-h5ffj_9f704b3a-26b9-43b8-8598-02e769dfbb87/extract-content/0.log" Dec 01 09:12:40 crc kubenswrapper[4813]: I1201 09:12:40.906951 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-h5ffj_9f704b3a-26b9-43b8-8598-02e769dfbb87/extract-content/0.log" Dec 01 09:12:41 crc kubenswrapper[4813]: I1201 09:12:41.079923 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-h5ffj_9f704b3a-26b9-43b8-8598-02e769dfbb87/extract-utilities/0.log" Dec 01 09:12:41 crc kubenswrapper[4813]: I1201 09:12:41.088813 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-h5ffj_9f704b3a-26b9-43b8-8598-02e769dfbb87/extract-content/0.log" Dec 01 09:12:41 crc kubenswrapper[4813]: I1201 09:12:41.395113 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-h5ffj_9f704b3a-26b9-43b8-8598-02e769dfbb87/registry-server/0.log" Dec 01 09:12:47 crc kubenswrapper[4813]: I1201 09:12:47.221552 4813 patch_prober.go:28] interesting pod/machine-config-daemon-9wn4q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 09:12:47 crc kubenswrapper[4813]: I1201 09:12:47.221896 4813 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9wn4q" podUID="3d423402-e08f-42e9-b60c-6eb9bf2f6d48" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 09:13:17 crc kubenswrapper[4813]: I1201 09:13:17.222118 4813 patch_prober.go:28] interesting pod/machine-config-daemon-9wn4q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 09:13:17 crc kubenswrapper[4813]: I1201 09:13:17.222791 4813 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9wn4q" podUID="3d423402-e08f-42e9-b60c-6eb9bf2f6d48" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 09:13:47 crc kubenswrapper[4813]: I1201 09:13:47.222248 4813 patch_prober.go:28] interesting pod/machine-config-daemon-9wn4q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 09:13:47 crc kubenswrapper[4813]: I1201 09:13:47.222889 4813 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9wn4q" podUID="3d423402-e08f-42e9-b60c-6eb9bf2f6d48" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 09:13:47 crc kubenswrapper[4813]: I1201 09:13:47.223081 4813 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-9wn4q" Dec 01 09:13:47 crc kubenswrapper[4813]: I1201 09:13:47.224032 4813 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"47c278315d64e3056cb6f536af9317d9d67616d501c03a3c61eb1b5cb51b2405"} pod="openshift-machine-config-operator/machine-config-daemon-9wn4q" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 01 09:13:47 crc kubenswrapper[4813]: I1201 09:13:47.224176 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-9wn4q" podUID="3d423402-e08f-42e9-b60c-6eb9bf2f6d48" containerName="machine-config-daemon" containerID="cri-o://47c278315d64e3056cb6f536af9317d9d67616d501c03a3c61eb1b5cb51b2405" gracePeriod=600 Dec 01 09:13:48 crc kubenswrapper[4813]: I1201 09:13:48.193414 4813 generic.go:334] "Generic (PLEG): container finished" podID="fa60e61e-2a74-4649-9956-eee9c0bb15cc" containerID="72b7a4f9f471289de5a4eb203f9d31da96ee153933bc45286a303d28a09fd295" exitCode=0 Dec 01 09:13:48 crc kubenswrapper[4813]: I1201 09:13:48.193491 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-sd6gm/must-gather-mpdwz" event={"ID":"fa60e61e-2a74-4649-9956-eee9c0bb15cc","Type":"ContainerDied","Data":"72b7a4f9f471289de5a4eb203f9d31da96ee153933bc45286a303d28a09fd295"} Dec 01 09:13:48 crc kubenswrapper[4813]: I1201 09:13:48.194622 4813 scope.go:117] "RemoveContainer" containerID="72b7a4f9f471289de5a4eb203f9d31da96ee153933bc45286a303d28a09fd295" Dec 01 09:13:48 crc kubenswrapper[4813]: I1201 09:13:48.197275 4813 generic.go:334] "Generic (PLEG): container finished" podID="3d423402-e08f-42e9-b60c-6eb9bf2f6d48" containerID="47c278315d64e3056cb6f536af9317d9d67616d501c03a3c61eb1b5cb51b2405" exitCode=0 Dec 01 09:13:48 crc kubenswrapper[4813]: I1201 09:13:48.197330 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-9wn4q" event={"ID":"3d423402-e08f-42e9-b60c-6eb9bf2f6d48","Type":"ContainerDied","Data":"47c278315d64e3056cb6f536af9317d9d67616d501c03a3c61eb1b5cb51b2405"} Dec 01 09:13:48 crc kubenswrapper[4813]: I1201 09:13:48.197401 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-9wn4q" event={"ID":"3d423402-e08f-42e9-b60c-6eb9bf2f6d48","Type":"ContainerStarted","Data":"e0826a7d333367829ec415cbaa9314fd5e1db5cbc96bf148dadb2adfedeb3d6a"} Dec 01 09:13:48 crc kubenswrapper[4813]: I1201 09:13:48.197430 4813 scope.go:117] "RemoveContainer" containerID="09b49517239f156b4ac50c72e4d184fde70e2daaf7b602adcfc3b42ec90a71e0" Dec 01 09:13:49 crc kubenswrapper[4813]: I1201 09:13:49.047495 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-sd6gm_must-gather-mpdwz_fa60e61e-2a74-4649-9956-eee9c0bb15cc/gather/0.log" Dec 01 09:13:57 crc kubenswrapper[4813]: I1201 09:13:57.349380 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-sd6gm/must-gather-mpdwz"] Dec 01 09:13:57 crc kubenswrapper[4813]: I1201 09:13:57.350188 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-sd6gm/must-gather-mpdwz" podUID="fa60e61e-2a74-4649-9956-eee9c0bb15cc" containerName="copy" containerID="cri-o://8772a396422cc9a7ceafad873cd441791ce54573e060515d4029cbc636dfa146" gracePeriod=2 Dec 01 09:13:57 crc kubenswrapper[4813]: I1201 09:13:57.356084 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-sd6gm/must-gather-mpdwz"] Dec 01 09:13:57 crc kubenswrapper[4813]: I1201 09:13:57.711881 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-sd6gm_must-gather-mpdwz_fa60e61e-2a74-4649-9956-eee9c0bb15cc/copy/0.log" Dec 01 09:13:57 crc kubenswrapper[4813]: I1201 09:13:57.712672 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-sd6gm/must-gather-mpdwz" Dec 01 09:13:57 crc kubenswrapper[4813]: I1201 09:13:57.883431 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/fa60e61e-2a74-4649-9956-eee9c0bb15cc-must-gather-output\") pod \"fa60e61e-2a74-4649-9956-eee9c0bb15cc\" (UID: \"fa60e61e-2a74-4649-9956-eee9c0bb15cc\") " Dec 01 09:13:57 crc kubenswrapper[4813]: I1201 09:13:57.883594 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bl9dc\" (UniqueName: \"kubernetes.io/projected/fa60e61e-2a74-4649-9956-eee9c0bb15cc-kube-api-access-bl9dc\") pod \"fa60e61e-2a74-4649-9956-eee9c0bb15cc\" (UID: \"fa60e61e-2a74-4649-9956-eee9c0bb15cc\") " Dec 01 09:13:57 crc kubenswrapper[4813]: I1201 09:13:57.891343 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fa60e61e-2a74-4649-9956-eee9c0bb15cc-kube-api-access-bl9dc" (OuterVolumeSpecName: "kube-api-access-bl9dc") pod "fa60e61e-2a74-4649-9956-eee9c0bb15cc" (UID: "fa60e61e-2a74-4649-9956-eee9c0bb15cc"). InnerVolumeSpecName "kube-api-access-bl9dc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:13:57 crc kubenswrapper[4813]: I1201 09:13:57.947686 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fa60e61e-2a74-4649-9956-eee9c0bb15cc-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "fa60e61e-2a74-4649-9956-eee9c0bb15cc" (UID: "fa60e61e-2a74-4649-9956-eee9c0bb15cc"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 09:13:57 crc kubenswrapper[4813]: I1201 09:13:57.985645 4813 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/fa60e61e-2a74-4649-9956-eee9c0bb15cc-must-gather-output\") on node \"crc\" DevicePath \"\"" Dec 01 09:13:57 crc kubenswrapper[4813]: I1201 09:13:57.985679 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bl9dc\" (UniqueName: \"kubernetes.io/projected/fa60e61e-2a74-4649-9956-eee9c0bb15cc-kube-api-access-bl9dc\") on node \"crc\" DevicePath \"\"" Dec 01 09:13:58 crc kubenswrapper[4813]: I1201 09:13:58.264932 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-sd6gm_must-gather-mpdwz_fa60e61e-2a74-4649-9956-eee9c0bb15cc/copy/0.log" Dec 01 09:13:58 crc kubenswrapper[4813]: I1201 09:13:58.266203 4813 generic.go:334] "Generic (PLEG): container finished" podID="fa60e61e-2a74-4649-9956-eee9c0bb15cc" containerID="8772a396422cc9a7ceafad873cd441791ce54573e060515d4029cbc636dfa146" exitCode=143 Dec 01 09:13:58 crc kubenswrapper[4813]: I1201 09:13:58.266283 4813 scope.go:117] "RemoveContainer" containerID="8772a396422cc9a7ceafad873cd441791ce54573e060515d4029cbc636dfa146" Dec 01 09:13:58 crc kubenswrapper[4813]: I1201 09:13:58.266428 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-sd6gm/must-gather-mpdwz" Dec 01 09:13:58 crc kubenswrapper[4813]: I1201 09:13:58.287831 4813 scope.go:117] "RemoveContainer" containerID="72b7a4f9f471289de5a4eb203f9d31da96ee153933bc45286a303d28a09fd295" Dec 01 09:13:58 crc kubenswrapper[4813]: I1201 09:13:58.334920 4813 scope.go:117] "RemoveContainer" containerID="8772a396422cc9a7ceafad873cd441791ce54573e060515d4029cbc636dfa146" Dec 01 09:13:58 crc kubenswrapper[4813]: E1201 09:13:58.336026 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8772a396422cc9a7ceafad873cd441791ce54573e060515d4029cbc636dfa146\": container with ID starting with 8772a396422cc9a7ceafad873cd441791ce54573e060515d4029cbc636dfa146 not found: ID does not exist" containerID="8772a396422cc9a7ceafad873cd441791ce54573e060515d4029cbc636dfa146" Dec 01 09:13:58 crc kubenswrapper[4813]: I1201 09:13:58.336105 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8772a396422cc9a7ceafad873cd441791ce54573e060515d4029cbc636dfa146"} err="failed to get container status \"8772a396422cc9a7ceafad873cd441791ce54573e060515d4029cbc636dfa146\": rpc error: code = NotFound desc = could not find container \"8772a396422cc9a7ceafad873cd441791ce54573e060515d4029cbc636dfa146\": container with ID starting with 8772a396422cc9a7ceafad873cd441791ce54573e060515d4029cbc636dfa146 not found: ID does not exist" Dec 01 09:13:58 crc kubenswrapper[4813]: I1201 09:13:58.336134 4813 scope.go:117] "RemoveContainer" containerID="72b7a4f9f471289de5a4eb203f9d31da96ee153933bc45286a303d28a09fd295" Dec 01 09:13:58 crc kubenswrapper[4813]: E1201 09:13:58.336581 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"72b7a4f9f471289de5a4eb203f9d31da96ee153933bc45286a303d28a09fd295\": container with ID starting with 72b7a4f9f471289de5a4eb203f9d31da96ee153933bc45286a303d28a09fd295 not found: ID does not exist" containerID="72b7a4f9f471289de5a4eb203f9d31da96ee153933bc45286a303d28a09fd295" Dec 01 09:13:58 crc kubenswrapper[4813]: I1201 09:13:58.336661 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"72b7a4f9f471289de5a4eb203f9d31da96ee153933bc45286a303d28a09fd295"} err="failed to get container status \"72b7a4f9f471289de5a4eb203f9d31da96ee153933bc45286a303d28a09fd295\": rpc error: code = NotFound desc = could not find container \"72b7a4f9f471289de5a4eb203f9d31da96ee153933bc45286a303d28a09fd295\": container with ID starting with 72b7a4f9f471289de5a4eb203f9d31da96ee153933bc45286a303d28a09fd295 not found: ID does not exist" Dec 01 09:13:58 crc kubenswrapper[4813]: I1201 09:13:58.401254 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fa60e61e-2a74-4649-9956-eee9c0bb15cc" path="/var/lib/kubelet/pods/fa60e61e-2a74-4649-9956-eee9c0bb15cc/volumes" Dec 01 09:14:00 crc kubenswrapper[4813]: I1201 09:14:00.364706 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-sq2pm"] Dec 01 09:14:00 crc kubenswrapper[4813]: E1201 09:14:00.366319 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fa60e61e-2a74-4649-9956-eee9c0bb15cc" containerName="gather" Dec 01 09:14:00 crc kubenswrapper[4813]: I1201 09:14:00.366593 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="fa60e61e-2a74-4649-9956-eee9c0bb15cc" containerName="gather" Dec 01 09:14:00 crc kubenswrapper[4813]: E1201 09:14:00.366714 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fa60e61e-2a74-4649-9956-eee9c0bb15cc" containerName="copy" Dec 01 09:14:00 crc kubenswrapper[4813]: I1201 09:14:00.366803 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="fa60e61e-2a74-4649-9956-eee9c0bb15cc" containerName="copy" Dec 01 09:14:00 crc kubenswrapper[4813]: I1201 09:14:00.367165 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="fa60e61e-2a74-4649-9956-eee9c0bb15cc" containerName="copy" Dec 01 09:14:00 crc kubenswrapper[4813]: I1201 09:14:00.367278 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="fa60e61e-2a74-4649-9956-eee9c0bb15cc" containerName="gather" Dec 01 09:14:00 crc kubenswrapper[4813]: I1201 09:14:00.368635 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-sq2pm" Dec 01 09:14:00 crc kubenswrapper[4813]: I1201 09:14:00.378810 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-sq2pm"] Dec 01 09:14:00 crc kubenswrapper[4813]: I1201 09:14:00.520187 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/46594ad2-8136-490e-adf4-545878b6b777-utilities\") pod \"redhat-marketplace-sq2pm\" (UID: \"46594ad2-8136-490e-adf4-545878b6b777\") " pod="openshift-marketplace/redhat-marketplace-sq2pm" Dec 01 09:14:00 crc kubenswrapper[4813]: I1201 09:14:00.520273 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/46594ad2-8136-490e-adf4-545878b6b777-catalog-content\") pod \"redhat-marketplace-sq2pm\" (UID: \"46594ad2-8136-490e-adf4-545878b6b777\") " pod="openshift-marketplace/redhat-marketplace-sq2pm" Dec 01 09:14:00 crc kubenswrapper[4813]: I1201 09:14:00.520593 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d4g4z\" (UniqueName: \"kubernetes.io/projected/46594ad2-8136-490e-adf4-545878b6b777-kube-api-access-d4g4z\") pod \"redhat-marketplace-sq2pm\" (UID: \"46594ad2-8136-490e-adf4-545878b6b777\") " pod="openshift-marketplace/redhat-marketplace-sq2pm" Dec 01 09:14:00 crc kubenswrapper[4813]: I1201 09:14:00.622511 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/46594ad2-8136-490e-adf4-545878b6b777-catalog-content\") pod \"redhat-marketplace-sq2pm\" (UID: \"46594ad2-8136-490e-adf4-545878b6b777\") " pod="openshift-marketplace/redhat-marketplace-sq2pm" Dec 01 09:14:00 crc kubenswrapper[4813]: I1201 09:14:00.622626 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d4g4z\" (UniqueName: \"kubernetes.io/projected/46594ad2-8136-490e-adf4-545878b6b777-kube-api-access-d4g4z\") pod \"redhat-marketplace-sq2pm\" (UID: \"46594ad2-8136-490e-adf4-545878b6b777\") " pod="openshift-marketplace/redhat-marketplace-sq2pm" Dec 01 09:14:00 crc kubenswrapper[4813]: I1201 09:14:00.622699 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/46594ad2-8136-490e-adf4-545878b6b777-utilities\") pod \"redhat-marketplace-sq2pm\" (UID: \"46594ad2-8136-490e-adf4-545878b6b777\") " pod="openshift-marketplace/redhat-marketplace-sq2pm" Dec 01 09:14:00 crc kubenswrapper[4813]: I1201 09:14:00.623234 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/46594ad2-8136-490e-adf4-545878b6b777-catalog-content\") pod \"redhat-marketplace-sq2pm\" (UID: \"46594ad2-8136-490e-adf4-545878b6b777\") " pod="openshift-marketplace/redhat-marketplace-sq2pm" Dec 01 09:14:00 crc kubenswrapper[4813]: I1201 09:14:00.623271 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/46594ad2-8136-490e-adf4-545878b6b777-utilities\") pod \"redhat-marketplace-sq2pm\" (UID: \"46594ad2-8136-490e-adf4-545878b6b777\") " pod="openshift-marketplace/redhat-marketplace-sq2pm" Dec 01 09:14:00 crc kubenswrapper[4813]: I1201 09:14:00.644475 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d4g4z\" (UniqueName: \"kubernetes.io/projected/46594ad2-8136-490e-adf4-545878b6b777-kube-api-access-d4g4z\") pod \"redhat-marketplace-sq2pm\" (UID: \"46594ad2-8136-490e-adf4-545878b6b777\") " pod="openshift-marketplace/redhat-marketplace-sq2pm" Dec 01 09:14:00 crc kubenswrapper[4813]: I1201 09:14:00.694158 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-sq2pm" Dec 01 09:14:01 crc kubenswrapper[4813]: I1201 09:14:01.480078 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-sq2pm"] Dec 01 09:14:02 crc kubenswrapper[4813]: I1201 09:14:02.296566 4813 generic.go:334] "Generic (PLEG): container finished" podID="46594ad2-8136-490e-adf4-545878b6b777" containerID="ac5da40214efbad5240ddaf8582eaf320a12b45ba1a436c40108599fab533b52" exitCode=0 Dec 01 09:14:02 crc kubenswrapper[4813]: I1201 09:14:02.296649 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-sq2pm" event={"ID":"46594ad2-8136-490e-adf4-545878b6b777","Type":"ContainerDied","Data":"ac5da40214efbad5240ddaf8582eaf320a12b45ba1a436c40108599fab533b52"} Dec 01 09:14:02 crc kubenswrapper[4813]: I1201 09:14:02.297139 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-sq2pm" event={"ID":"46594ad2-8136-490e-adf4-545878b6b777","Type":"ContainerStarted","Data":"7878874eda214ba9acfbc0bf2b168964eac2c3aef548af520b2f78c7d4082f53"} Dec 01 09:14:02 crc kubenswrapper[4813]: I1201 09:14:02.301319 4813 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 01 09:14:04 crc kubenswrapper[4813]: I1201 09:14:04.312391 4813 generic.go:334] "Generic (PLEG): container finished" podID="46594ad2-8136-490e-adf4-545878b6b777" containerID="44a68e5e62c54b9f31e8401fa5a0f9c9b670770ed0f16a5ba64dfc68d32c4098" exitCode=0 Dec 01 09:14:04 crc kubenswrapper[4813]: I1201 09:14:04.312516 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-sq2pm" event={"ID":"46594ad2-8136-490e-adf4-545878b6b777","Type":"ContainerDied","Data":"44a68e5e62c54b9f31e8401fa5a0f9c9b670770ed0f16a5ba64dfc68d32c4098"} Dec 01 09:14:05 crc kubenswrapper[4813]: I1201 09:14:05.319390 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-sq2pm" event={"ID":"46594ad2-8136-490e-adf4-545878b6b777","Type":"ContainerStarted","Data":"1c15cee1850552422a619f8efc475999c47821e8d6f33eca5d2cbe708f9ebdae"} Dec 01 09:14:05 crc kubenswrapper[4813]: I1201 09:14:05.334411 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-sq2pm" podStartSLOduration=2.785963707 podStartE2EDuration="5.334368367s" podCreationTimestamp="2025-12-01 09:14:00 +0000 UTC" firstStartedPulling="2025-12-01 09:14:02.300881153 +0000 UTC m=+2026.923802739" lastFinishedPulling="2025-12-01 09:14:04.849285813 +0000 UTC m=+2029.472207399" observedRunningTime="2025-12-01 09:14:05.334067388 +0000 UTC m=+2029.956988974" watchObservedRunningTime="2025-12-01 09:14:05.334368367 +0000 UTC m=+2029.957289953" Dec 01 09:14:10 crc kubenswrapper[4813]: I1201 09:14:10.695160 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-sq2pm" Dec 01 09:14:10 crc kubenswrapper[4813]: I1201 09:14:10.695752 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-sq2pm" Dec 01 09:14:10 crc kubenswrapper[4813]: I1201 09:14:10.744281 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-sq2pm" Dec 01 09:14:11 crc kubenswrapper[4813]: I1201 09:14:11.403873 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-sq2pm" Dec 01 09:14:11 crc kubenswrapper[4813]: I1201 09:14:11.454310 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-sq2pm"] Dec 01 09:14:13 crc kubenswrapper[4813]: I1201 09:14:13.366606 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-sq2pm" podUID="46594ad2-8136-490e-adf4-545878b6b777" containerName="registry-server" containerID="cri-o://1c15cee1850552422a619f8efc475999c47821e8d6f33eca5d2cbe708f9ebdae" gracePeriod=2 Dec 01 09:14:14 crc kubenswrapper[4813]: I1201 09:14:14.442360 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-sq2pm" Dec 01 09:14:14 crc kubenswrapper[4813]: I1201 09:14:14.447291 4813 generic.go:334] "Generic (PLEG): container finished" podID="46594ad2-8136-490e-adf4-545878b6b777" containerID="1c15cee1850552422a619f8efc475999c47821e8d6f33eca5d2cbe708f9ebdae" exitCode=0 Dec 01 09:14:14 crc kubenswrapper[4813]: I1201 09:14:14.447379 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-sq2pm" event={"ID":"46594ad2-8136-490e-adf4-545878b6b777","Type":"ContainerDied","Data":"1c15cee1850552422a619f8efc475999c47821e8d6f33eca5d2cbe708f9ebdae"} Dec 01 09:14:14 crc kubenswrapper[4813]: I1201 09:14:14.447450 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-sq2pm" event={"ID":"46594ad2-8136-490e-adf4-545878b6b777","Type":"ContainerDied","Data":"7878874eda214ba9acfbc0bf2b168964eac2c3aef548af520b2f78c7d4082f53"} Dec 01 09:14:14 crc kubenswrapper[4813]: I1201 09:14:14.447566 4813 scope.go:117] "RemoveContainer" containerID="1c15cee1850552422a619f8efc475999c47821e8d6f33eca5d2cbe708f9ebdae" Dec 01 09:14:14 crc kubenswrapper[4813]: I1201 09:14:14.482995 4813 scope.go:117] "RemoveContainer" containerID="44a68e5e62c54b9f31e8401fa5a0f9c9b670770ed0f16a5ba64dfc68d32c4098" Dec 01 09:14:14 crc kubenswrapper[4813]: I1201 09:14:14.498032 4813 scope.go:117] "RemoveContainer" containerID="ac5da40214efbad5240ddaf8582eaf320a12b45ba1a436c40108599fab533b52" Dec 01 09:14:14 crc kubenswrapper[4813]: I1201 09:14:14.515614 4813 scope.go:117] "RemoveContainer" containerID="1c15cee1850552422a619f8efc475999c47821e8d6f33eca5d2cbe708f9ebdae" Dec 01 09:14:14 crc kubenswrapper[4813]: E1201 09:14:14.516298 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1c15cee1850552422a619f8efc475999c47821e8d6f33eca5d2cbe708f9ebdae\": container with ID starting with 1c15cee1850552422a619f8efc475999c47821e8d6f33eca5d2cbe708f9ebdae not found: ID does not exist" containerID="1c15cee1850552422a619f8efc475999c47821e8d6f33eca5d2cbe708f9ebdae" Dec 01 09:14:14 crc kubenswrapper[4813]: I1201 09:14:14.516391 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1c15cee1850552422a619f8efc475999c47821e8d6f33eca5d2cbe708f9ebdae"} err="failed to get container status \"1c15cee1850552422a619f8efc475999c47821e8d6f33eca5d2cbe708f9ebdae\": rpc error: code = NotFound desc = could not find container \"1c15cee1850552422a619f8efc475999c47821e8d6f33eca5d2cbe708f9ebdae\": container with ID starting with 1c15cee1850552422a619f8efc475999c47821e8d6f33eca5d2cbe708f9ebdae not found: ID does not exist" Dec 01 09:14:14 crc kubenswrapper[4813]: I1201 09:14:14.516438 4813 scope.go:117] "RemoveContainer" containerID="44a68e5e62c54b9f31e8401fa5a0f9c9b670770ed0f16a5ba64dfc68d32c4098" Dec 01 09:14:14 crc kubenswrapper[4813]: E1201 09:14:14.516808 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"44a68e5e62c54b9f31e8401fa5a0f9c9b670770ed0f16a5ba64dfc68d32c4098\": container with ID starting with 44a68e5e62c54b9f31e8401fa5a0f9c9b670770ed0f16a5ba64dfc68d32c4098 not found: ID does not exist" containerID="44a68e5e62c54b9f31e8401fa5a0f9c9b670770ed0f16a5ba64dfc68d32c4098" Dec 01 09:14:14 crc kubenswrapper[4813]: I1201 09:14:14.516874 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"44a68e5e62c54b9f31e8401fa5a0f9c9b670770ed0f16a5ba64dfc68d32c4098"} err="failed to get container status \"44a68e5e62c54b9f31e8401fa5a0f9c9b670770ed0f16a5ba64dfc68d32c4098\": rpc error: code = NotFound desc = could not find container \"44a68e5e62c54b9f31e8401fa5a0f9c9b670770ed0f16a5ba64dfc68d32c4098\": container with ID starting with 44a68e5e62c54b9f31e8401fa5a0f9c9b670770ed0f16a5ba64dfc68d32c4098 not found: ID does not exist" Dec 01 09:14:14 crc kubenswrapper[4813]: I1201 09:14:14.516948 4813 scope.go:117] "RemoveContainer" containerID="ac5da40214efbad5240ddaf8582eaf320a12b45ba1a436c40108599fab533b52" Dec 01 09:14:14 crc kubenswrapper[4813]: E1201 09:14:14.517207 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ac5da40214efbad5240ddaf8582eaf320a12b45ba1a436c40108599fab533b52\": container with ID starting with ac5da40214efbad5240ddaf8582eaf320a12b45ba1a436c40108599fab533b52 not found: ID does not exist" containerID="ac5da40214efbad5240ddaf8582eaf320a12b45ba1a436c40108599fab533b52" Dec 01 09:14:14 crc kubenswrapper[4813]: I1201 09:14:14.517265 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ac5da40214efbad5240ddaf8582eaf320a12b45ba1a436c40108599fab533b52"} err="failed to get container status \"ac5da40214efbad5240ddaf8582eaf320a12b45ba1a436c40108599fab533b52\": rpc error: code = NotFound desc = could not find container \"ac5da40214efbad5240ddaf8582eaf320a12b45ba1a436c40108599fab533b52\": container with ID starting with ac5da40214efbad5240ddaf8582eaf320a12b45ba1a436c40108599fab533b52 not found: ID does not exist" Dec 01 09:14:14 crc kubenswrapper[4813]: I1201 09:14:14.633543 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/46594ad2-8136-490e-adf4-545878b6b777-catalog-content\") pod \"46594ad2-8136-490e-adf4-545878b6b777\" (UID: \"46594ad2-8136-490e-adf4-545878b6b777\") " Dec 01 09:14:14 crc kubenswrapper[4813]: I1201 09:14:14.633743 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4g4z\" (UniqueName: \"kubernetes.io/projected/46594ad2-8136-490e-adf4-545878b6b777-kube-api-access-d4g4z\") pod \"46594ad2-8136-490e-adf4-545878b6b777\" (UID: \"46594ad2-8136-490e-adf4-545878b6b777\") " Dec 01 09:14:14 crc kubenswrapper[4813]: I1201 09:14:14.633805 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/46594ad2-8136-490e-adf4-545878b6b777-utilities\") pod \"46594ad2-8136-490e-adf4-545878b6b777\" (UID: \"46594ad2-8136-490e-adf4-545878b6b777\") " Dec 01 09:14:14 crc kubenswrapper[4813]: I1201 09:14:14.636292 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/46594ad2-8136-490e-adf4-545878b6b777-utilities" (OuterVolumeSpecName: "utilities") pod "46594ad2-8136-490e-adf4-545878b6b777" (UID: "46594ad2-8136-490e-adf4-545878b6b777"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 09:14:14 crc kubenswrapper[4813]: I1201 09:14:14.642561 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/46594ad2-8136-490e-adf4-545878b6b777-kube-api-access-d4g4z" (OuterVolumeSpecName: "kube-api-access-d4g4z") pod "46594ad2-8136-490e-adf4-545878b6b777" (UID: "46594ad2-8136-490e-adf4-545878b6b777"). InnerVolumeSpecName "kube-api-access-d4g4z". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:14:14 crc kubenswrapper[4813]: I1201 09:14:14.654888 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/46594ad2-8136-490e-adf4-545878b6b777-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "46594ad2-8136-490e-adf4-545878b6b777" (UID: "46594ad2-8136-490e-adf4-545878b6b777"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 09:14:14 crc kubenswrapper[4813]: I1201 09:14:14.735545 4813 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/46594ad2-8136-490e-adf4-545878b6b777-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 09:14:14 crc kubenswrapper[4813]: I1201 09:14:14.735584 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4g4z\" (UniqueName: \"kubernetes.io/projected/46594ad2-8136-490e-adf4-545878b6b777-kube-api-access-d4g4z\") on node \"crc\" DevicePath \"\"" Dec 01 09:14:14 crc kubenswrapper[4813]: I1201 09:14:14.735602 4813 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/46594ad2-8136-490e-adf4-545878b6b777-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 09:14:15 crc kubenswrapper[4813]: I1201 09:14:15.455136 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-sq2pm" Dec 01 09:14:15 crc kubenswrapper[4813]: I1201 09:14:15.485826 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-sq2pm"] Dec 01 09:14:15 crc kubenswrapper[4813]: I1201 09:14:15.491245 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-sq2pm"] Dec 01 09:14:16 crc kubenswrapper[4813]: I1201 09:14:16.506882 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="46594ad2-8136-490e-adf4-545878b6b777" path="/var/lib/kubelet/pods/46594ad2-8136-490e-adf4-545878b6b777/volumes" Dec 01 09:14:17 crc kubenswrapper[4813]: E1201 09:14:17.410947 4813 configmap.go:193] Couldn't get configMap glance-kuttl-tests/openstack-config: configmap "openstack-config" not found Dec 01 09:14:17 crc kubenswrapper[4813]: E1201 09:14:17.411138 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/4d0873f6-c7e7-4a9b-bd6e-a0263fd2bc5a-openstack-config podName:4d0873f6-c7e7-4a9b-bd6e-a0263fd2bc5a nodeName:}" failed. No retries permitted until 2025-12-01 09:16:19.411082109 +0000 UTC m=+2164.034003695 (durationBeforeRetry 2m2s). Error: MountVolume.SetUp failed for volume "openstack-config" (UniqueName: "kubernetes.io/configmap/4d0873f6-c7e7-4a9b-bd6e-a0263fd2bc5a-openstack-config") pod "openstackclient" (UID: "4d0873f6-c7e7-4a9b-bd6e-a0263fd2bc5a") : configmap "openstack-config" not found Dec 01 09:14:17 crc kubenswrapper[4813]: E1201 09:14:17.411697 4813 secret.go:188] Couldn't get secret glance-kuttl-tests/openstack-config-secret: secret "openstack-config-secret" not found Dec 01 09:14:17 crc kubenswrapper[4813]: E1201 09:14:17.411740 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/4d0873f6-c7e7-4a9b-bd6e-a0263fd2bc5a-openstack-config-secret podName:4d0873f6-c7e7-4a9b-bd6e-a0263fd2bc5a nodeName:}" failed. No retries permitted until 2025-12-01 09:16:19.411732497 +0000 UTC m=+2164.034654073 (durationBeforeRetry 2m2s). Error: MountVolume.SetUp failed for volume "openstack-config-secret" (UniqueName: "kubernetes.io/secret/4d0873f6-c7e7-4a9b-bd6e-a0263fd2bc5a-openstack-config-secret") pod "openstackclient" (UID: "4d0873f6-c7e7-4a9b-bd6e-a0263fd2bc5a") : secret "openstack-config-secret" not found Dec 01 09:15:00 crc kubenswrapper[4813]: I1201 09:15:00.156652 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29409675-k9t7r"] Dec 01 09:15:00 crc kubenswrapper[4813]: E1201 09:15:00.157726 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="46594ad2-8136-490e-adf4-545878b6b777" containerName="registry-server" Dec 01 09:15:00 crc kubenswrapper[4813]: I1201 09:15:00.157765 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="46594ad2-8136-490e-adf4-545878b6b777" containerName="registry-server" Dec 01 09:15:00 crc kubenswrapper[4813]: E1201 09:15:00.157792 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="46594ad2-8136-490e-adf4-545878b6b777" containerName="extract-utilities" Dec 01 09:15:00 crc kubenswrapper[4813]: I1201 09:15:00.157801 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="46594ad2-8136-490e-adf4-545878b6b777" containerName="extract-utilities" Dec 01 09:15:00 crc kubenswrapper[4813]: E1201 09:15:00.157831 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="46594ad2-8136-490e-adf4-545878b6b777" containerName="extract-content" Dec 01 09:15:00 crc kubenswrapper[4813]: I1201 09:15:00.157838 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="46594ad2-8136-490e-adf4-545878b6b777" containerName="extract-content" Dec 01 09:15:00 crc kubenswrapper[4813]: I1201 09:15:00.158042 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="46594ad2-8136-490e-adf4-545878b6b777" containerName="registry-server" Dec 01 09:15:00 crc kubenswrapper[4813]: I1201 09:15:00.158772 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29409675-k9t7r" Dec 01 09:15:00 crc kubenswrapper[4813]: I1201 09:15:00.162911 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 01 09:15:00 crc kubenswrapper[4813]: I1201 09:15:00.163237 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 01 09:15:00 crc kubenswrapper[4813]: I1201 09:15:00.165404 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29409675-k9t7r"] Dec 01 09:15:00 crc kubenswrapper[4813]: I1201 09:15:00.306996 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b4xj6\" (UniqueName: \"kubernetes.io/projected/d8924eb0-830c-44a6-9042-3bdba9d75844-kube-api-access-b4xj6\") pod \"collect-profiles-29409675-k9t7r\" (UID: \"d8924eb0-830c-44a6-9042-3bdba9d75844\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29409675-k9t7r" Dec 01 09:15:00 crc kubenswrapper[4813]: I1201 09:15:00.307539 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d8924eb0-830c-44a6-9042-3bdba9d75844-config-volume\") pod \"collect-profiles-29409675-k9t7r\" (UID: \"d8924eb0-830c-44a6-9042-3bdba9d75844\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29409675-k9t7r" Dec 01 09:15:00 crc kubenswrapper[4813]: I1201 09:15:00.307591 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d8924eb0-830c-44a6-9042-3bdba9d75844-secret-volume\") pod \"collect-profiles-29409675-k9t7r\" (UID: \"d8924eb0-830c-44a6-9042-3bdba9d75844\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29409675-k9t7r" Dec 01 09:15:00 crc kubenswrapper[4813]: I1201 09:15:00.408923 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b4xj6\" (UniqueName: \"kubernetes.io/projected/d8924eb0-830c-44a6-9042-3bdba9d75844-kube-api-access-b4xj6\") pod \"collect-profiles-29409675-k9t7r\" (UID: \"d8924eb0-830c-44a6-9042-3bdba9d75844\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29409675-k9t7r" Dec 01 09:15:00 crc kubenswrapper[4813]: I1201 09:15:00.409042 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d8924eb0-830c-44a6-9042-3bdba9d75844-config-volume\") pod \"collect-profiles-29409675-k9t7r\" (UID: \"d8924eb0-830c-44a6-9042-3bdba9d75844\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29409675-k9t7r" Dec 01 09:15:00 crc kubenswrapper[4813]: I1201 09:15:00.409070 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d8924eb0-830c-44a6-9042-3bdba9d75844-secret-volume\") pod \"collect-profiles-29409675-k9t7r\" (UID: \"d8924eb0-830c-44a6-9042-3bdba9d75844\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29409675-k9t7r" Dec 01 09:15:00 crc kubenswrapper[4813]: I1201 09:15:00.410797 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d8924eb0-830c-44a6-9042-3bdba9d75844-config-volume\") pod \"collect-profiles-29409675-k9t7r\" (UID: \"d8924eb0-830c-44a6-9042-3bdba9d75844\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29409675-k9t7r" Dec 01 09:15:00 crc kubenswrapper[4813]: I1201 09:15:00.418827 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d8924eb0-830c-44a6-9042-3bdba9d75844-secret-volume\") pod \"collect-profiles-29409675-k9t7r\" (UID: \"d8924eb0-830c-44a6-9042-3bdba9d75844\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29409675-k9t7r" Dec 01 09:15:00 crc kubenswrapper[4813]: I1201 09:15:00.428235 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b4xj6\" (UniqueName: \"kubernetes.io/projected/d8924eb0-830c-44a6-9042-3bdba9d75844-kube-api-access-b4xj6\") pod \"collect-profiles-29409675-k9t7r\" (UID: \"d8924eb0-830c-44a6-9042-3bdba9d75844\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29409675-k9t7r" Dec 01 09:15:00 crc kubenswrapper[4813]: I1201 09:15:00.479215 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29409675-k9t7r" Dec 01 09:15:00 crc kubenswrapper[4813]: I1201 09:15:00.696263 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29409675-k9t7r"] Dec 01 09:15:00 crc kubenswrapper[4813]: I1201 09:15:00.813254 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29409675-k9t7r" event={"ID":"d8924eb0-830c-44a6-9042-3bdba9d75844","Type":"ContainerStarted","Data":"a63ccb7c4606440b40adffc05f03ca32657818d13e0e3f22636ae569406501ee"} Dec 01 09:15:01 crc kubenswrapper[4813]: I1201 09:15:01.824042 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29409675-k9t7r" event={"ID":"d8924eb0-830c-44a6-9042-3bdba9d75844","Type":"ContainerDied","Data":"d34d47bc47beb016e081e07c49378a430c02deabbb7aec0d75e1ba691c297c90"} Dec 01 09:15:01 crc kubenswrapper[4813]: I1201 09:15:01.823999 4813 generic.go:334] "Generic (PLEG): container finished" podID="d8924eb0-830c-44a6-9042-3bdba9d75844" containerID="d34d47bc47beb016e081e07c49378a430c02deabbb7aec0d75e1ba691c297c90" exitCode=0 Dec 01 09:15:03 crc kubenswrapper[4813]: I1201 09:15:03.131939 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29409675-k9t7r" Dec 01 09:15:03 crc kubenswrapper[4813]: I1201 09:15:03.254238 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d8924eb0-830c-44a6-9042-3bdba9d75844-secret-volume\") pod \"d8924eb0-830c-44a6-9042-3bdba9d75844\" (UID: \"d8924eb0-830c-44a6-9042-3bdba9d75844\") " Dec 01 09:15:03 crc kubenswrapper[4813]: I1201 09:15:03.254370 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b4xj6\" (UniqueName: \"kubernetes.io/projected/d8924eb0-830c-44a6-9042-3bdba9d75844-kube-api-access-b4xj6\") pod \"d8924eb0-830c-44a6-9042-3bdba9d75844\" (UID: \"d8924eb0-830c-44a6-9042-3bdba9d75844\") " Dec 01 09:15:03 crc kubenswrapper[4813]: I1201 09:15:03.254424 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d8924eb0-830c-44a6-9042-3bdba9d75844-config-volume\") pod \"d8924eb0-830c-44a6-9042-3bdba9d75844\" (UID: \"d8924eb0-830c-44a6-9042-3bdba9d75844\") " Dec 01 09:15:03 crc kubenswrapper[4813]: I1201 09:15:03.255912 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d8924eb0-830c-44a6-9042-3bdba9d75844-config-volume" (OuterVolumeSpecName: "config-volume") pod "d8924eb0-830c-44a6-9042-3bdba9d75844" (UID: "d8924eb0-830c-44a6-9042-3bdba9d75844"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 09:15:03 crc kubenswrapper[4813]: I1201 09:15:03.261217 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d8924eb0-830c-44a6-9042-3bdba9d75844-kube-api-access-b4xj6" (OuterVolumeSpecName: "kube-api-access-b4xj6") pod "d8924eb0-830c-44a6-9042-3bdba9d75844" (UID: "d8924eb0-830c-44a6-9042-3bdba9d75844"). InnerVolumeSpecName "kube-api-access-b4xj6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:15:03 crc kubenswrapper[4813]: I1201 09:15:03.264114 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d8924eb0-830c-44a6-9042-3bdba9d75844-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "d8924eb0-830c-44a6-9042-3bdba9d75844" (UID: "d8924eb0-830c-44a6-9042-3bdba9d75844"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 09:15:03 crc kubenswrapper[4813]: I1201 09:15:03.356232 4813 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d8924eb0-830c-44a6-9042-3bdba9d75844-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 01 09:15:03 crc kubenswrapper[4813]: I1201 09:15:03.356274 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b4xj6\" (UniqueName: \"kubernetes.io/projected/d8924eb0-830c-44a6-9042-3bdba9d75844-kube-api-access-b4xj6\") on node \"crc\" DevicePath \"\"" Dec 01 09:15:03 crc kubenswrapper[4813]: I1201 09:15:03.356293 4813 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d8924eb0-830c-44a6-9042-3bdba9d75844-config-volume\") on node \"crc\" DevicePath \"\"" Dec 01 09:15:03 crc kubenswrapper[4813]: I1201 09:15:03.843496 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29409675-k9t7r" event={"ID":"d8924eb0-830c-44a6-9042-3bdba9d75844","Type":"ContainerDied","Data":"a63ccb7c4606440b40adffc05f03ca32657818d13e0e3f22636ae569406501ee"} Dec 01 09:15:03 crc kubenswrapper[4813]: I1201 09:15:03.843557 4813 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a63ccb7c4606440b40adffc05f03ca32657818d13e0e3f22636ae569406501ee" Dec 01 09:15:03 crc kubenswrapper[4813]: I1201 09:15:03.844005 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29409675-k9t7r" Dec 01 09:15:04 crc kubenswrapper[4813]: I1201 09:15:04.208133 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29409630-45pn5"] Dec 01 09:15:04 crc kubenswrapper[4813]: I1201 09:15:04.212120 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29409630-45pn5"] Dec 01 09:15:04 crc kubenswrapper[4813]: I1201 09:15:04.402402 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="95e0dc94-8aec-42d0-8b19-7cdf12fa7dee" path="/var/lib/kubelet/pods/95e0dc94-8aec-42d0-8b19-7cdf12fa7dee/volumes" Dec 01 09:15:20 crc kubenswrapper[4813]: I1201 09:15:20.074069 4813 scope.go:117] "RemoveContainer" containerID="f6d3eb682a8ac8a11cfdaff1cc058f21bb6b5cfef699c8f6fdf59a4c88c3f826" Dec 01 09:16:15 crc kubenswrapper[4813]: I1201 09:16:15.845777 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-s69x9"] Dec 01 09:16:15 crc kubenswrapper[4813]: E1201 09:16:15.846802 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d8924eb0-830c-44a6-9042-3bdba9d75844" containerName="collect-profiles" Dec 01 09:16:15 crc kubenswrapper[4813]: I1201 09:16:15.846839 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="d8924eb0-830c-44a6-9042-3bdba9d75844" containerName="collect-profiles" Dec 01 09:16:15 crc kubenswrapper[4813]: I1201 09:16:15.847080 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="d8924eb0-830c-44a6-9042-3bdba9d75844" containerName="collect-profiles" Dec 01 09:16:15 crc kubenswrapper[4813]: I1201 09:16:15.852482 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-s69x9" Dec 01 09:16:15 crc kubenswrapper[4813]: I1201 09:16:15.865767 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-s69x9"] Dec 01 09:16:15 crc kubenswrapper[4813]: I1201 09:16:15.885216 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rbt25\" (UniqueName: \"kubernetes.io/projected/609748aa-5b9f-499e-89af-255a5c08c463-kube-api-access-rbt25\") pod \"certified-operators-s69x9\" (UID: \"609748aa-5b9f-499e-89af-255a5c08c463\") " pod="openshift-marketplace/certified-operators-s69x9" Dec 01 09:16:15 crc kubenswrapper[4813]: I1201 09:16:15.885275 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/609748aa-5b9f-499e-89af-255a5c08c463-catalog-content\") pod \"certified-operators-s69x9\" (UID: \"609748aa-5b9f-499e-89af-255a5c08c463\") " pod="openshift-marketplace/certified-operators-s69x9" Dec 01 09:16:15 crc kubenswrapper[4813]: I1201 09:16:15.885431 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/609748aa-5b9f-499e-89af-255a5c08c463-utilities\") pod \"certified-operators-s69x9\" (UID: \"609748aa-5b9f-499e-89af-255a5c08c463\") " pod="openshift-marketplace/certified-operators-s69x9" Dec 01 09:16:15 crc kubenswrapper[4813]: I1201 09:16:15.986600 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rbt25\" (UniqueName: \"kubernetes.io/projected/609748aa-5b9f-499e-89af-255a5c08c463-kube-api-access-rbt25\") pod \"certified-operators-s69x9\" (UID: \"609748aa-5b9f-499e-89af-255a5c08c463\") " pod="openshift-marketplace/certified-operators-s69x9" Dec 01 09:16:15 crc kubenswrapper[4813]: I1201 09:16:15.986664 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/609748aa-5b9f-499e-89af-255a5c08c463-catalog-content\") pod \"certified-operators-s69x9\" (UID: \"609748aa-5b9f-499e-89af-255a5c08c463\") " pod="openshift-marketplace/certified-operators-s69x9" Dec 01 09:16:15 crc kubenswrapper[4813]: I1201 09:16:15.986778 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/609748aa-5b9f-499e-89af-255a5c08c463-utilities\") pod \"certified-operators-s69x9\" (UID: \"609748aa-5b9f-499e-89af-255a5c08c463\") " pod="openshift-marketplace/certified-operators-s69x9" Dec 01 09:16:15 crc kubenswrapper[4813]: I1201 09:16:15.987334 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/609748aa-5b9f-499e-89af-255a5c08c463-catalog-content\") pod \"certified-operators-s69x9\" (UID: \"609748aa-5b9f-499e-89af-255a5c08c463\") " pod="openshift-marketplace/certified-operators-s69x9" Dec 01 09:16:15 crc kubenswrapper[4813]: I1201 09:16:15.987411 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/609748aa-5b9f-499e-89af-255a5c08c463-utilities\") pod \"certified-operators-s69x9\" (UID: \"609748aa-5b9f-499e-89af-255a5c08c463\") " pod="openshift-marketplace/certified-operators-s69x9" Dec 01 09:16:16 crc kubenswrapper[4813]: I1201 09:16:16.015801 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rbt25\" (UniqueName: \"kubernetes.io/projected/609748aa-5b9f-499e-89af-255a5c08c463-kube-api-access-rbt25\") pod \"certified-operators-s69x9\" (UID: \"609748aa-5b9f-499e-89af-255a5c08c463\") " pod="openshift-marketplace/certified-operators-s69x9" Dec 01 09:16:16 crc kubenswrapper[4813]: I1201 09:16:16.176369 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-s69x9" Dec 01 09:16:16 crc kubenswrapper[4813]: I1201 09:16:16.672116 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-s69x9"] Dec 01 09:16:16 crc kubenswrapper[4813]: W1201 09:16:16.678074 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod609748aa_5b9f_499e_89af_255a5c08c463.slice/crio-95635960fc6928151bdfc4aee14d7d4811a5108c6fcc800a3d767fc2a991db51 WatchSource:0}: Error finding container 95635960fc6928151bdfc4aee14d7d4811a5108c6fcc800a3d767fc2a991db51: Status 404 returned error can't find the container with id 95635960fc6928151bdfc4aee14d7d4811a5108c6fcc800a3d767fc2a991db51 Dec 01 09:16:17 crc kubenswrapper[4813]: I1201 09:16:17.222094 4813 patch_prober.go:28] interesting pod/machine-config-daemon-9wn4q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 09:16:17 crc kubenswrapper[4813]: I1201 09:16:17.223269 4813 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9wn4q" podUID="3d423402-e08f-42e9-b60c-6eb9bf2f6d48" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 09:16:17 crc kubenswrapper[4813]: I1201 09:16:17.426021 4813 generic.go:334] "Generic (PLEG): container finished" podID="609748aa-5b9f-499e-89af-255a5c08c463" containerID="a274c699396862b155ed01dcce70105fd9d181f992bc978485d814af3feb19ee" exitCode=0 Dec 01 09:16:17 crc kubenswrapper[4813]: I1201 09:16:17.427877 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-s69x9" event={"ID":"609748aa-5b9f-499e-89af-255a5c08c463","Type":"ContainerDied","Data":"a274c699396862b155ed01dcce70105fd9d181f992bc978485d814af3feb19ee"} Dec 01 09:16:17 crc kubenswrapper[4813]: I1201 09:16:17.428315 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-s69x9" event={"ID":"609748aa-5b9f-499e-89af-255a5c08c463","Type":"ContainerStarted","Data":"95635960fc6928151bdfc4aee14d7d4811a5108c6fcc800a3d767fc2a991db51"} Dec 01 09:16:18 crc kubenswrapper[4813]: I1201 09:16:18.443900 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-s69x9" event={"ID":"609748aa-5b9f-499e-89af-255a5c08c463","Type":"ContainerStarted","Data":"4f8ffadf20ca7dc4229b12dea79e8edaa12e56f7858fc060b33525e84acb9c79"} Dec 01 09:16:19 crc kubenswrapper[4813]: E1201 09:16:19.450891 4813 configmap.go:193] Couldn't get configMap glance-kuttl-tests/openstack-config: configmap "openstack-config" not found Dec 01 09:16:19 crc kubenswrapper[4813]: E1201 09:16:19.451040 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/4d0873f6-c7e7-4a9b-bd6e-a0263fd2bc5a-openstack-config podName:4d0873f6-c7e7-4a9b-bd6e-a0263fd2bc5a nodeName:}" failed. No retries permitted until 2025-12-01 09:18:21.451005945 +0000 UTC m=+2286.073927531 (durationBeforeRetry 2m2s). Error: MountVolume.SetUp failed for volume "openstack-config" (UniqueName: "kubernetes.io/configmap/4d0873f6-c7e7-4a9b-bd6e-a0263fd2bc5a-openstack-config") pod "openstackclient" (UID: "4d0873f6-c7e7-4a9b-bd6e-a0263fd2bc5a") : configmap "openstack-config" not found Dec 01 09:16:19 crc kubenswrapper[4813]: E1201 09:16:19.451055 4813 secret.go:188] Couldn't get secret glance-kuttl-tests/openstack-config-secret: secret "openstack-config-secret" not found Dec 01 09:16:19 crc kubenswrapper[4813]: E1201 09:16:19.451215 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/4d0873f6-c7e7-4a9b-bd6e-a0263fd2bc5a-openstack-config-secret podName:4d0873f6-c7e7-4a9b-bd6e-a0263fd2bc5a nodeName:}" failed. No retries permitted until 2025-12-01 09:18:21.451174289 +0000 UTC m=+2286.074095915 (durationBeforeRetry 2m2s). Error: MountVolume.SetUp failed for volume "openstack-config-secret" (UniqueName: "kubernetes.io/secret/4d0873f6-c7e7-4a9b-bd6e-a0263fd2bc5a-openstack-config-secret") pod "openstackclient" (UID: "4d0873f6-c7e7-4a9b-bd6e-a0263fd2bc5a") : secret "openstack-config-secret" not found Dec 01 09:16:19 crc kubenswrapper[4813]: I1201 09:16:19.456429 4813 generic.go:334] "Generic (PLEG): container finished" podID="609748aa-5b9f-499e-89af-255a5c08c463" containerID="4f8ffadf20ca7dc4229b12dea79e8edaa12e56f7858fc060b33525e84acb9c79" exitCode=0 Dec 01 09:16:19 crc kubenswrapper[4813]: I1201 09:16:19.456494 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-s69x9" event={"ID":"609748aa-5b9f-499e-89af-255a5c08c463","Type":"ContainerDied","Data":"4f8ffadf20ca7dc4229b12dea79e8edaa12e56f7858fc060b33525e84acb9c79"} Dec 01 09:16:21 crc kubenswrapper[4813]: I1201 09:16:21.471197 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-s69x9" event={"ID":"609748aa-5b9f-499e-89af-255a5c08c463","Type":"ContainerStarted","Data":"bce84c3f3c8889a7544a84df160f7a3fce610c497bf46bf3ea3b0893b274493a"} Dec 01 09:16:26 crc kubenswrapper[4813]: I1201 09:16:26.177259 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-s69x9" Dec 01 09:16:26 crc kubenswrapper[4813]: I1201 09:16:26.178565 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-s69x9" Dec 01 09:16:26 crc kubenswrapper[4813]: I1201 09:16:26.241070 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-s69x9" Dec 01 09:16:26 crc kubenswrapper[4813]: I1201 09:16:26.275446 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-s69x9" podStartSLOduration=8.292929572 podStartE2EDuration="11.275395518s" podCreationTimestamp="2025-12-01 09:16:15 +0000 UTC" firstStartedPulling="2025-12-01 09:16:17.430639977 +0000 UTC m=+2162.053561603" lastFinishedPulling="2025-12-01 09:16:20.413105943 +0000 UTC m=+2165.036027549" observedRunningTime="2025-12-01 09:16:21.495419634 +0000 UTC m=+2166.118341220" watchObservedRunningTime="2025-12-01 09:16:26.275395518 +0000 UTC m=+2170.898317104" Dec 01 09:16:26 crc kubenswrapper[4813]: I1201 09:16:26.562974 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-s69x9" Dec 01 09:16:26 crc kubenswrapper[4813]: I1201 09:16:26.607401 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-s69x9"] Dec 01 09:16:28 crc kubenswrapper[4813]: I1201 09:16:28.526989 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-s69x9" podUID="609748aa-5b9f-499e-89af-255a5c08c463" containerName="registry-server" containerID="cri-o://bce84c3f3c8889a7544a84df160f7a3fce610c497bf46bf3ea3b0893b274493a" gracePeriod=2 Dec 01 09:16:29 crc kubenswrapper[4813]: I1201 09:16:29.536258 4813 generic.go:334] "Generic (PLEG): container finished" podID="609748aa-5b9f-499e-89af-255a5c08c463" containerID="bce84c3f3c8889a7544a84df160f7a3fce610c497bf46bf3ea3b0893b274493a" exitCode=0 Dec 01 09:16:29 crc kubenswrapper[4813]: I1201 09:16:29.536319 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-s69x9" event={"ID":"609748aa-5b9f-499e-89af-255a5c08c463","Type":"ContainerDied","Data":"bce84c3f3c8889a7544a84df160f7a3fce610c497bf46bf3ea3b0893b274493a"} Dec 01 09:16:29 crc kubenswrapper[4813]: I1201 09:16:29.988698 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-s69x9" Dec 01 09:16:30 crc kubenswrapper[4813]: I1201 09:16:30.049104 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rbt25\" (UniqueName: \"kubernetes.io/projected/609748aa-5b9f-499e-89af-255a5c08c463-kube-api-access-rbt25\") pod \"609748aa-5b9f-499e-89af-255a5c08c463\" (UID: \"609748aa-5b9f-499e-89af-255a5c08c463\") " Dec 01 09:16:30 crc kubenswrapper[4813]: I1201 09:16:30.049169 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/609748aa-5b9f-499e-89af-255a5c08c463-catalog-content\") pod \"609748aa-5b9f-499e-89af-255a5c08c463\" (UID: \"609748aa-5b9f-499e-89af-255a5c08c463\") " Dec 01 09:16:30 crc kubenswrapper[4813]: I1201 09:16:30.049220 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/609748aa-5b9f-499e-89af-255a5c08c463-utilities\") pod \"609748aa-5b9f-499e-89af-255a5c08c463\" (UID: \"609748aa-5b9f-499e-89af-255a5c08c463\") " Dec 01 09:16:30 crc kubenswrapper[4813]: I1201 09:16:30.050101 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/609748aa-5b9f-499e-89af-255a5c08c463-utilities" (OuterVolumeSpecName: "utilities") pod "609748aa-5b9f-499e-89af-255a5c08c463" (UID: "609748aa-5b9f-499e-89af-255a5c08c463"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 09:16:30 crc kubenswrapper[4813]: I1201 09:16:30.057864 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/609748aa-5b9f-499e-89af-255a5c08c463-kube-api-access-rbt25" (OuterVolumeSpecName: "kube-api-access-rbt25") pod "609748aa-5b9f-499e-89af-255a5c08c463" (UID: "609748aa-5b9f-499e-89af-255a5c08c463"). InnerVolumeSpecName "kube-api-access-rbt25". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:16:30 crc kubenswrapper[4813]: I1201 09:16:30.102408 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/609748aa-5b9f-499e-89af-255a5c08c463-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "609748aa-5b9f-499e-89af-255a5c08c463" (UID: "609748aa-5b9f-499e-89af-255a5c08c463"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 09:16:30 crc kubenswrapper[4813]: I1201 09:16:30.151455 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rbt25\" (UniqueName: \"kubernetes.io/projected/609748aa-5b9f-499e-89af-255a5c08c463-kube-api-access-rbt25\") on node \"crc\" DevicePath \"\"" Dec 01 09:16:30 crc kubenswrapper[4813]: I1201 09:16:30.151504 4813 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/609748aa-5b9f-499e-89af-255a5c08c463-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 09:16:30 crc kubenswrapper[4813]: I1201 09:16:30.151515 4813 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/609748aa-5b9f-499e-89af-255a5c08c463-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 09:16:30 crc kubenswrapper[4813]: I1201 09:16:30.550035 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-s69x9" event={"ID":"609748aa-5b9f-499e-89af-255a5c08c463","Type":"ContainerDied","Data":"95635960fc6928151bdfc4aee14d7d4811a5108c6fcc800a3d767fc2a991db51"} Dec 01 09:16:30 crc kubenswrapper[4813]: I1201 09:16:30.550112 4813 scope.go:117] "RemoveContainer" containerID="bce84c3f3c8889a7544a84df160f7a3fce610c497bf46bf3ea3b0893b274493a" Dec 01 09:16:30 crc kubenswrapper[4813]: I1201 09:16:30.550454 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-s69x9" Dec 01 09:16:30 crc kubenswrapper[4813]: I1201 09:16:30.573576 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-s69x9"] Dec 01 09:16:30 crc kubenswrapper[4813]: I1201 09:16:30.578406 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-s69x9"] Dec 01 09:16:30 crc kubenswrapper[4813]: I1201 09:16:30.583459 4813 scope.go:117] "RemoveContainer" containerID="4f8ffadf20ca7dc4229b12dea79e8edaa12e56f7858fc060b33525e84acb9c79" Dec 01 09:16:30 crc kubenswrapper[4813]: I1201 09:16:30.604772 4813 scope.go:117] "RemoveContainer" containerID="a274c699396862b155ed01dcce70105fd9d181f992bc978485d814af3feb19ee" Dec 01 09:16:32 crc kubenswrapper[4813]: I1201 09:16:32.400833 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="609748aa-5b9f-499e-89af-255a5c08c463" path="/var/lib/kubelet/pods/609748aa-5b9f-499e-89af-255a5c08c463/volumes" Dec 01 09:16:47 crc kubenswrapper[4813]: I1201 09:16:47.221433 4813 patch_prober.go:28] interesting pod/machine-config-daemon-9wn4q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 09:16:47 crc kubenswrapper[4813]: I1201 09:16:47.222167 4813 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9wn4q" podUID="3d423402-e08f-42e9-b60c-6eb9bf2f6d48" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 09:16:51 crc kubenswrapper[4813]: I1201 09:16:51.518051 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-rw775"] Dec 01 09:16:51 crc kubenswrapper[4813]: E1201 09:16:51.518740 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="609748aa-5b9f-499e-89af-255a5c08c463" containerName="registry-server" Dec 01 09:16:51 crc kubenswrapper[4813]: I1201 09:16:51.518768 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="609748aa-5b9f-499e-89af-255a5c08c463" containerName="registry-server" Dec 01 09:16:51 crc kubenswrapper[4813]: E1201 09:16:51.518803 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="609748aa-5b9f-499e-89af-255a5c08c463" containerName="extract-content" Dec 01 09:16:51 crc kubenswrapper[4813]: I1201 09:16:51.518815 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="609748aa-5b9f-499e-89af-255a5c08c463" containerName="extract-content" Dec 01 09:16:51 crc kubenswrapper[4813]: E1201 09:16:51.518842 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="609748aa-5b9f-499e-89af-255a5c08c463" containerName="extract-utilities" Dec 01 09:16:51 crc kubenswrapper[4813]: I1201 09:16:51.518852 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="609748aa-5b9f-499e-89af-255a5c08c463" containerName="extract-utilities" Dec 01 09:16:51 crc kubenswrapper[4813]: I1201 09:16:51.519031 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="609748aa-5b9f-499e-89af-255a5c08c463" containerName="registry-server" Dec 01 09:16:51 crc kubenswrapper[4813]: I1201 09:16:51.520408 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rw775" Dec 01 09:16:51 crc kubenswrapper[4813]: I1201 09:16:51.520766 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-rw775"] Dec 01 09:16:51 crc kubenswrapper[4813]: I1201 09:16:51.600235 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vmstn\" (UniqueName: \"kubernetes.io/projected/fe4672ae-5104-4635-8e60-7daaea7b2772-kube-api-access-vmstn\") pod \"redhat-operators-rw775\" (UID: \"fe4672ae-5104-4635-8e60-7daaea7b2772\") " pod="openshift-marketplace/redhat-operators-rw775" Dec 01 09:16:51 crc kubenswrapper[4813]: I1201 09:16:51.600293 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fe4672ae-5104-4635-8e60-7daaea7b2772-catalog-content\") pod \"redhat-operators-rw775\" (UID: \"fe4672ae-5104-4635-8e60-7daaea7b2772\") " pod="openshift-marketplace/redhat-operators-rw775" Dec 01 09:16:51 crc kubenswrapper[4813]: I1201 09:16:51.600367 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fe4672ae-5104-4635-8e60-7daaea7b2772-utilities\") pod \"redhat-operators-rw775\" (UID: \"fe4672ae-5104-4635-8e60-7daaea7b2772\") " pod="openshift-marketplace/redhat-operators-rw775" Dec 01 09:16:51 crc kubenswrapper[4813]: I1201 09:16:51.701686 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fe4672ae-5104-4635-8e60-7daaea7b2772-utilities\") pod \"redhat-operators-rw775\" (UID: \"fe4672ae-5104-4635-8e60-7daaea7b2772\") " pod="openshift-marketplace/redhat-operators-rw775" Dec 01 09:16:51 crc kubenswrapper[4813]: I1201 09:16:51.701765 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vmstn\" (UniqueName: \"kubernetes.io/projected/fe4672ae-5104-4635-8e60-7daaea7b2772-kube-api-access-vmstn\") pod \"redhat-operators-rw775\" (UID: \"fe4672ae-5104-4635-8e60-7daaea7b2772\") " pod="openshift-marketplace/redhat-operators-rw775" Dec 01 09:16:51 crc kubenswrapper[4813]: I1201 09:16:51.701805 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fe4672ae-5104-4635-8e60-7daaea7b2772-catalog-content\") pod \"redhat-operators-rw775\" (UID: \"fe4672ae-5104-4635-8e60-7daaea7b2772\") " pod="openshift-marketplace/redhat-operators-rw775" Dec 01 09:16:51 crc kubenswrapper[4813]: I1201 09:16:51.702345 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fe4672ae-5104-4635-8e60-7daaea7b2772-catalog-content\") pod \"redhat-operators-rw775\" (UID: \"fe4672ae-5104-4635-8e60-7daaea7b2772\") " pod="openshift-marketplace/redhat-operators-rw775" Dec 01 09:16:51 crc kubenswrapper[4813]: I1201 09:16:51.702510 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fe4672ae-5104-4635-8e60-7daaea7b2772-utilities\") pod \"redhat-operators-rw775\" (UID: \"fe4672ae-5104-4635-8e60-7daaea7b2772\") " pod="openshift-marketplace/redhat-operators-rw775" Dec 01 09:16:51 crc kubenswrapper[4813]: I1201 09:16:51.723914 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vmstn\" (UniqueName: \"kubernetes.io/projected/fe4672ae-5104-4635-8e60-7daaea7b2772-kube-api-access-vmstn\") pod \"redhat-operators-rw775\" (UID: \"fe4672ae-5104-4635-8e60-7daaea7b2772\") " pod="openshift-marketplace/redhat-operators-rw775" Dec 01 09:16:51 crc kubenswrapper[4813]: I1201 09:16:51.853399 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rw775" Dec 01 09:16:52 crc kubenswrapper[4813]: I1201 09:16:52.049392 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-rw775"] Dec 01 09:16:52 crc kubenswrapper[4813]: I1201 09:16:52.704664 4813 generic.go:334] "Generic (PLEG): container finished" podID="fe4672ae-5104-4635-8e60-7daaea7b2772" containerID="379aa4f265921ca184b43f0570a34451884612ff1dc410ca051d71a196b8f073" exitCode=0 Dec 01 09:16:52 crc kubenswrapper[4813]: I1201 09:16:52.704721 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rw775" event={"ID":"fe4672ae-5104-4635-8e60-7daaea7b2772","Type":"ContainerDied","Data":"379aa4f265921ca184b43f0570a34451884612ff1dc410ca051d71a196b8f073"} Dec 01 09:16:52 crc kubenswrapper[4813]: I1201 09:16:52.704753 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rw775" event={"ID":"fe4672ae-5104-4635-8e60-7daaea7b2772","Type":"ContainerStarted","Data":"ae09eecbd6200badf3b791378f50de05490bed230065d0baaae33573b98cc9e6"} Dec 01 09:16:53 crc kubenswrapper[4813]: I1201 09:16:53.712479 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rw775" event={"ID":"fe4672ae-5104-4635-8e60-7daaea7b2772","Type":"ContainerStarted","Data":"0aa359302f91353e7694964abf827eaf2dc2c600cafac508b9f1679ec1f31abb"} Dec 01 09:16:53 crc kubenswrapper[4813]: I1201 09:16:53.908390 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-v2gmq"] Dec 01 09:16:53 crc kubenswrapper[4813]: I1201 09:16:53.909694 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-v2gmq" Dec 01 09:16:53 crc kubenswrapper[4813]: I1201 09:16:53.919360 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-v2gmq"] Dec 01 09:16:54 crc kubenswrapper[4813]: I1201 09:16:54.040238 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8f263451-06f4-4f65-8d59-9a34bdcb350e-utilities\") pod \"community-operators-v2gmq\" (UID: \"8f263451-06f4-4f65-8d59-9a34bdcb350e\") " pod="openshift-marketplace/community-operators-v2gmq" Dec 01 09:16:54 crc kubenswrapper[4813]: I1201 09:16:54.040978 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kgjbn\" (UniqueName: \"kubernetes.io/projected/8f263451-06f4-4f65-8d59-9a34bdcb350e-kube-api-access-kgjbn\") pod \"community-operators-v2gmq\" (UID: \"8f263451-06f4-4f65-8d59-9a34bdcb350e\") " pod="openshift-marketplace/community-operators-v2gmq" Dec 01 09:16:54 crc kubenswrapper[4813]: I1201 09:16:54.041120 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8f263451-06f4-4f65-8d59-9a34bdcb350e-catalog-content\") pod \"community-operators-v2gmq\" (UID: \"8f263451-06f4-4f65-8d59-9a34bdcb350e\") " pod="openshift-marketplace/community-operators-v2gmq" Dec 01 09:16:54 crc kubenswrapper[4813]: I1201 09:16:54.142500 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8f263451-06f4-4f65-8d59-9a34bdcb350e-catalog-content\") pod \"community-operators-v2gmq\" (UID: \"8f263451-06f4-4f65-8d59-9a34bdcb350e\") " pod="openshift-marketplace/community-operators-v2gmq" Dec 01 09:16:54 crc kubenswrapper[4813]: I1201 09:16:54.142576 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8f263451-06f4-4f65-8d59-9a34bdcb350e-utilities\") pod \"community-operators-v2gmq\" (UID: \"8f263451-06f4-4f65-8d59-9a34bdcb350e\") " pod="openshift-marketplace/community-operators-v2gmq" Dec 01 09:16:54 crc kubenswrapper[4813]: I1201 09:16:54.142656 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kgjbn\" (UniqueName: \"kubernetes.io/projected/8f263451-06f4-4f65-8d59-9a34bdcb350e-kube-api-access-kgjbn\") pod \"community-operators-v2gmq\" (UID: \"8f263451-06f4-4f65-8d59-9a34bdcb350e\") " pod="openshift-marketplace/community-operators-v2gmq" Dec 01 09:16:54 crc kubenswrapper[4813]: I1201 09:16:54.143489 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8f263451-06f4-4f65-8d59-9a34bdcb350e-catalog-content\") pod \"community-operators-v2gmq\" (UID: \"8f263451-06f4-4f65-8d59-9a34bdcb350e\") " pod="openshift-marketplace/community-operators-v2gmq" Dec 01 09:16:54 crc kubenswrapper[4813]: I1201 09:16:54.143496 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8f263451-06f4-4f65-8d59-9a34bdcb350e-utilities\") pod \"community-operators-v2gmq\" (UID: \"8f263451-06f4-4f65-8d59-9a34bdcb350e\") " pod="openshift-marketplace/community-operators-v2gmq" Dec 01 09:16:54 crc kubenswrapper[4813]: I1201 09:16:54.165475 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kgjbn\" (UniqueName: \"kubernetes.io/projected/8f263451-06f4-4f65-8d59-9a34bdcb350e-kube-api-access-kgjbn\") pod \"community-operators-v2gmq\" (UID: \"8f263451-06f4-4f65-8d59-9a34bdcb350e\") " pod="openshift-marketplace/community-operators-v2gmq" Dec 01 09:16:54 crc kubenswrapper[4813]: I1201 09:16:54.226489 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-v2gmq" Dec 01 09:16:54 crc kubenswrapper[4813]: I1201 09:16:54.714422 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-v2gmq"] Dec 01 09:16:54 crc kubenswrapper[4813]: I1201 09:16:54.737500 4813 generic.go:334] "Generic (PLEG): container finished" podID="fe4672ae-5104-4635-8e60-7daaea7b2772" containerID="0aa359302f91353e7694964abf827eaf2dc2c600cafac508b9f1679ec1f31abb" exitCode=0 Dec 01 09:16:54 crc kubenswrapper[4813]: I1201 09:16:54.737815 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rw775" event={"ID":"fe4672ae-5104-4635-8e60-7daaea7b2772","Type":"ContainerDied","Data":"0aa359302f91353e7694964abf827eaf2dc2c600cafac508b9f1679ec1f31abb"} Dec 01 09:16:55 crc kubenswrapper[4813]: I1201 09:16:55.746792 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rw775" event={"ID":"fe4672ae-5104-4635-8e60-7daaea7b2772","Type":"ContainerStarted","Data":"0af9eae3225d82881e3f882c9e6a54df1b4ec6f79d931a5d2a5e25aad6b5d6b5"} Dec 01 09:16:55 crc kubenswrapper[4813]: I1201 09:16:55.749743 4813 generic.go:334] "Generic (PLEG): container finished" podID="8f263451-06f4-4f65-8d59-9a34bdcb350e" containerID="17a21d8962c8a087d5c8174c75944056ffa66988c07fa7aeb9e2e02ae8d17372" exitCode=0 Dec 01 09:16:55 crc kubenswrapper[4813]: I1201 09:16:55.749807 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-v2gmq" event={"ID":"8f263451-06f4-4f65-8d59-9a34bdcb350e","Type":"ContainerDied","Data":"17a21d8962c8a087d5c8174c75944056ffa66988c07fa7aeb9e2e02ae8d17372"} Dec 01 09:16:55 crc kubenswrapper[4813]: I1201 09:16:55.749877 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-v2gmq" event={"ID":"8f263451-06f4-4f65-8d59-9a34bdcb350e","Type":"ContainerStarted","Data":"fd8f0f92afe9ed77b9461ac74847303dad42b481c0961f04d8c43de0380277b0"} Dec 01 09:16:55 crc kubenswrapper[4813]: I1201 09:16:55.773861 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-rw775" podStartSLOduration=2.200031047 podStartE2EDuration="4.773836008s" podCreationTimestamp="2025-12-01 09:16:51 +0000 UTC" firstStartedPulling="2025-12-01 09:16:52.706450772 +0000 UTC m=+2197.329372358" lastFinishedPulling="2025-12-01 09:16:55.280255693 +0000 UTC m=+2199.903177319" observedRunningTime="2025-12-01 09:16:55.773125738 +0000 UTC m=+2200.396047324" watchObservedRunningTime="2025-12-01 09:16:55.773836008 +0000 UTC m=+2200.396757594" Dec 01 09:16:57 crc kubenswrapper[4813]: I1201 09:16:57.764843 4813 generic.go:334] "Generic (PLEG): container finished" podID="8f263451-06f4-4f65-8d59-9a34bdcb350e" containerID="570000a6c911ed0e1e68fe92cc5938ab77265464b3542227305c8c73a0a3176f" exitCode=0 Dec 01 09:16:57 crc kubenswrapper[4813]: I1201 09:16:57.765279 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-v2gmq" event={"ID":"8f263451-06f4-4f65-8d59-9a34bdcb350e","Type":"ContainerDied","Data":"570000a6c911ed0e1e68fe92cc5938ab77265464b3542227305c8c73a0a3176f"} Dec 01 09:16:58 crc kubenswrapper[4813]: I1201 09:16:58.774241 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-v2gmq" event={"ID":"8f263451-06f4-4f65-8d59-9a34bdcb350e","Type":"ContainerStarted","Data":"49a6185bcf76689a59bc4473d2fd06c44698c0c4a3402bf5f6dff7cdca67d5f7"} Dec 01 09:16:58 crc kubenswrapper[4813]: I1201 09:16:58.792939 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-v2gmq" podStartSLOduration=3.245686116 podStartE2EDuration="5.79292144s" podCreationTimestamp="2025-12-01 09:16:53 +0000 UTC" firstStartedPulling="2025-12-01 09:16:55.75313422 +0000 UTC m=+2200.376055806" lastFinishedPulling="2025-12-01 09:16:58.300369524 +0000 UTC m=+2202.923291130" observedRunningTime="2025-12-01 09:16:58.78977325 +0000 UTC m=+2203.412694836" watchObservedRunningTime="2025-12-01 09:16:58.79292144 +0000 UTC m=+2203.415843026" Dec 01 09:17:01 crc kubenswrapper[4813]: I1201 09:17:01.854051 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-rw775" Dec 01 09:17:01 crc kubenswrapper[4813]: I1201 09:17:01.854415 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-rw775" Dec 01 09:17:01 crc kubenswrapper[4813]: I1201 09:17:01.922710 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-rw775" Dec 01 09:17:02 crc kubenswrapper[4813]: I1201 09:17:02.848867 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-rw775" Dec 01 09:17:03 crc kubenswrapper[4813]: I1201 09:17:03.500695 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-rw775"] Dec 01 09:17:04 crc kubenswrapper[4813]: I1201 09:17:04.227147 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-v2gmq" Dec 01 09:17:04 crc kubenswrapper[4813]: I1201 09:17:04.229084 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-v2gmq" Dec 01 09:17:04 crc kubenswrapper[4813]: I1201 09:17:04.280712 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-v2gmq" Dec 01 09:17:04 crc kubenswrapper[4813]: I1201 09:17:04.816825 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-rw775" podUID="fe4672ae-5104-4635-8e60-7daaea7b2772" containerName="registry-server" containerID="cri-o://0af9eae3225d82881e3f882c9e6a54df1b4ec6f79d931a5d2a5e25aad6b5d6b5" gracePeriod=2 Dec 01 09:17:04 crc kubenswrapper[4813]: I1201 09:17:04.857760 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-v2gmq" Dec 01 09:17:05 crc kubenswrapper[4813]: I1201 09:17:05.297184 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-v2gmq"] Dec 01 09:17:06 crc kubenswrapper[4813]: I1201 09:17:06.325108 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rw775" Dec 01 09:17:06 crc kubenswrapper[4813]: I1201 09:17:06.453528 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fe4672ae-5104-4635-8e60-7daaea7b2772-utilities\") pod \"fe4672ae-5104-4635-8e60-7daaea7b2772\" (UID: \"fe4672ae-5104-4635-8e60-7daaea7b2772\") " Dec 01 09:17:06 crc kubenswrapper[4813]: I1201 09:17:06.453856 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vmstn\" (UniqueName: \"kubernetes.io/projected/fe4672ae-5104-4635-8e60-7daaea7b2772-kube-api-access-vmstn\") pod \"fe4672ae-5104-4635-8e60-7daaea7b2772\" (UID: \"fe4672ae-5104-4635-8e60-7daaea7b2772\") " Dec 01 09:17:06 crc kubenswrapper[4813]: I1201 09:17:06.453914 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fe4672ae-5104-4635-8e60-7daaea7b2772-catalog-content\") pod \"fe4672ae-5104-4635-8e60-7daaea7b2772\" (UID: \"fe4672ae-5104-4635-8e60-7daaea7b2772\") " Dec 01 09:17:06 crc kubenswrapper[4813]: I1201 09:17:06.454700 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fe4672ae-5104-4635-8e60-7daaea7b2772-utilities" (OuterVolumeSpecName: "utilities") pod "fe4672ae-5104-4635-8e60-7daaea7b2772" (UID: "fe4672ae-5104-4635-8e60-7daaea7b2772"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 09:17:06 crc kubenswrapper[4813]: I1201 09:17:06.471606 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fe4672ae-5104-4635-8e60-7daaea7b2772-kube-api-access-vmstn" (OuterVolumeSpecName: "kube-api-access-vmstn") pod "fe4672ae-5104-4635-8e60-7daaea7b2772" (UID: "fe4672ae-5104-4635-8e60-7daaea7b2772"). InnerVolumeSpecName "kube-api-access-vmstn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:17:06 crc kubenswrapper[4813]: I1201 09:17:06.554939 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fe4672ae-5104-4635-8e60-7daaea7b2772-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "fe4672ae-5104-4635-8e60-7daaea7b2772" (UID: "fe4672ae-5104-4635-8e60-7daaea7b2772"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 09:17:06 crc kubenswrapper[4813]: I1201 09:17:06.555267 4813 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fe4672ae-5104-4635-8e60-7daaea7b2772-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 09:17:06 crc kubenswrapper[4813]: I1201 09:17:06.555289 4813 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fe4672ae-5104-4635-8e60-7daaea7b2772-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 09:17:06 crc kubenswrapper[4813]: I1201 09:17:06.555300 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vmstn\" (UniqueName: \"kubernetes.io/projected/fe4672ae-5104-4635-8e60-7daaea7b2772-kube-api-access-vmstn\") on node \"crc\" DevicePath \"\"" Dec 01 09:17:06 crc kubenswrapper[4813]: I1201 09:17:06.838675 4813 generic.go:334] "Generic (PLEG): container finished" podID="fe4672ae-5104-4635-8e60-7daaea7b2772" containerID="0af9eae3225d82881e3f882c9e6a54df1b4ec6f79d931a5d2a5e25aad6b5d6b5" exitCode=0 Dec 01 09:17:06 crc kubenswrapper[4813]: I1201 09:17:06.838816 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rw775" Dec 01 09:17:06 crc kubenswrapper[4813]: I1201 09:17:06.838933 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-v2gmq" podUID="8f263451-06f4-4f65-8d59-9a34bdcb350e" containerName="registry-server" containerID="cri-o://49a6185bcf76689a59bc4473d2fd06c44698c0c4a3402bf5f6dff7cdca67d5f7" gracePeriod=2 Dec 01 09:17:06 crc kubenswrapper[4813]: I1201 09:17:06.838809 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rw775" event={"ID":"fe4672ae-5104-4635-8e60-7daaea7b2772","Type":"ContainerDied","Data":"0af9eae3225d82881e3f882c9e6a54df1b4ec6f79d931a5d2a5e25aad6b5d6b5"} Dec 01 09:17:06 crc kubenswrapper[4813]: I1201 09:17:06.839559 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rw775" event={"ID":"fe4672ae-5104-4635-8e60-7daaea7b2772","Type":"ContainerDied","Data":"ae09eecbd6200badf3b791378f50de05490bed230065d0baaae33573b98cc9e6"} Dec 01 09:17:06 crc kubenswrapper[4813]: I1201 09:17:06.839610 4813 scope.go:117] "RemoveContainer" containerID="0af9eae3225d82881e3f882c9e6a54df1b4ec6f79d931a5d2a5e25aad6b5d6b5" Dec 01 09:17:06 crc kubenswrapper[4813]: I1201 09:17:06.885472 4813 scope.go:117] "RemoveContainer" containerID="0aa359302f91353e7694964abf827eaf2dc2c600cafac508b9f1679ec1f31abb" Dec 01 09:17:06 crc kubenswrapper[4813]: I1201 09:17:06.892362 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-rw775"] Dec 01 09:17:06 crc kubenswrapper[4813]: I1201 09:17:06.898540 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-rw775"] Dec 01 09:17:06 crc kubenswrapper[4813]: I1201 09:17:06.987481 4813 scope.go:117] "RemoveContainer" containerID="379aa4f265921ca184b43f0570a34451884612ff1dc410ca051d71a196b8f073" Dec 01 09:17:07 crc kubenswrapper[4813]: I1201 09:17:07.005978 4813 scope.go:117] "RemoveContainer" containerID="0af9eae3225d82881e3f882c9e6a54df1b4ec6f79d931a5d2a5e25aad6b5d6b5" Dec 01 09:17:07 crc kubenswrapper[4813]: E1201 09:17:07.006686 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0af9eae3225d82881e3f882c9e6a54df1b4ec6f79d931a5d2a5e25aad6b5d6b5\": container with ID starting with 0af9eae3225d82881e3f882c9e6a54df1b4ec6f79d931a5d2a5e25aad6b5d6b5 not found: ID does not exist" containerID="0af9eae3225d82881e3f882c9e6a54df1b4ec6f79d931a5d2a5e25aad6b5d6b5" Dec 01 09:17:07 crc kubenswrapper[4813]: I1201 09:17:07.006774 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0af9eae3225d82881e3f882c9e6a54df1b4ec6f79d931a5d2a5e25aad6b5d6b5"} err="failed to get container status \"0af9eae3225d82881e3f882c9e6a54df1b4ec6f79d931a5d2a5e25aad6b5d6b5\": rpc error: code = NotFound desc = could not find container \"0af9eae3225d82881e3f882c9e6a54df1b4ec6f79d931a5d2a5e25aad6b5d6b5\": container with ID starting with 0af9eae3225d82881e3f882c9e6a54df1b4ec6f79d931a5d2a5e25aad6b5d6b5 not found: ID does not exist" Dec 01 09:17:07 crc kubenswrapper[4813]: I1201 09:17:07.006841 4813 scope.go:117] "RemoveContainer" containerID="0aa359302f91353e7694964abf827eaf2dc2c600cafac508b9f1679ec1f31abb" Dec 01 09:17:07 crc kubenswrapper[4813]: E1201 09:17:07.007350 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0aa359302f91353e7694964abf827eaf2dc2c600cafac508b9f1679ec1f31abb\": container with ID starting with 0aa359302f91353e7694964abf827eaf2dc2c600cafac508b9f1679ec1f31abb not found: ID does not exist" containerID="0aa359302f91353e7694964abf827eaf2dc2c600cafac508b9f1679ec1f31abb" Dec 01 09:17:07 crc kubenswrapper[4813]: I1201 09:17:07.007385 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0aa359302f91353e7694964abf827eaf2dc2c600cafac508b9f1679ec1f31abb"} err="failed to get container status \"0aa359302f91353e7694964abf827eaf2dc2c600cafac508b9f1679ec1f31abb\": rpc error: code = NotFound desc = could not find container \"0aa359302f91353e7694964abf827eaf2dc2c600cafac508b9f1679ec1f31abb\": container with ID starting with 0aa359302f91353e7694964abf827eaf2dc2c600cafac508b9f1679ec1f31abb not found: ID does not exist" Dec 01 09:17:07 crc kubenswrapper[4813]: I1201 09:17:07.007420 4813 scope.go:117] "RemoveContainer" containerID="379aa4f265921ca184b43f0570a34451884612ff1dc410ca051d71a196b8f073" Dec 01 09:17:07 crc kubenswrapper[4813]: E1201 09:17:07.007802 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"379aa4f265921ca184b43f0570a34451884612ff1dc410ca051d71a196b8f073\": container with ID starting with 379aa4f265921ca184b43f0570a34451884612ff1dc410ca051d71a196b8f073 not found: ID does not exist" containerID="379aa4f265921ca184b43f0570a34451884612ff1dc410ca051d71a196b8f073" Dec 01 09:17:07 crc kubenswrapper[4813]: I1201 09:17:07.007858 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"379aa4f265921ca184b43f0570a34451884612ff1dc410ca051d71a196b8f073"} err="failed to get container status \"379aa4f265921ca184b43f0570a34451884612ff1dc410ca051d71a196b8f073\": rpc error: code = NotFound desc = could not find container \"379aa4f265921ca184b43f0570a34451884612ff1dc410ca051d71a196b8f073\": container with ID starting with 379aa4f265921ca184b43f0570a34451884612ff1dc410ca051d71a196b8f073 not found: ID does not exist" Dec 01 09:17:07 crc kubenswrapper[4813]: I1201 09:17:07.214022 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-v2gmq" Dec 01 09:17:07 crc kubenswrapper[4813]: I1201 09:17:07.371845 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8f263451-06f4-4f65-8d59-9a34bdcb350e-utilities\") pod \"8f263451-06f4-4f65-8d59-9a34bdcb350e\" (UID: \"8f263451-06f4-4f65-8d59-9a34bdcb350e\") " Dec 01 09:17:07 crc kubenswrapper[4813]: I1201 09:17:07.371980 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8f263451-06f4-4f65-8d59-9a34bdcb350e-catalog-content\") pod \"8f263451-06f4-4f65-8d59-9a34bdcb350e\" (UID: \"8f263451-06f4-4f65-8d59-9a34bdcb350e\") " Dec 01 09:17:07 crc kubenswrapper[4813]: I1201 09:17:07.372051 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kgjbn\" (UniqueName: \"kubernetes.io/projected/8f263451-06f4-4f65-8d59-9a34bdcb350e-kube-api-access-kgjbn\") pod \"8f263451-06f4-4f65-8d59-9a34bdcb350e\" (UID: \"8f263451-06f4-4f65-8d59-9a34bdcb350e\") " Dec 01 09:17:07 crc kubenswrapper[4813]: I1201 09:17:07.372897 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f263451-06f4-4f65-8d59-9a34bdcb350e-utilities" (OuterVolumeSpecName: "utilities") pod "8f263451-06f4-4f65-8d59-9a34bdcb350e" (UID: "8f263451-06f4-4f65-8d59-9a34bdcb350e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 09:17:07 crc kubenswrapper[4813]: I1201 09:17:07.375891 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f263451-06f4-4f65-8d59-9a34bdcb350e-kube-api-access-kgjbn" (OuterVolumeSpecName: "kube-api-access-kgjbn") pod "8f263451-06f4-4f65-8d59-9a34bdcb350e" (UID: "8f263451-06f4-4f65-8d59-9a34bdcb350e"). InnerVolumeSpecName "kube-api-access-kgjbn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 09:17:07 crc kubenswrapper[4813]: I1201 09:17:07.464334 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f263451-06f4-4f65-8d59-9a34bdcb350e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8f263451-06f4-4f65-8d59-9a34bdcb350e" (UID: "8f263451-06f4-4f65-8d59-9a34bdcb350e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 09:17:07 crc kubenswrapper[4813]: I1201 09:17:07.473497 4813 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8f263451-06f4-4f65-8d59-9a34bdcb350e-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 09:17:07 crc kubenswrapper[4813]: I1201 09:17:07.473529 4813 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8f263451-06f4-4f65-8d59-9a34bdcb350e-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 09:17:07 crc kubenswrapper[4813]: I1201 09:17:07.473539 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kgjbn\" (UniqueName: \"kubernetes.io/projected/8f263451-06f4-4f65-8d59-9a34bdcb350e-kube-api-access-kgjbn\") on node \"crc\" DevicePath \"\"" Dec 01 09:17:07 crc kubenswrapper[4813]: I1201 09:17:07.855473 4813 generic.go:334] "Generic (PLEG): container finished" podID="8f263451-06f4-4f65-8d59-9a34bdcb350e" containerID="49a6185bcf76689a59bc4473d2fd06c44698c0c4a3402bf5f6dff7cdca67d5f7" exitCode=0 Dec 01 09:17:07 crc kubenswrapper[4813]: I1201 09:17:07.855552 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-v2gmq" event={"ID":"8f263451-06f4-4f65-8d59-9a34bdcb350e","Type":"ContainerDied","Data":"49a6185bcf76689a59bc4473d2fd06c44698c0c4a3402bf5f6dff7cdca67d5f7"} Dec 01 09:17:07 crc kubenswrapper[4813]: I1201 09:17:07.855599 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-v2gmq" Dec 01 09:17:07 crc kubenswrapper[4813]: I1201 09:17:07.855632 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-v2gmq" event={"ID":"8f263451-06f4-4f65-8d59-9a34bdcb350e","Type":"ContainerDied","Data":"fd8f0f92afe9ed77b9461ac74847303dad42b481c0961f04d8c43de0380277b0"} Dec 01 09:17:07 crc kubenswrapper[4813]: I1201 09:17:07.855659 4813 scope.go:117] "RemoveContainer" containerID="49a6185bcf76689a59bc4473d2fd06c44698c0c4a3402bf5f6dff7cdca67d5f7" Dec 01 09:17:07 crc kubenswrapper[4813]: I1201 09:17:07.876428 4813 scope.go:117] "RemoveContainer" containerID="570000a6c911ed0e1e68fe92cc5938ab77265464b3542227305c8c73a0a3176f" Dec 01 09:17:07 crc kubenswrapper[4813]: I1201 09:17:07.899092 4813 scope.go:117] "RemoveContainer" containerID="17a21d8962c8a087d5c8174c75944056ffa66988c07fa7aeb9e2e02ae8d17372" Dec 01 09:17:07 crc kubenswrapper[4813]: I1201 09:17:07.909728 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-v2gmq"] Dec 01 09:17:07 crc kubenswrapper[4813]: I1201 09:17:07.917446 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-v2gmq"] Dec 01 09:17:07 crc kubenswrapper[4813]: I1201 09:17:07.926042 4813 scope.go:117] "RemoveContainer" containerID="49a6185bcf76689a59bc4473d2fd06c44698c0c4a3402bf5f6dff7cdca67d5f7" Dec 01 09:17:07 crc kubenswrapper[4813]: E1201 09:17:07.926696 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"49a6185bcf76689a59bc4473d2fd06c44698c0c4a3402bf5f6dff7cdca67d5f7\": container with ID starting with 49a6185bcf76689a59bc4473d2fd06c44698c0c4a3402bf5f6dff7cdca67d5f7 not found: ID does not exist" containerID="49a6185bcf76689a59bc4473d2fd06c44698c0c4a3402bf5f6dff7cdca67d5f7" Dec 01 09:17:07 crc kubenswrapper[4813]: I1201 09:17:07.926762 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"49a6185bcf76689a59bc4473d2fd06c44698c0c4a3402bf5f6dff7cdca67d5f7"} err="failed to get container status \"49a6185bcf76689a59bc4473d2fd06c44698c0c4a3402bf5f6dff7cdca67d5f7\": rpc error: code = NotFound desc = could not find container \"49a6185bcf76689a59bc4473d2fd06c44698c0c4a3402bf5f6dff7cdca67d5f7\": container with ID starting with 49a6185bcf76689a59bc4473d2fd06c44698c0c4a3402bf5f6dff7cdca67d5f7 not found: ID does not exist" Dec 01 09:17:07 crc kubenswrapper[4813]: I1201 09:17:07.926803 4813 scope.go:117] "RemoveContainer" containerID="570000a6c911ed0e1e68fe92cc5938ab77265464b3542227305c8c73a0a3176f" Dec 01 09:17:07 crc kubenswrapper[4813]: E1201 09:17:07.927592 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"570000a6c911ed0e1e68fe92cc5938ab77265464b3542227305c8c73a0a3176f\": container with ID starting with 570000a6c911ed0e1e68fe92cc5938ab77265464b3542227305c8c73a0a3176f not found: ID does not exist" containerID="570000a6c911ed0e1e68fe92cc5938ab77265464b3542227305c8c73a0a3176f" Dec 01 09:17:07 crc kubenswrapper[4813]: I1201 09:17:07.927631 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"570000a6c911ed0e1e68fe92cc5938ab77265464b3542227305c8c73a0a3176f"} err="failed to get container status \"570000a6c911ed0e1e68fe92cc5938ab77265464b3542227305c8c73a0a3176f\": rpc error: code = NotFound desc = could not find container \"570000a6c911ed0e1e68fe92cc5938ab77265464b3542227305c8c73a0a3176f\": container with ID starting with 570000a6c911ed0e1e68fe92cc5938ab77265464b3542227305c8c73a0a3176f not found: ID does not exist" Dec 01 09:17:07 crc kubenswrapper[4813]: I1201 09:17:07.927659 4813 scope.go:117] "RemoveContainer" containerID="17a21d8962c8a087d5c8174c75944056ffa66988c07fa7aeb9e2e02ae8d17372" Dec 01 09:17:07 crc kubenswrapper[4813]: E1201 09:17:07.927997 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"17a21d8962c8a087d5c8174c75944056ffa66988c07fa7aeb9e2e02ae8d17372\": container with ID starting with 17a21d8962c8a087d5c8174c75944056ffa66988c07fa7aeb9e2e02ae8d17372 not found: ID does not exist" containerID="17a21d8962c8a087d5c8174c75944056ffa66988c07fa7aeb9e2e02ae8d17372" Dec 01 09:17:07 crc kubenswrapper[4813]: I1201 09:17:07.928020 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"17a21d8962c8a087d5c8174c75944056ffa66988c07fa7aeb9e2e02ae8d17372"} err="failed to get container status \"17a21d8962c8a087d5c8174c75944056ffa66988c07fa7aeb9e2e02ae8d17372\": rpc error: code = NotFound desc = could not find container \"17a21d8962c8a087d5c8174c75944056ffa66988c07fa7aeb9e2e02ae8d17372\": container with ID starting with 17a21d8962c8a087d5c8174c75944056ffa66988c07fa7aeb9e2e02ae8d17372 not found: ID does not exist" Dec 01 09:17:08 crc kubenswrapper[4813]: I1201 09:17:08.408808 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f263451-06f4-4f65-8d59-9a34bdcb350e" path="/var/lib/kubelet/pods/8f263451-06f4-4f65-8d59-9a34bdcb350e/volumes" Dec 01 09:17:08 crc kubenswrapper[4813]: I1201 09:17:08.411895 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fe4672ae-5104-4635-8e60-7daaea7b2772" path="/var/lib/kubelet/pods/fe4672ae-5104-4635-8e60-7daaea7b2772/volumes" Dec 01 09:17:17 crc kubenswrapper[4813]: I1201 09:17:17.222735 4813 patch_prober.go:28] interesting pod/machine-config-daemon-9wn4q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 09:17:17 crc kubenswrapper[4813]: I1201 09:17:17.223907 4813 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9wn4q" podUID="3d423402-e08f-42e9-b60c-6eb9bf2f6d48" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 09:17:17 crc kubenswrapper[4813]: I1201 09:17:17.224029 4813 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-9wn4q" Dec 01 09:17:17 crc kubenswrapper[4813]: I1201 09:17:17.225637 4813 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"e0826a7d333367829ec415cbaa9314fd5e1db5cbc96bf148dadb2adfedeb3d6a"} pod="openshift-machine-config-operator/machine-config-daemon-9wn4q" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 01 09:17:17 crc kubenswrapper[4813]: I1201 09:17:17.225752 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-9wn4q" podUID="3d423402-e08f-42e9-b60c-6eb9bf2f6d48" containerName="machine-config-daemon" containerID="cri-o://e0826a7d333367829ec415cbaa9314fd5e1db5cbc96bf148dadb2adfedeb3d6a" gracePeriod=600 Dec 01 09:17:17 crc kubenswrapper[4813]: E1201 09:17:17.370929 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9wn4q_openshift-machine-config-operator(3d423402-e08f-42e9-b60c-6eb9bf2f6d48)\"" pod="openshift-machine-config-operator/machine-config-daemon-9wn4q" podUID="3d423402-e08f-42e9-b60c-6eb9bf2f6d48" Dec 01 09:17:17 crc kubenswrapper[4813]: I1201 09:17:17.930257 4813 generic.go:334] "Generic (PLEG): container finished" podID="3d423402-e08f-42e9-b60c-6eb9bf2f6d48" containerID="e0826a7d333367829ec415cbaa9314fd5e1db5cbc96bf148dadb2adfedeb3d6a" exitCode=0 Dec 01 09:17:17 crc kubenswrapper[4813]: I1201 09:17:17.930323 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-9wn4q" event={"ID":"3d423402-e08f-42e9-b60c-6eb9bf2f6d48","Type":"ContainerDied","Data":"e0826a7d333367829ec415cbaa9314fd5e1db5cbc96bf148dadb2adfedeb3d6a"} Dec 01 09:17:17 crc kubenswrapper[4813]: I1201 09:17:17.930806 4813 scope.go:117] "RemoveContainer" containerID="47c278315d64e3056cb6f536af9317d9d67616d501c03a3c61eb1b5cb51b2405" Dec 01 09:17:17 crc kubenswrapper[4813]: I1201 09:17:17.933773 4813 scope.go:117] "RemoveContainer" containerID="e0826a7d333367829ec415cbaa9314fd5e1db5cbc96bf148dadb2adfedeb3d6a" Dec 01 09:17:17 crc kubenswrapper[4813]: E1201 09:17:17.934221 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9wn4q_openshift-machine-config-operator(3d423402-e08f-42e9-b60c-6eb9bf2f6d48)\"" pod="openshift-machine-config-operator/machine-config-daemon-9wn4q" podUID="3d423402-e08f-42e9-b60c-6eb9bf2f6d48" var/home/core/zuul-output/logs/crc-cloud-workdir-crc-all-logs.tar.gz0000644000175000000000000000005515113256457024456 0ustar coreroot‹íÁ  ÷Om7 €7šÞ'(var/home/core/zuul-output/logs/crc-cloud/0000755000175000000000000000000015113256460017365 5ustar corerootvar/home/core/zuul-output/artifacts/0000755000175000017500000000000015113251635016507 5ustar corecorevar/home/core/zuul-output/docs/0000755000175000017500000000000015113251636015460 5ustar corecore